kinect h4x n.
Skip this Video
Download Presentation
Kinect H4x

Loading in 2 Seconds...

play fullscreen
1 / 16

Kinect H4x - PowerPoint PPT Presentation

  • Uploaded on

Kinect H4x. Gesture Recognition and Playback Tools (+Inspiration). SDK Version 1.0 - Out TODAY What's New?. Ability to control which user(s) are under full skeletal tracking. "Near mode" enables interaction as close as 40cm from the device. Includes "too far" and "too close" depth indicators:

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
Download Presentation

PowerPoint Slideshow about 'Kinect H4x' - bertille

Download Now An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript
kinect h4x

Kinect H4x

Gesture Recognition and Playback Tools (+Inspiration)

sdk version 1 0 out today what s new
SDK Version 1.0 - Out TODAYWhat's New?

Ability to control which user(s) are under full skeletal tracking.

"Near mode" enables interaction as close as 40cm from the device. Includes "too far" and "too close"

depth indicators:

Better everything

Mother of All Kinect Demos - Kinect Explorer sample app shows off all features (camera tilt, audio beam angles, etc).

gesture recognition dynamic time warping
Gesture Recognition: Dynamic Time Warping

A sequence matching algorithm that can adapt to sequences that vary in speed and time. (think Levenshtein distances, but generalized to matching any sort of input to stored data)

It measures the similarity between two sequences based on a cost function of how much it needs to "warp" the points forward/backward in time to have them line up.

In Kinect-land, this means an algorithm that can take streaming joint data and quickly find the closest match to a gesture 'on record'.

but how does it work and how can i work into my own project
But how does it work?And how can I work into my own project?

Three classes:

Skeleton2DDataExtract.cs -- Takes in Kinect SDK joint data, spits out normalized 2d skeleton points.

Skeleton2DdataCoordEventArgs.cs -- Defines the event args that get emitted from the Skeleton2DDataExtract event handler once processed.

DtwGestureRecognizer.cs - Parse the 2d skeleton data and call Recognize() to match against loaded gestures (see code for loading/saving example).

really advanced hacks
Really Advanced Hacks

The DtwGestureRecognizer can flexibly match any vectorized data stream.

We happened to use skeleton data in our example, but it should be fairly simple to incorporate a depth stream or color stream. Just ensure that each of the sequence objects you pass into AddOrUpdate and Recognize is an array of doubles (e.g. double[] observationPoint).

working on p4 collaboratively recording and replaying skeleton data
Working on P4 Collaboratively:Recording and Replaying Skeleton Data

In light of the fact that many teams share a single Kinect, it might be of use to you to be able to record a sequence of skeleton data, write it to a file, and replay it back through a dummy nui.SkeletonFrameReady handler.

Fortunately, the Kinect Toolbox (not to be confused with the Coding4Fun Kinect Toolkit), allows us to do just that.

Get it at:

candescent nui demo
Candescent NUI (Demo)

The Cool News: Hand + Finger Tracking!

The Bad News: Behemoth, undocumented code library

Project available at

Kinect SDK only provides depth values from 800mm out. The finger tracking code only works in the range of .8 - 1m, so using this project in conjunction with the Kinect SDK will prove difficult. 

UPDATE: NEW SDK 1.0 provides depth values from 400mm out!

Compatible With Candescent Alternative: OpenNI + NITE uses the raw point cloud to make best-guess tracking estimates < 800mm away. Good community, documentation at /

inspirational project
Inspirational Project:

Deixis Application To Children's Education Games

Main Idea: Ask children to point and verbally identify ("this one!") a subset of objects (numbers, colors, animals).

Description + Video

inspirational hacks
Inspirational Hacks

Gesture Enabled Garden Hose

Main Idea: Use servos (simple motors) in conjunction with netduinos (network-enabled microcontrollers) to control the servo via Kinect gestures:

In Practice:

inspirational hacks pt 2
Inspirational Hacks Pt. 2

Gesture Based Electronic Music Performance

inspirational hacks pt 3
Inspirational Hacks Pt. 3

EDEN: Interactive Ecosystem Simulation Software

Main Idea: Create a topographical landscape on the iPad, fill it with (simulated) water, project it onto a sandscape via depth data with an overhead Kinect + projector. Play with the sand to change the climate and topography to terraform your own sandscape.


Kinect Telepresence -

Home Security Camera -

Living Paintings -

Visually Impaired Navigation Tool-

ZigFu: Single bundle install NITE, OpenNI, PrimeSense Sensor, everything you need to work outside of the official SDK: