Shows the user's avatar as a set of cubes, with one cube per tracked joint. The orientation of each cube corresponds to the orientation of the joint.
An avatar facing the user. This is a good starting place for some augmented reality applications, where the user can see their own body from the front.
The user can move their hand left and right to hover over an item, then do a tap gesture to select it. Start here for building a gesture-based user interface.
The streaming depth data of the user from the 3D sensor is shown as a set of fiery particles.
The user controls multiple skeletons, each of which mimic the user's motions.
Displays the depth stream, image stream (from the webcam in the sensor), and user map (which parts of the depth image correspond to which user). On its own, this is a good test to make sure everything is configured properly. It also demonstrates how to get access to the various data streams.
A top-down view of users' locations relative to the sensor. Useful to provide visual feedback of which users are being tracked.