Kinect for Windows v2 brings you the latest in human computing technologies, enabling the development of applications that allow people to interact naturally with computers by simply gesturing and speaking.
The Kinect for Windows sensor is a physical device with depth sensing technology, a built-in color camera, an infrared (IR) emitter, and a microphone array, enabling it to sense the location and movements of individual humans as well as their voices. With up to 3x higher depth fidelity, the v2 sensor provides significant improvements in visualizing small objects—and all objects more clearly.
The Kinect for Window software development kit (SDK) provides developers with drivers, tools, APIs, device interfaces, and code samples to facilitate the development of Kinect-enabled applications for commercial deployment. With improved skeletal tracking in the SDK 2.0 public preview, you will be able to track as many as six people and 25 joints per person. With more anatomically correct positions, your application can discern movements more precisely.
Together, the sensor and SDK provide developers with the foundation to create and deploy interactive applications that respond to people's movements, gestures, and voice commands, resulting in natural computer interactions.
Individuals and businesses worldwide are using the sensor to run Kinect for Windows natural human computing applications in a range of settings.
Learn more about how people are using Kinect for Windows >
Kinect for Windows v2
The Kinect for Windows v2 sensor and the SDK 2.0 take natural user interactions with computers to the next level, offering greater overall precision, responsiveness, and intuitive capabilities to accelerate the development of voice and gesture experiences on computers. The v2 sensor’s color camera is enhanced with full 1080p video that can be displayed in the same resolution as the viewing screen. The higher depth fidelity makes it significantly easier to see smaller objects, to see all objects more clearly, and to view objects in 3-D. In addition to now tracking as many as six people and 25 skeletal joints per person, the tracked positions are more anatomically correct and stable—and the range of tracking is broader.
When the final version of the SDK 2.0 releases to the public later this year, you will be able to commercially deploy your solutions and sell your Kinect v2 applications directly in the Windows Store, allowing you to reach more potential customers than ever, from businesses to end users to other developers.
Key SDK 2.0 public preview features and benefits
Improved skeletal, hand and joint orientation
With the ability to track as many as six people and 25 skeletal joints per person—including new joints for hand tips, thumbs, and shoulder center—and Improved understanding of the soft connective tissue and body positioning, you get more anatomically correct positions for crisp interactions and more accurate avateering and more lifelike avatars.
New and better scenarios in fitness, wellness, education and training, entertainment, gaming, movies, and communications.
Support for new development environments
New Unity support provides faster, cost-efficient, and high quality support for cross-platform development, enabling developers to build apps for the Windows Store by using tools you already know.
Build and publish apps to the Windows Store using tools you already know across multiple platforms.
With Kinect Studio’s enhanced recording and playback, developers can develop on the go, without the need to carry the Kinect sensor with them. And visual gesture builder lets developers build their own custom gestures that the system recognizes and uses to write code by using machine learning. These features increase productivity and keep costs down.
Increase your productivity and cost efficiency.
Advanced face tracking
With the resolution increased 20 times, applications can capture a face with a 2,000-point mesh that looks more true to life. This means that avatars will look more lifelike.
Build more lifelike avatars.
Simultaneous multi-app support
Improved multi-app support enables multiple applications to access a single sensor simultaneously.
For instance, by enabling a retail app and a business intelligence app access to the same sensor, you can get analytics in real time while customers are using a Kinect experience you have deployed in your store.
Key sensor features and benefits
Improved skeletal tracking
The enhanced fidelity of the depth camera, combined with improvements in the software, have led to a number skeletal tracking developments. In addition to now tracking as many as six complete skeletons (compared to two with the original sensor), and tracking 25 joints per person (as compared to 20 with the original sensor), the tracked positions are more anatomically correct and stable—and the range of tracking is broader.
Enables more stable avateering, more accurate body position evaluation, crisper interactions, and more bystander involvement in interactive scenarios
512 x 484
FOV: 70 x 60
One mode: .5–4.5 meters
With higher depth fidelity and a significantly improved noise floor, the v2 sensor gives you improved 3-D visualization, improved ability to see smaller objects and all objects more clearly, and improves the stability of skeletal tracking.
Improved 3-D visualization for fitness, wellness, and entertainment scenarios
1080p color camera
30 Hz (15 Hz in low light)
The color camera captures full, beautiful 1080p video that can be displayed in the same resolution as the viewing screen, allowing for a broad range of powerful scenarios. In addition to improving video communications and video analytics applications, this provides a great input on which to build high-quality, augmented reality scenarios, digital signage, and more.
Build high-quality, augmented reality scenarios. For example, digital signage for retail, museums, lobbies, and public spaces.
New active infrared (IR) capabilities 512x484, 30 Hz
In addition to allowing the Kinect for Windows v2 sensor to see in the dark, the new infrared (IR) capabilities produce a lighting-independent view, which makes machine learning or computer-vision–based tasks much easier—because you don’t have to account for or model lighting-based variation. And, you can now use IR and color at the same time. We look forward to the many new and innovative uses that the community will develop to use this fundamentally new capability.
New machine learning applications
Wider/expanded field of view
The expanded field of view enables a larger area of a scene to be captured by the camera. As a result, users can be closer to the camera and still in view, and the camera is effective over a larger total area.
More effective over a larger area
More sensor information
Sensor support and software updates
Both versions of the Kinect for Windows sensor include a one-year limited warranty, support, and access to software updates for both speech and human tracking.
Sensor technical support