So now we have a depth image as shown below and not only colored image. As active vision is used, result is sure better than passive stereo vision.
So now we have a depth image as shown on the right and not only a colored image. But what is the range and the precision of the Kinect ?
Thanks to the UCL Department of Civil, Environmental and Geomatic Engineering, we have a Depth resolution vs distance study.
- The first step now is to extract a user out of this Depth data: In the University of Texas at Austin
Department of Electrical and Computer Engineering, they combine different techniques to detect the human.
- The second step is to use Decision forests to learn how a person looks like, learning it part by part, or guessing where the closest joint is with regression.
The rest of the story is that today, a bunch of players are jumping in their living room in front of their TV. Tomorrow, we might be not necessarily jumping for gaming, but making hand signs to control things around as a Natural User Interface.
The key point to these kind of applications goes beyond the Kinect sensor, and relies onto machine learning for what do we want to recognize, hands, objects…