Originally posted by DrSpike
View Post
But it's in a completely another league. The PS3 solution is another implementation of the Wii Waggle system, which I find to be wholly inadequate.
The MS solution is a "me too" in it's motion control, but without a controller. There's a huge difference.
- It also integrates a mic array for voice recognition and spatial recognition (where is the sound coming from, which speaker is it coming from?)
- It integrates IR projectors which determine with pretty amazing accuracy the depth of all objects in its image (a z-buffer) and two optical HD cameras to judge x and y-axis positioning (and provide reference for the z-buffer)
- It has an onboard chip to do realtime 3D modelling of the scene in question (does not require additional CPU usage to merge the two HD images and apply the z-buffer IR results). It also maps the skeletal points of people on the screen to the 3D scene which it can send to the 360 for processing.
I've been digging around at the Microsoft Research work Molyneaux mentioned they're leveraging for this project. I've only looked at the MS Research "Speech Group" first, but with very interesting results:
Understanding user's intent from speech: http://research.microsoft.com/en-us/...g/default.aspx
Microphone array processing and spatial sound: http://research.microsoft.com/en-us/...g/default.aspx
Language modeling: http://research.microsoft.com/en-us/...g/default.aspx
Multimodal conversational user interface: http://research.microsoft.com/en-us/...u/default.aspx
Speaker identification (who is speaking?): http://research.microsoft.com/en-us/...d/default.aspx
SAPI (Speech API): http://research.microsoft.com/en-us/...i/default.aspx
A lot of this stuff was explicitly developed with an Xbox 360-like environment in mind:
Comment