Hacker News new | past | comments | ask | show | jobs | submit login

I feel like that cuts out some of the uses... perhaps you could us ML techniques to separate out different voices and then individually transcribe them?

edit: wow this is a complicated problem now that I am thinking about it... the amount of processing the brain does...




I may be wrong, but I believe Kinect uses the information from its vision system that identifies people, to help inform the microphone array system with data that it can use to identify which of the people in front of it is talking at any one time...




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: