AlterEgo is a project from MIT to allow you to interface with a computer via neuromuscular face twitches, translating subtle movements into text. You need to wear a sensor which looks a little like a confused facehugger that attached to your hear from the rear. The headset uses bone conduction to detect sub-vocalized speach, you do not even need to move your mouth in order for your words to register. The Register reports that AlterEgo has an an impressive 92% accuracy rate, good enough for use but with the possibility of some humorous results. This will be a huge boon for those who lack the ability to speak as well as offering a way to interface with VR by removing the need for a keyboard to enter text. Check it out here.
"Boffins at MIT have developed a silent speech interface called AlterEgo that allows wearers to communicate via text translated from neuromuscular signals – tiny face twitches – without a visible tell."
Here is some more Tech News from around the web:
- Microsoft will be adding 64-bit support to Windows for ARM soon @ The Inquirer
- The Supreme Court Fight Over Microsoft's Foreign Servers Is Over @ Slashdot
- Office 365 goes down across Europe, again @ The Inquirer
- Cisco switch flaw lead to attacks on critical infrastructure @ The Inquirer
“You need to wear a sensor
“You need to wear a sensor which looks a little like a confused facehugger that attached to your hear from the rear.”
Here from the rear…
“speach” “an an” “hear”
“speach” “an an” “hear”
Guys…
Not as impressive as the
Not as impressive as the recent software that can detect if someone is a homosexual from photographs. Next they will be able to tell if you like chocolate or vanilla from the curve of your eyebrow.
Is that data obtained from
Is that data obtained from facial features or is it taken from the clothing worn by said individuals.