Neuroengineers from the Mortimer B. Zuckerman Mind Brain Behavior Institute at Columbia University in the US have developed a brain-computer interface that directly translates thoughts into intelligible, recognisable speech.
The new system offers hope for people with limited or no ability to speak, including patients with amyotrophic lateral sclerosis (ALS) or those recovering from a stroke. Around one in three patients who have had a stroke have some kind of problem with speech.
Based on speech synthesisers and artificial intelligence (AI), the interface tracks brain activity and clearly reconstructs the words heard by a person.
The team believes that the new system has the potential to facilitate computers that can directly communicate with the brain.
Zuckerman Mind Brain Behavior Institute principal investigator Dr Nima Mesgarani said: “Our voices help connect us to our friends, family and the world around us, which is why losing the power of one’s voice due to injury or disease is so devastating.
“With today’s study, we have a potential way to restore that power. We’ve shown that, with the right technology, these people’s thoughts could be decoded and understood by any listener.”
How well do you really know your competitors?
Access the most comprehensive Company Profiles on the market, powered by GlobalData. Save hours of research. Gain competitive edge.
Thank you!
Your download email will arrive shortly
Not ready to buy yet? Download a free sample
We are confident about the unique quality of our Company Profiles. However, we want you to make the most beneficial decision for your business, so we offer a free sample that you can download by submitting the below form
By GlobalDataTo create the system, the team used a computer algorithm called vocoder, which synthesises speech after being trained on recordings of people talking.
Commonly, this technology is leveraged by Amazon Echo and Apple Siri.
The researchers partnered with Northwell Health Physician Partners Neuroscience Institute neurosurgeon Dr Ashesh Dinesh Mehta to train the vocoder in interpreting brain activity.
Epilepsy patients treated by Mehta were asked to listen to certain sentences and digits, and their brain signals were recorded to be run through the vocoder.
The sound generated by the vocoder in response to these signals was analysed and cleaned up by AI-based neural networks, which mimic the structure of biological neurons.
This led to the production of a robotic-sounding voice. People were able to understand and repeat the produced sounds in about 75% of the cases, which is said to be significantly higher than previous attempts.
While the system requires further training and testing, it is hoped to be applied in implants that can be worn and directly translate the user’s thoughts into words.
Mesgarani added: “This would be a game changer. It would give anyone who has lost their ability to speak, whether through injury or disease, the renewed chance to connect to the world around them.”
Additional reporting by Charlotte Edwards.