Published by UC Davis Health on June 11, the study marks a significant advance beyond previous technologies that were largely limited to translating brain signals into text. The new system restored a level of communication once thought impossible for a 45-year-old volunteer with amyotrophic lateral sclerosis (ALS), a neurodegenerative disease that had robbed him of his voice.
The “brain-to-voice neuroprosthesis” works by decoding the brain's intent to speak. Researchers implanted 256 microelectrodes into the region of the patient's brain responsible for controlling speech muscles. When the man attempts to talk, the BCI intercepts these signals and, with the help of an advanced AI model, converts them into audible speech in just 25 milliseconds.
The researchers used advanced AI algorithms to translate brain activity into synthesized speech in real time. The system was trained using neural recordings taken while the participant tried to read sentences shown on a screen. By aligning the firing patterns of hundreds of neurons with the intended speech sounds, the algorithm learned to accurately reconstruct the participant’s voice directly from his brain signals.
The BCI doesn't just produce monotonic words; it successfully captures and reproduces vocal intonation — the subtle shifts in pitch and tone that are fundamental to how humans communicate meaning and emotion. In a series of powerful demonstrations, the patient was able to:
- Ask questions by generating a rising pitch at the end of a sentence.
- Emphasize specific words to completely change a sentence's meaning, such as stressing different words in the phrase: I never said she stole my money.
- Sing simple 3-pitch melodies, showcasing a fine level of neural control over the synthesized voice.
To make the breakthrough even more profound, the researchers used a voice-cloning AI trained on old recordings of the patient made prior to his condition. The result was a synthesized voice that sounded like his own, a feature the patient reported “made me feel happy, and it felt like my real voice.”
While the technology represents a major breakthrough, the researchers caution that it is still a proof-of-concept. In tests where human listeners transcribed the BCI’s output, they understood what the patient was saying correctly about 56% of the time. More development will be needed to increase its efficiency.