ScienceAn AI can decode speech from brain activity with...

An AI can decode speech from brain activity with surprising accuracy


An artificial intelligence can decode words and sentences from brain activity with surprising — but still limited — accuracy. Using only a few seconds of brain activity data, the AI guesses what a person has heard. It lists the correct answer in its top 10 possibilities up to 73 percent of the time, researchers found in a preliminary study.

The AI’s “performance was above what many people thought was possible at this stage,” says Giovanni Di Liberto, a computer scientist at Trinity College Dublin who was not involved in the research.

Developed at the parent company of Facebook, Meta, the AI could eventually be used to help thousands of people around the world unable to communicate through speech, typing or gestures, researchers report August 25 at arXiv.org. That includes many patients in minimally conscious, locked-in or “vegetative states” — what’s now generally known as unresponsive wakefulness syndrome (SN: 2/8/19).

Most existing technologies to help such patients communicate require risky brain surgeries to implant electrodes. This new approach “could provide a viable path to help patients with communication deficits … without the use of invasive methods,” says neuroscientist Jean-Rémi King, a Meta AI researcher currently at the École Normale Supérieure in Paris.

King and his colleagues trained a computational tool to detect words and sentences on 56,000 hours of speech recordings from 53 languages. The tool, also known as a language model, learned how to recognize specific features of language both at a fine-grained level — think letters or syllables — and at a broader level, such as a word or sentence.

The team applied an AI with this language model to databases from four institutions that included brain activity from 169 volunteers. In these databases, participants listened to various stories and sentences from, for example, Ernest Hemingway’s The Old Man and the Sea and Lewis Carroll’s Alices Adventures in Wonderland while the people’s brains were scanned using either magnetoencephalography or electroencephalography. Those techniques measure the magnetic or electrical component of brain signals.

Then with the help of a computational method that helps account for physical differences among actual brains, the team tried to decode what participants had heard using just three seconds of brain activity data from each person. The team instructed the AI to align the speech sounds from the story recordings to patterns of brain activity that the AI computed as corresponding to what people were hearing. It then made predictions about what the person might have been hearing during that short time, given more than 1,000 possibilities.

Using magnetoencephalography, or MEG, the correct answer was in the AI’s top 10 guesses up to 73 percent of the time, the researchers found. With electroencephalography, that value dropped to no more than 30 percent. “[That MEG] performance is very good,” Di Liberto says, but he’s less optimistic about its practical use. “What can we do with it? Nothing. Absolutely nothing.”

The reason, he says, is that MEG requires a bulky and expensive machine. Bringing this technology to clinics will require scientific innovations that make the machines cheaper and easier to use.

It’s also important to understand what “decoding” really means in this study, says Jonathan Brennan, a linguist at the University of Michigan in Ann Arbor. The word is often used to describe the process of deciphering information directly from a source — in this case, speech from brain activity. But the AI could do this only because it was provided a finite list of possible correct answers to make its guesses.

“With language, that’s not going to cut it if we want to scale to practical use, because language is infinite,” Brennan says. 

What’s more, Di Liberto says, the AI decoded information of participants passively listening to audio, which is not directly relevant to nonverbal patients. For it to become a meaningful communication tool, scientists will need to learn how to decrypt from brain activity what these patients intend on saying, including expressions of hunger, discomfort or a simple “yes” or “no.”

The new study is “decoding of speech perception, not production,” King agrees. Though speech production is the ultimate goal, for now, “we’re quite a long way away.”



Original Source Link

Latest News

Mark Zuckerberg warns Meta staff of ‘intense year’ ahead as he announces low performers will be cut

Mark Zuckerberg is taking aim at Meta's low performers. The CEO told employees to prepare for an "intense...

Malaysia mulls crypto policy after talks with UAE and Binance founder CZ

Malaysia is reportedly exploring cryptocurrency regulations after its prime minister held discussions with Abu Dhabi leaders and Binance...

World Bank approves 10-year $20bn Pakistan lending package

Unlock the Editor’s Digest for freeRoula Khalaf, Editor of the FT, selects her favourite stories in this weekly...

Quantum Brilliance, which makes quantum computing hardware using synthetic diamond qubits that work at room temperature in any setting, raised a $20M Series A...

Mike Wheatley / SiliconANGLE: Quantum Brilliance, which makes quantum computing hardware using synthetic diamond qubits that work at...

The House In Chaos As Nancy Mace Challenges Jasmine Crockett To A Fight

To support PoliticusUSA and our new ad-free platform, please consider becoming a subscriber.At a House hearing, this was...

Hand-feeding squirrels accidentally changed their skulls

Soft diet, weak jaws. If red squirrels eat too many peanuts, their jaws end up weaker than the...

Must Read

California wildfire season should be over. So why is L.A. burning?

Unusually dry conditions and hurricane-force seasonal winds are...

Ethereum Sees $1.4 Billion In Exchange Outflows This Week – Strong Accumulation Trend?

Este artículo también está disponible en español. Ethereum has...
- Advertisement -

You might also likeRELATED
Recommended to you