Nature: New AI technology helps patients with facial paralysis to communicate

Two paralyzed patients have been able to communicate with unprecedented accuracy and speed through the use of artificial intelligence (AI)-enhanced brain-machine implants.

a6023299cdb09cccaa15aa7c8cb08353.png

A brain-computer interface converts participants' brain signals into animated speech and facial movements. Source: Noah Berger

Published separately in the journal Nature (IF2022=64.8) on August 23, two research teams describe a brain-computer interface (BCI) that converts brain signals into text or synthesized speech. These BCIs decode speech at rates of 62 words per minute and 78 words per minute, respectively. These new techniques are all faster than any previous attempt, despite the rate of natural conversation at about 160 words per minute.

"It is now conceivable in the future that it might be possible to restore fluent communication to paralyzed patients, allowing them to say whatever they want with sufficient accuracy," neuroscientist Francis Willett of Stanford University in California said in an August 22 news conference. High, can be reliably understood."

Such devices "could become products in the near future," says Christian Herff, a computational neuroscientist at Maastricht University in the Netherlands.

Electrodes and Algorithms

Willett and colleagues developed a BCI that interprets neural activity at the cellular level and translates it into text [1]. They collaborated on the study with Pat Bennett, a 67-year-old patient with motor neurone disease, also known as amyotrophic lateral sclerosis, a condition in which the gradual loss of muscle control leads to movement and difficulty speaking.

76b0e05029fa141f0a00fa64ae3ef764.png

Paralyzed people use their brain activity to control prosthetics. Source: Petter/UPMC

First, the researchers performed surgery on Bennett, implanting a set of small silicon electrodes a few millimeters below the language-related brain region. They then trained a deep-learning algorithm to identify unique signals in Bennett's brain as he tried to utter various phrases using the large vocabulary of 125,000 words and the small vocabulary of 50 words. The AI ​​decodes words from phonemes, the subunits of speech that form spoken language. For a vocabulary of 50 words, the BCI works 2.7 times faster than the earlier state-of-the-art BCI [3] with a word error rate of 9.1%. For a vocabulary of 125,000 words, the error rate rises to 23.8%. "About three-quarters of the words were correctly decoded," Willett said in a news release.

"For those who cannot speak, it means they can stay in touch with the larger world, perhaps continue to work and maintain friends and family relationships, " Bennett said in a statement to reporters.

reading brain activity

In another study [2], neurosurgeon Edward Chang of the University of California, San Francisco, and colleagues worked with Ann, a 47-year-old woman who lost her ability to speak after a brainstem stroke 18 years earlier.

They took a different approach from Willett's team, placing a thin rectangle containing 253 electrodes on the surface of the cortex to record neural activity. The technique, called electrocorticography (ECoG), is considered less invasive and can record the combined activity of thousands of neurons simultaneously. The team trained the AI ​​algorithm to recognize patterns in the brain activity associated with Ann's attempts to speak 249 sentences, using a vocabulary of 1,024 words. The device ended up producing 78 words per minute with a median word error rate of 25.5%.

Although the implant used by Willett's team captures neural activity more precisely and performs better on larger vocabularies, Blaise Yvert, a neurotechnology researcher at the Krenoble Institute for Neuroscience, said, "Seeing that ECoG can achieve low error Rates are a good thing."

Chang and his team also created custom algorithms, trained using recordings from Ann's wedding footage, to convert Ann's brain signals into animated avatars with synthetic voices and simulated facial expressions. Ultimately, they created a personalized voice that sounded like Ann did before her injury.

In a feedback session after the study, Ann told the researchers: "The simple fact of hearing a voice similar to my own is emotionally charged. When I can speak for myself, it means a lot to me! "

"Voice is a very important part of our identity," Chang said. "It's not just about communication, it's about who we are."

Clinical application

b1bde198aaf74e8b503f8143499e2f45.png

An exoskeleton controlled by brain activity was tested on an able-bodied boy. Source: 160over90

Many improvements are needed before BCIs can be used clinically. "The ideal would be a wireless connection," Ann told the researchers. A BCI suitable for everyday use would have to be a fully implantable system with no visible connectors or cables, Yvert added. Both teams hope to continue improving the speed and accuracy of their devices with more powerful decoding algorithms.

Participants in both studies were still able to use facial muscles when thinking about speaking, and brain regions associated with language were intact, Herff said. "It doesn't necessarily work for every patient."

"We see it as a proof of concept, just to provide momentum for people in the field to turn it into a usable product," Willett said.

The devices must also be tested in larger groups of people to prove their reliability. "No matter how elegant the data and how sophisticated the technology, we have to understand them in a very careful way," says Judy Illes, a neuroethics researcher at the University of British Columbia in Vancouver, Canada. "We have to be careful not to overcommit," she adds. Its broad applicability to large populations, I'm not sure we've done that yet."

References :

[1].  Willett, F. R. et al. Nature https://doi.org/10.1038/s41586-023-06377-x (2023).

[2]. Metzger, SL et al. Nature https://doi.org/10.1038/s41586-023-06443-4 (2023).

[3]. Moses, DA et al. N. Engl. J. Med. 385, 217–227 (2021).

Read the original content :

https://doi.org/10.1038/d41586-023-02682-7

Past products (click on the picture to go directly to the text corresponding tutorial)

5768f9b9a271113616326c825774dc30.jpeg

8f24163ba9877b551597f766380ff40f.jpeg

4908e2cf1322c288a709fca093861251.jpeg

9ac1a42ea655d5e511afe9d30263d0a0.jpeg

55fe05c317ce183aa906c55860495679.jpeg

66e0b15895b3355cb3f1aa7513290739.jpeg

b95d173901a37ac5082c4733f611390e.jpeg

c2a4909fa2865e57a1f30b3bd547c7c4.jpeg

6e8182afd9f7c9b30027db51681a413f.jpeg

cfd5668d86ec53c5388c8181c4d129dc.jpeg

2fe05972b67e9580521dc14033c6ebae.jpeg

121449dd74edfb9ccd8f9346353045b9.jpeg

b3a7008aa64d5ef3fd11120772e37974.png

700637f70257113b95c0cf24993063ca.png

b49ed5f910403f8f71f6aead43b6abf7.png

3d81b915a33dbd0b880d5dec92596eb6.png

aec13034849077e14972629191fc275b.jpeg

fb697a84bb44ca0f90afede202aa1e3f.jpeg

16b59ce492d717c5093db25ad1b42e36.jpeg

1d492cde8a0a37ad984d5cdd9b4bc2ff.jpeg

64a2df14b65e0260a45f2c32ba961f2a.png

581387cf7f5294daf38e013d94f558d0.png

b4f6853ddcab09390040ae59199e43c7.jpeg

09f20f34e039c40b8f6c21fe6aa202a1.png

9ad13a7c81f5042431d383dbb0120cbf.png

95acc3ee203a76980154d6375e407e2f.jpeg

714ed1a8a434a535550ae40bba334eb2.png

0ec3677bcce68ece2be22b66b7e080e9.png

machine learning

6504f8ce8adfdc7a911983f81ffb37e4.png

eb8ee32f0d8cc71673c418c418f011ba.jpeg

508109895c0949678b1f042b092a709b.jpeg

f68b1d88ce4764816cc9536538dc4604.png

Guess you like

Origin blog.csdn.net/qazplm12_3/article/details/132550990