From thoughts to words: How AI deciphers neural signals to help a man with ALS speak (2024)

From thoughts to words: How AI deciphers neural signals to help a man with ALS speak (1)

Brain-computer interfaces are a groundbreaking technology that can help paralyzed people regain functions they've lost, like moving a hand. These devices record signals from the brain and decipher the user's intended action, bypassing damaged or degraded nerves that would normally transmit those brain signals to control muscles.

Since 2006, demonstrations of brain-computer interfaces in humans have primarily focused on restoring arm and hand movements by enabling people to control computer cursors or robotic arms. Recently, researchers have begun developing speech brain-computer interfaces to restore communication for people who cannot speak.

As the user attempts to talk, these brain-computer interfaces record the person's unique brain signals associated with attempted muscle movements for speaking and then translate them into words. These words can then be displayed as text on a screen or spoken aloud using text-to-speech software.

I'm a researcher in the Neuroprosthetics Lab at the University of California, Davis, which is part of the BrainGate2 clinical trial.

My colleagues and I recently demonstrated a speech brain-computer interface that deciphers the attempted speech of a man with ALS, or amyotrophic lateral sclerosis, also known as Lou Gehrig's disease. The interface converts neural signals into text with over 97% accuracy. Key to our system is a set of artificial intelligence language models—artificial neural networks that help interpret natural ones.

Recording brain signals

The first step in our speech brain-computer interface is recording brain signals. There are several sources of brain signals, some of which require surgery to record. Surgically implanted recording devices can capture high-quality brain signals because they are placed closer to neurons, resulting in stronger signals with less interference. These neural recording devices include grids of electrodes placed on the brain's surface or electrodes implanted directly into brain tissue.

In our study, we used electrode arrays surgically placed in the speech motor cortex, the part of the brain that controls muscles related to speech, of the participant, Casey Harrell. We recorded neural activity from 256 electrodes as Harrell attempted to speak.

Decoding brain signals

The next challenge is relating the complex brain signals to the words the user is trying to say.

One approach is to map neural activity patterns directly to spoken words. This method requires recording brain signals corresponding to each word multiple times to identify the average relationship between neural activity and specific words.

While this strategy works well for small vocabularies, as demonstrated in a 2021 study with a 50-word vocabulary, it becomes impractical for larger ones. Imagine asking the brain-computer interface user to try to say every word in the dictionary multiple times—it could take months, and it still wouldn't work for new words.

Instead, we use an alternative strategy: mapping brain signals to phonemes, the basic units of sound that make up words. In English, there are 39 phonemes, including ch, er, oo, pl and sh, that can be combined to form any word.

We can measure the neural activity associated with every phoneme multiple times just by asking the participant to read a few sentences aloud. By accurately mapping neural activity to phonemes, we can assemble them into any English word, even ones the system wasn't explicitly trained with.

To map brain signals to phonemes, we use advanced machine learning models. These models are particularly well-suited for this task due to their ability to find patterns in large amounts of complex data that would be impossible for humans to discern.

Think of these models as super-smart listeners that can pick out important information from noisy brain signals, much like you might focus on a conversation in a crowded room. Using these models, we were able to decipher phoneme sequences during attempted speech with over 90% accuracy.

From phonemes to words

Once we have the deciphered phoneme sequences, we need to convert them into words and sentences. This is challenging, especially if the deciphered phoneme sequence isn't perfectly accurate. To solve this puzzle, we use two complementary types of machine learning language models.

From thoughts to words: How AI deciphers neural signals to help a man with ALS speak (2)

The first is n-gram language models, which predict which word is most likely to follow a set of "n" words. We trained a 5-gram, or five-word, language model on millions of sentences to predict the likelihood of a word based on the previous four words, capturing local context and common phrases. For example, after "I am very good," it might suggest "today" as more likely than "potato."

Using this model, we convert our phoneme sequences into the 100 most likely word sequences, each with an associated probability.

The second is large language models, which power AI chatbots and also predict which words most likely follow others. We use large language models to refine our choices. These models, trained on vast amounts of diverse text, have a broader understanding of language structure and meaning. They help us determine which of our 100 candidate sentences makes the most sense in a wider context.

By carefully balancing probabilities from the n-gram model, the large language model and our initial phoneme predictions, we can make a highly educated guess about what the brain-computer interface user is trying to say. This multistep process allows us to handle the uncertainties in phoneme decoding and produce coherent, contextually appropriate sentences.

Real-world benefits

In practice, this speech decoding strategy has been remarkably successful. We've enabled Casey Harrell, a man with ALS, to "speak" with over 97% accuracy using just his thoughts. This breakthrough allows him to easily converse with his family and friends for the first time in years, all in the comfort of his own home.

Speech brain-computer interfaces represent a significant step forward in restoring communication. As we continue to refine these devices, they hold the promise of giving a voice to those who have lost the ability to speak, reconnecting them with their loved ones and the world around them.

However, challenges remain, such as making the technology more accessible, portable and durable over years of use. Despite these hurdles, speech brain-computer interfaces are a powerful example of how science and technology can come together to solve complex problems and dramatically improve people's lives.

Provided byThe Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.From thoughts to words: How AI deciphers neural signals to help a man with ALS speak (3)

Citation: From thoughts to words: How AI deciphers neural signals to help a man with ALS speak (2024, August 24) retrieved 1 September 2024 from https://medicalxpress.com/news/2024-08-thoughts-words-ai-deciphers-neural.html

This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

From thoughts to words: How AI deciphers neural signals to help a man with ALS speak (2024)

References

Top Articles
Becoming a Living Kidney Donor: 6 Things to Know About the Process - UF Health
When will stores open on bank holiday Monday as weekend shopping hours revealed?
No Hard Feelings Showtimes Near Metropolitan Fiesta 5 Theatre
Faketoks Twitter
Louisville Kentucky Craigslist Cars And Trucks By Owner
Inmate Inquiry Mendocino
Black Adam Movies123
Rick Lee Oaklawn Park Picks Today
Dryers At Abc Warehouse
S10 Mpg
Family Guy Wiki Peter
1V1.Lol Pizza Edition
Craigslist Pets Peoria Il
Jennette Mccurdy Tmz Hawaii
Best Builder Hall 5 Base
Shore Lodge Webcam
Four-Legged Friday: Meet Tuscaloosa's Adoptable All-Stars Cub & Pickle
Amy Riley Electric Video
New Jersey Map | Map of New Jersey | NJ Map
Stafford Rotoworld
Christmas Song Figgerits
Lerntools und Lösungen für Bildungseinrichtungen - Google for Education
What is a Nutmeg in Soccer? (Explained!) - Soccer Knowledge Hub
Retire Early Wsbtv.com Free Book
Craigslist Caldwell Id
Highway 420 East Bremerton
MovieHaX.Click
Pa Lottery Remaining Prizes Scratch Offs
Elizabeth Nj Garbage Schedule 2022
Two Brothers Pizza Middletown Pa
Realidades 2 Capitulo 2B Answers
Acnh Picnic Table
Antonio Brown Football Reference
Walmart Neighborhood Market Pharmacy Phone Number
Us 25 Yard Sale Map
MAELLE MAGNETISEUSE A ST-MALO ATTENUE VOTRE LUMBAGO
Sacramento Library Overdrive
Ice Quartz Osrs
Rachel Pizzolato Age, Height, Wiki, Net Worth, Measurement
Fallen Avatar Mythic Solo
Advanced Auto Body Hilton Head
Wells Fargo Careers Log In
Bfri Forum
Uw Oshkosh Wrestling
Rub Md Okc
Rabbi Raps
Craigslist For Pets For Sale
Old Navy Student Discount Unidays
Crustless Pizza Bowl Pizza Hut
The Eye Doctors North Topeka
Temperature At 12 Pm Today
Pike County Buy Sale And Trade
Latest Posts
Article information

Author: Terrell Hackett

Last Updated:

Views: 6134

Rating: 4.1 / 5 (52 voted)

Reviews: 91% of readers found this page helpful

Author information

Name: Terrell Hackett

Birthday: 1992-03-17

Address: Suite 453 459 Gibson Squares, East Adriane, AK 71925-5692

Phone: +21811810803470

Job: Chief Representative

Hobby: Board games, Rock climbing, Ghost hunting, Origami, Kabaddi, Mushroom hunting, Gaming

Introduction: My name is Terrell Hackett, I am a gleaming, brainy, courageous, helpful, healthy, cooperative, graceful person who loves writing and wants to share my knowledge and understanding with you.