Brain-Computer Interface Lets Users Communicate Using Thoughts

By decoding neural signals linked to a person’s inner monologue, researchers enabled communication for individuals with severe speech impairments.

Written bySahana Sitaraman, PhD
| 3 min read
An abstract illustration that represents brain-computer interface, showing an arm and a robotic hand pointing at a human brain from opposite sides.
Register for free to listen to this article
Listen with Speechify
0:00
3:00
Share

There’s a voice inside most people’s minds that comes alive when they listen, read, or prepare to speak. This “internal monologue” is thought to support complex cognitive processes like working memory, logical reasoning, and motivation.1 In fact, inner speech continues to thrive in many individuals who are unable to speak owing to injury or disease.2 More than half a century ago, Jacques Vidal, a computer scientist at the University of California, Los Angeles, proposed the idea for brain-computer interfaces (BCIs); systems that could use electrical signals in the brain to control prosthetic devices.3

Since then, scientists have designed and developed BCIs that have enabled people with quadriplegia to control a computer cursor, a robotic arm, and even move their own limb. Recently, a person with amyotrophic lateral sclerosis (ALS)—a neurodegenerative disease—who had severe difficulties speaking, was able to carry out a freeform conversation with the help of a speech BCI.4 The neuroprosthesis accurately translated brain activity into coherent sentences while the person tried to speak to the best of their ability. However, the reliance on attempted speech can fatigue the user and limit communication speed.

To address this gap, Francis Willett, a neurosurgeon at Stanford University, and his team developed a speech BCI that could decipher imagined sentences.5 Their findings, published in Cell, enabled people with speech paralysis to audibly communicate by simply thinking about what they wanted to say in real time. “If you just have to think about speech instead of actually trying to speak, it’s potentially easier and faster for people,” said Benyamin Meschede-Krasa, a graduate student at Stanford University and coauthor of the study, in a statement.

The researchers recruited four participants enrolled in the BrainGate2 clinical trial, who had varying degrees of speech defects due to ALS or stroke and implanted devices to record neuronal signals from their motor cortex. Then, they asked the participants to either imagine speaking a set of English words or try to vocalize them. While both speech behaviors activated the same regions of the brain, the neuronal responses to inner speech were weaker than those produced by talking attempts. The team then trained a machine learning model on the neuronal signals linked to a library of imagined words to create the speech BCI. To test its performance, the researchers asked the study participants to imagine various cued sentences, all based on a vocabulary of 125,000 words, and observed that the neuroprosthesis could decode inner speech with an accuracy of 74 percent.

“This is the first time we’ve managed to understand what brain activity looks like when you just think about speaking,” said Erin Kunz, an electrical engineer at Stanford University and coauthor of the study. “For people with severe speech and motor impairments, BCIs capable of decoding inner speech could help them communicate much more easily and more naturally.”

A woman uses an inner speech neuroprosthetic device to communicate. The machine decodes sentences imagined by the study participant in real time and displays it on a screen.

A patient with speech paralysis imagines a cued sentence (above) in her mind, which is decoded by a brain-computer interface in real time (below).

Emory BrainGate Team

But the real test was to see if the BCI could translate uninstructed freeform verbal thoughts. For this, Willett and his team showed two participants a stair-shaped pattern that they had to replicate using a joystick. They asked them to remember the shape either as a set of directions, using inner speech (up, left, up), or by relying solely on their visual memory. The speech BCI could decode the former, but not the latter. In another task, when the individuals counted the number of times a certain colored shape appeared in a grid, the BCI could decode the internal monologue of numbers, albeit with significant mistakes.

Continue reading below...

Like this story? Sign up for FREE Newsletter updates:

Latest science news storiesTopic-tailored resources and eventsCustomized newsletter content
Subscribe

While the speech BCI offers substantial benefits to users, it also raises the possibility of unintentional decoding of private verbal thoughts. To prevent this, the team trained the system to detect the password “chittychittybangbang” with 98 percent accuracy. Users could unlock the system by simply vocalizing the phrase internally, followed by what they wished to say.

While Willette and his team acknowledged that the current system is not ready to accurately decode spontaneous thinking, improvements to recording technology could eventually make this a reality.

“The future of BCIs is bright,” Willett said. “This work gives real hope that speech BCIs can one day restore communication that is as fluent, natural, and comfortable as conversational speech.”

  1. Alderson-Day B, Fernyhough C. Inner speech: Development, cognitive functions, phenomenology, and neurobiology. Psychol Bull. 2015;141(5):931-965.
  2. Hurlburt RT, et al. Toward a phenomenology of inner speaking. Conscious Cogn. 2013;22(4):1477-1494.
  3. Vidal JJ. Toward direct brain-computer communication. Ann Rev Biophys. 1973;2:157-180.
  4. Card NS, et al. An accurate and rapidly calibrating speech neuroprosthesis. N Engl J Med. 2024;391(7):609-618.
  5. Kunz EM, et al. Inner speech in motor cortex and implications for speech neuroprostheses. Cell. 2025.

Related Topics

Meet the Author

  • Photograph of Sahana Sitaraman. The photograph is in grayscale. Sahana has short, curly hair, round-framed glasses, and is wearing a windbreaker jacket.

    Sahana is an Assistant Editor at The Scientist, where she crafts stories that bring the wonders and oddities of science to life. In 2022, she earned a PhD in neuroscience from the National Centre for Biological Sciences, India, studying how neurons develop their stereotypical tree-like shapes. In a parallel universe, Sahana is a passionate singer and an enthusiastic hiker.

    View Full Profile
Share
You might also be interested in...
Loading Next Article...
You might also be interested in...
Loading Next Article...
December digest cover image of a wooden sculpture comprised of multiple wooden neurons that form a seahorse.
December 2025, Issue 1

Wooden Neurons: An Artistic Vision of the Brain

A neurobiologist, who loves the morphology of cells, turns these shapes into works of art made from wood.

View this Issue
Stacks of cell culture dishes, plates, and flasks with pink cell culture medium on a white background.

Driving Innovation with Cell Culture Essentials

Merck
Stacks of cell culture dishes, plates, and flasks with pink cell culture medium on a white background.

Driving Innovation with Cell Culture Essentials

MilliporeSigma purple logo
Human iPSC-derived Models for Brain Disease Research

Human iPSC-derived Models for Neurodegenerative Disease Research

Fujifilm
Abstract wireframe sphere with colorful dots and connecting lines representing the complex cellular and molecular interactions within the tumor microenvironment.

Exploring the Inflammatory Tumor Microenvironment 

Cellecta logo

Products

Beckman Logo

Beckman Coulter Life Sciences Introduces the Biomek i3 Benchtop Liquid Handler, a Small but Mighty Addition to its Portfolio of Automated Workstations

brandtech logo

BRANDTECH® Scientific Announces Strategic Partnership with Copia Scientific to Strengthen Sales and Service of the BRAND® Liquid Handling Station (LHS) 

Top Innovations 2026 Contest Image

Enter Our 2026 Top Innovations Contest

Biotium Logo

Biotium Expands Tyramide Signal Amplification Portfolio with Brighter and More Stable Dyes for Enhanced Spatial Imaging