Breaking Silence: Giving the Silent a Voice Through Thoughts – Neuroscience News

by thinkia.org.in
0 comment


Summary: Researchers enabled a silent person to produce speech using thought alone. Depth electrodes in the participant’s brain transmitted electrical signals to a computer, which then vocalized imagined syllables.

This technology offers hope for paralyzed individuals to regain speech. The study marks a significant step towards brain-computer interfaces for voluntary communication.

Key Facts:

  1. Technology: Depth electrodes transmit brain signals to a computer for speech.
  2. Participants: Experiment involved an epilepsy patient with implanted electrodes.
  3. Future Impact: Could enable paralyzed individuals to communicate through thought.

Source: Tel Aviv University

A scientific breakthrough by researchers from Tel Aviv University and Tel Aviv Sourasky Medical Center (Ichilov Hospital) has demonstrated the potential for speech by a silent person using the power of thought only.

In an experiment, a silent participant imagined saying one of two syllables. Depth electrodes implanted in his brain transmitted the electrical signals to a computer, which then vocalized the syllables.

The study was led by Dr. Ariel Tankus of Tel Aviv University’s School of Medical and Health Sciences and Tel Aviv Sourasky Medical Center (Ichilov Hospital), along with Dr. Ido Strauss of Tel Aviv University’s School of Medical and Health Sciences and director of the Functional Neurosurgery Unit at Ichilov Hospital.

The results of this study were published in the journal Neurosurgery.

These findings offer hope for enabling people who are completely paralyzed—due to conditions such as ALS, brainstem stroke, or brain injury—to regain the ability to speak voluntarily.

“The patient in the study is an epilepsy patient who was hospitalized in order to undergo resection of the epileptic focus in his brain,” explains Dr. Tankus. “In order to do this, of course, you need to locate the focal point, which is the source of the ‘short’ that sends powerful electrical waves through the brain.

“This situation pertains to a smaller subset of epilepsy patients who do not respond well to medication and require neurosurgical intervention, and an even smaller subset of epilepsy patients whose suspected focus is located deep within the brain, rather than on the surface of the cortex.

“To identify the exact location, electrodes have to be implanted into deep structures of their brains. They are then hospitalized, awaiting the next seizure.

“When a seizure occurs, the electrodes will tell the neurologists and neurosurgeons where the focus is, allowing them to operate precisely. From a scientific perspective, this provides a rare opportunity to get a glimpse into the depths a living human brain.

“Fortunately, the epilepsy patient hospitalized at Ichilov agreed to participate in the experiment, which may ultimately help completely paralyzed individuals to express themselves again through artificial speech.”

In the first stage of the experiment, with the depth electrodes already implanted in the patient’s brain, the Tel Aviv University researchers asked him to say two syllables out loud: /a/ and /e/.

They recorded the brain activity as he articulated these sounds. Using deep learning and machine learning, the researchers trained artificial intelligence models to identify the specific brain cells whose electrical activity indicated the desire to say /a/ or /e/.

Once the computer learned to recognize the pattern of electrical activity associated with these two syllables in the patient’s brain, he was asked to only imagine that he was saying /a/ and /e/. The computer then translated the electrical signals and played the pre-recorded sounds of /a/ or /e/ accordingly.

“My field of research deals with the encoding and decoding of speech, that is, how individual brain cells participate in the speech process—the production of speech, the hearing of speech, and the imagination of speech, or ‘speaking silently,’” says Dr. Tankus.

“In this experiment, for the first time in history, we were able to connect the parts of speech to the activity of individual cells from the regions of the brain from which we recorded.

“This allowed us to distinguish between the electrical signals that represent the sounds /a/ and /e/. At the moment, our research involves two building blocks of speech, two syllables.

“Of course, our ambition is to get to complete speech, but even two different syllables can enable a fully paralyzed person to signal ‘yes’ and ‘no.’ For example, in the future it will be possible to train a computer for an ALS patient in the early stages of the disease, while they can still speak.

“The computer would learn to recognize the electrical signals in the patient’s brain, enabling it to interpret these signals even after the patient loses the ability to move their muscles. And that is just one example.

“Our study is a significant step toward developing a brain-computer interface that can replace the brain’s control pathways for speech production, allowing completely paralyzed individuals to communicate voluntarily with their surroundings once again.”

About this BCI and neurotech research news

Author: Ariel Tankus
Source: Tel Aviv University
Contact: Ariel Tankus – Tel Aviv University
Image: The image is credited to Neuroscience News

Original Research: Closed access.
A Speech Neuroprosthesis in the Frontal Lobe and Hippocampus: Decoding High-Frequency Activity into Phonemes” by Ariel Tankus et al. Neurosurgery


Abstract

A Speech Neuroprosthesis in the Frontal Lobe and Hippocampus: Decoding High-Frequency Activity into Phonemes

BACKGROUND AND OBJECTIVES: 

Loss of speech due to injury or disease is devastating. Here, we report a novel speech neuroprosthesis that artificially articulates building blocks of speech based on high-frequency activity in brain areas never harnessed for a neuroprosthesis before: anterior cingulate and orbitofrontal cortices, and hippocampus.

METHODS: 

A 37-year-old male neurosurgical epilepsy patient with intact speech, implanted with depth electrodes for clinical reasons only, silently controlled the neuroprosthesis almost immediately and in a natural way to voluntarily produce 2 vowel sounds.

RESULTS: 

During the first set of trials, the participant made the neuroprosthesis produce the different vowel sounds artificially with 85% accuracy. In the following trials, performance improved consistently, which may be attributed to neuroplasticity. We show that a neuroprosthesis trained on overt speech data may be controlled silently.

CONCLUSION: 

This may open the way for a novel strategy of neuroprosthesis implantation at earlier disease stages (eg, amyotrophic lateral sclerosis), while speech is intact, for improved training that still allows silent control at later stages. The results demonstrate clinical feasibility of direct decoding of high-frequency activity that includes spiking activity in the aforementioned areas for silent production of phonemes that may serve as a part of a neuroprosthesis for replacing lost speech control pathways.

You may also like

Thinkia is a professional platform where we provide informative content like current world news, all types of educational content, health awareness, food awareness, travel awareness, ideas and tips. We hope you like all the content provided by us.

Editors' Picks

Latest Posts

Copyright © 2024 | Thinkia | All Right Reserved