1
|
Ma J, Rui Z, Zou Y, Qin Z, Zhao Z, Zhang Y, Mao Z, Bai H, Zhang J. Neurosurgical and BCI approaches to visual rehabilitation in occipital lobe tumor patients. Heliyon 2024; 10:e39072. [PMID: 39687114 PMCID: PMC11647799 DOI: 10.1016/j.heliyon.2024.e39072] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2024] [Revised: 10/03/2024] [Accepted: 10/07/2024] [Indexed: 12/18/2024] Open
Abstract
This study investigates the effects of occipital lobe tumors on visual processing and the role of brain-computer interface (BCI) technologies in post-surgical visual rehabilitation. Through a combination of pre-surgical functional magnetic resonance imaging (fMRI) and Diffusion Tensor Imaging (DTI), intra-operative direct cortical stimulation (DCS) and Electrocorticography (ECoG), and post-surgical BCI interventions, we provide insight into the complex dynamics between occipital lobe tumors and visual function. Our results highlight a discrepancy between clinical assessments of visual field damage and the patient's reported visual experiences, suggesting a residual functional capacity within the damaged occipital regions. Additionally, the absence of expected visual phenomena during surgery and the promising outcomes from BCI-driven rehabilitation underscore the complexity of visual processing and the potential of technology-enhanced rehabilitation strategies. This work emphasizes the need for an interdisciplinary approach in developing effective treatments for visual impairments related to brain tumors, illustrating the significant implications for neurosurgical practices and the advancement of rehabilitation sciences.
Collapse
Affiliation(s)
- Jie Ma
- PLA Medical School, Beijing, 100853, PR China
- Department of Neurosurgery, Chinese PLA General Hospital, Beijing, 100853, PR China
| | - Zong Rui
- Department of Neurosurgery, Chinese PLA General Hospital, Beijing, 100853, PR China
| | - Yuhui Zou
- Department of Neurosurgery, General Hospital of the Southern Theater Command of PLA, Guangzhou, Guangzhou, 510051, PR China
| | - Zhizhen Qin
- PLA Medical School, Beijing, 100853, PR China
- Department of Neurosurgery, Chinese PLA General Hospital, Beijing, 100853, PR China
| | - Zhenyu Zhao
- Department of Neurosurgery, General Hospital of the Southern Theater Command of PLA, Guangzhou, Guangzhou, 510051, PR China
| | - Yanyang Zhang
- Department of Neurosurgery, Chinese PLA General Hospital, Beijing, 100853, PR China
| | - Zhiqi Mao
- Department of Neurosurgery, Chinese PLA General Hospital, Beijing, 100853, PR China
| | - Hongmin Bai
- Department of Neurosurgery, General Hospital of the Southern Theater Command of PLA, Guangzhou, Guangzhou, 510051, PR China
| | - Jianning Zhang
- Department of Neurosurgery, Chinese PLA General Hospital, Beijing, 100853, PR China
| |
Collapse
|
2
|
Hoyle AC, Stevenson R, Leonhardt M, Gillett T, Martinez-Hernandez U, Gompertz N, Clarke C, Cazzola D, Metcalfe BW. Exploring the 'EarSwitch' concept: a novel ear based control method for assistive technology. J Neuroeng Rehabil 2024; 21:210. [PMID: 39623474 PMCID: PMC11613744 DOI: 10.1186/s12984-024-01500-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2024] [Accepted: 10/30/2024] [Indexed: 12/06/2024] Open
Abstract
BACKGROUND Loss of communication with loved ones and carers is one of the most isolating and debilitating effects of many neurological disorders. Assistive technology (AT) supports individuals with communication, but the acceptability of AT solutions is highly variable. In this paper a novel ear based control method of AT, the concept of 'EarSwitch', is presented. This new approach is based on detecting ear rumbling, which is the voluntary contraction of the tensor tympani muscle (TTM), resulting in observable movement of the eardrum and a dull rumbling sound. 'EarSwitch' has the potential to be a discreet method that can complement existing AT control methods. However, only a subset of the population can ear rumble and little is known about the ability of rumbling in populations with neurological disorders. METHODS To explore the viability of the 'EarSwitch' concept as an AT control method we conducted in-depth online surveys with (N=1853) respondents from the general population and (N=170) respondents with self-declared neurological disorders including Motor Neurone Disease (MND) and Multiple Sclerosis (MS).This is the largest ever study to explore ear rumbling and the first to explore whether rumbling is preserved among individuals with neurological disorders. In addition, we validated rumbling, and investigated usability of the 'EarSwitch' concept as a control input, using in-person otoscopic examination with a subset of participants. RESULTS A significant proportion of the population with neurological disorders could benefit from 'EarSwitch' controllable AT. The upper bound prevalence of the ability to rumble without accompanying movements was 55% in the general population, 38% in the neurological population, and 20% of participants with MND (N=95) reported this ability. During the validation procedure, participants achieved high accuracy in self-reporting the ability to rumble (80%) and proved concept of using the 'EarSwitch' method to control a basic interface. DISCUSSION 'EarSwitch' is a potential new AT control method control, either by itself or as a supplement to other existing methods. Results demonstrate self-reported ear rumbling is present among patients with different neurological disorders, including MND. Further research should explore how well the ability to rumble is preserved in different types and stages of neurological disorders.
Collapse
Affiliation(s)
- Anna C Hoyle
- Department of Electronic and Electrical Engineering, University of Bath, Bath, UK
- Bath Institute for the Augmented Human (IAH), University of Bath, Bath, UK
| | | | - Martin Leonhardt
- Department of Electronic and Electrical Engineering, University of Bath, Bath, UK
| | - Thomas Gillett
- School of Engineering and Physical Science, Heriot-Watt University, Edinburgh, UK
| | - Uriel Martinez-Hernandez
- Department of Electronic and Electrical Engineering, University of Bath, Bath, UK
- Bath Institute for the Augmented Human (IAH), University of Bath, Bath, UK
| | | | - Christopher Clarke
- Department of Computer Science, University of Bath, Bath, UK
- Centre for Analysis of Motion and Entertainment Research and Application (CAMERA), University of Bath, Bath, UK
- Bath Institute for the Augmented Human (IAH), University of Bath, Bath, UK
| | - Dario Cazzola
- Department for Health, University of Bath, Bath, UK
- Centre for Analysis of Motion and Entertainment Research and Application (CAMERA), University of Bath, Bath, UK
- Bath Institute for the Augmented Human (IAH), University of Bath, Bath, UK
| | - Benjamin W Metcalfe
- Department of Electronic and Electrical Engineering, University of Bath, Bath, UK.
- Bath Institute for the Augmented Human (IAH), University of Bath, Bath, UK.
| |
Collapse
|
3
|
Rahman N, Khan DM, Masroor K, Arshad M, Rafiq A, Fahim SM. Advances in brain-computer interface for decoding speech imagery from EEG signals: a systematic review. Cogn Neurodyn 2024; 18:3565-3583. [PMID: 39712121 PMCID: PMC11655741 DOI: 10.1007/s11571-024-10167-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Revised: 07/16/2024] [Accepted: 08/17/2024] [Indexed: 12/24/2024] Open
Abstract
Numerous individuals encounter challenges in verbal communication due to various factors, including physical disabilities, neurological disorders, and strokes. In response to this pressing need, technology has actively pursued solutions to bridge the communication gap, recognizing the inherent difficulties faced in verbal communication, particularly in contexts where traditional methods may be inadequate. Electroencephalogram (EEG) has emerged as a primary non-invasive method for measuring brain activity, offering valuable insights from a cognitive neurodevelopmental perspective. It forms the basis for Brain-Computer Interfaces (BCIs) that provide a communication channel for individuals with neurological impairments, thereby empowering them to express themselves effectively. EEG-based BCIs, especially those adapted to decode imagined speech from EEG signals, represent a significant advancement in enabling individuals with speech disabilities to communicate through text or synthesized speech. By utilizing cognitive neurodevelopmental insights, researchers have been able to develop innovative approaches for interpreting EEG signals and translating them into meaningful communication outputs. To aid researchers in effectively addressing this complex challenge, this review article synthesizes key findings from state-of-the-art significant studies. It investigates into the methodologies employed by various researchers, including preprocessing techniques, feature extraction methods, and classification algorithms utilizing Deep Learning and Machine Learning approaches and their integration. Furthermore, the review outlines the potential avenues for future research, with the goal of advancing the practical implementation of EEG-based BCI systems for decoding imagined speech from a cognitive neurodevelopmental perspective.
Collapse
Affiliation(s)
- Nimra Rahman
- Department of Telecommunications Engineering, NED University of Engineering and Technology, Karachi, Sindh 75270 Pakistan
| | - Danish Mahmood Khan
- Department of Electronic Engineering, NED University of Engineering and Technology, Karachi, Sindh 75270 Pakistan
- Department of Computing and Information Systems, School of Engineering and Technology, Sunway University, Selangor 47500 Petaling Jaya, Malaysia
| | - Komal Masroor
- Department of Telecommunications Engineering, NED University of Engineering and Technology, Karachi, Sindh 75270 Pakistan
| | - Mehak Arshad
- Department of Telecommunications Engineering, NED University of Engineering and Technology, Karachi, Sindh 75270 Pakistan
| | - Amna Rafiq
- Department of Telecommunications Engineering, NED University of Engineering and Technology, Karachi, Sindh 75270 Pakistan
| | - Syeda Maham Fahim
- Department of Telecommunications Engineering, NED University of Engineering and Technology, Karachi, Sindh 75270 Pakistan
| |
Collapse
|
4
|
Singer-Clark T, Hou X, Card NS, Wairagkar M, Iacobacci C, Peracha H, Hochberg LR, Stavisky SD, Brandman DM. Speech motor cortex enables BCI cursor control and click. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.11.12.623096. [PMID: 39605556 PMCID: PMC11601350 DOI: 10.1101/2024.11.12.623096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 11/29/2024]
Abstract
Decoding neural activity from ventral (speech) motor cortex is known to enable high-performance speech brain-computer interface (BCI) control. It was previously unknown whether this brain area could also enable computer control via neural cursor and click, as is typically associated with dorsal (arm and hand) motor cortex. We recruited a clinical trial participant with ALS and implanted intracortical microelectrode arrays in ventral precentral gyrus (vPCG), which the participant used to operate a speech BCI in a prior study. We developed a cursor BCI driven by the participant's vPCG neural activity, and evaluated performance on a series of target selection tasks. The reported vPCG cursor BCI enabled rapidly-calibrating (40 seconds), accurate (2.90 bits per second) cursor control and click. The participant also used the BCI to control his own personal computer independently. These results suggest that placing electrodes in vPCG to optimize for speech decoding may also be a viable strategy for building a multi-modal BCI which enables both speech-based communication and computer control via cursor and click.
Collapse
Affiliation(s)
- Tyler Singer-Clark
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
- Department of Biomedical Engineering, University of California Davis, Davis, CA, USA
| | - Xianda Hou
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
- Department of Computer Science, University of California Davis, Davis, CA, USA
| | - Nicholas S. Card
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Maitreyee Wairagkar
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Carrina Iacobacci
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Hamza Peracha
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Leigh R. Hochberg
- School of Engineering and Carney Institute for Brain Sciences, Brown University, Providence, RI, USA
- VA RR&D Center for Neurorestoration and Neurotechnology, VA Providence Healthcare, Providence, RI
- Center for Neurotechnology and Neurorecovery, Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA
| | - Sergey D. Stavisky
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - David M. Brandman
- Department of Neurological Surgery, University of California Davis, Davis, CA, USA
| |
Collapse
|
5
|
Rybář M, Poli R, Daly I. Using data from cue presentations results in grossly overestimating semantic BCI performance. Sci Rep 2024; 14:28003. [PMID: 39543314 PMCID: PMC11564751 DOI: 10.1038/s41598-024-79309-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2024] [Accepted: 11/07/2024] [Indexed: 11/17/2024] Open
Abstract
Neuroimaging studies have reported the possibility of semantic neural decoding to identify specific semantic concepts from neural activity. This offers promise for brain-computer interfaces (BCIs) for communication. However, translating these findings into a BCI paradigm has proven challenging. Existing EEG-based semantic decoding studies often rely on neural activity recorded when a cue is present, raising concerns about decoding reliability. To address this, we investigate the effects of cue presentation on EEG-based semantic decoding. In an experiment with a clear separation between cue presentation and mental task periods, we attempt to differentiate between semantic categories of animals and tools in four mental tasks. By using state-of-the-art decoding analyses, we demonstrate significant mean classification accuracies up to 71.3% during cue presentation but not during mental tasks, even with adapted analyses from previous studies. These findings highlight a potential issue when using neural activity recorded during cue presentation periods for semantic decoding. Additionally, our results show that semantic decoding without external cues may be more challenging than current state-of-the-art research suggests. By bringing attention to these issues, we aim to stimulate discussion and drive advancements in the field toward more effective semantic BCI applications.
Collapse
Affiliation(s)
- Milan Rybář
- Brain-Computer Interfaces and Neural Engineering Laboratory, School of Computer Science and Electronic Engineering, University of Essex, Colchester, CO4 3SQ, UK.
| | - Riccardo Poli
- Brain-Computer Interfaces and Neural Engineering Laboratory, School of Computer Science and Electronic Engineering, University of Essex, Colchester, CO4 3SQ, UK
| | - Ian Daly
- Brain-Computer Interfaces and Neural Engineering Laboratory, School of Computer Science and Electronic Engineering, University of Essex, Colchester, CO4 3SQ, UK.
| |
Collapse
|
6
|
Xu S, Liu Y, Lee H, Li W. Neural interfaces: Bridging the brain to the world beyond healthcare. EXPLORATION (BEIJING, CHINA) 2024; 4:20230146. [PMID: 39439491 PMCID: PMC11491314 DOI: 10.1002/exp.20230146] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 02/02/2024] [Indexed: 10/25/2024]
Abstract
Neural interfaces, emerging at the intersection of neurotechnology and urban planning, promise to transform how we interact with our surroundings and communicate. By recording and decoding neural signals, these interfaces facilitate direct connections between the brain and external devices, enabling seamless information exchange and shared experiences. Nevertheless, their development is challenged by complexities in materials science, electrochemistry, and algorithmic design. Electrophysiological crosstalk and the mismatch between electrode rigidity and tissue flexibility further complicate signal fidelity and biocompatibility. Recent closed-loop brain-computer interfaces, while promising for mood regulation and cognitive enhancement, are limited by decoding accuracy and the adaptability of user interfaces. This perspective outlines these challenges and discusses the progress in neural interfaces, contrasting non-invasive and invasive approaches, and explores the dynamics between stimulation and direct interfacing. Emphasis is placed on applications beyond healthcare, highlighting the need for implantable interfaces with high-resolution recording and stimulation capabilities.
Collapse
Affiliation(s)
- Shumao Xu
- Department of Biomedical EngineeringThe Pennsylvania State UniversityPennsylvaniaUSA
| | - Yang Liu
- Brain Health and Brain Technology Center at Global Institute of Future TechnologyShanghai Jiao Tong UniversityShanghaiChina
| | - Hyunjin Lee
- Department of Biomedical EngineeringThe Pennsylvania State UniversityPennsylvaniaUSA
| | - Weidong Li
- Brain Health and Brain Technology Center at Global Institute of Future TechnologyShanghai Jiao Tong UniversityShanghaiChina
| |
Collapse
|
7
|
Angrick M, Luo S, Rabbani Q, Joshi S, Candrea DN, Milsap GW, Gordon CR, Rosenblatt K, Clawson L, Maragakis N, Tenore FV, Fifer MS, Ramsey NF, Crone NE. Real-time detection of spoken speech from unlabeled ECoG signals: A pilot study with an ALS participant. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2024:2024.09.18.24313755. [PMID: 39371161 PMCID: PMC11451764 DOI: 10.1101/2024.09.18.24313755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Indexed: 10/08/2024]
Abstract
Objective Brain-Computer Interfaces (BCIs) hold significant promise for restoring communication in individuals with partial or complete loss of the ability to speak due to paralysis from amyotrophic lateral sclerosis (ALS), brainstem stroke, and other neurological disorders. Many of the approaches to speech decoding reported in the BCI literature have required time-aligned target representations to allow successful training - a major challenge when translating such approaches to people who have already lost their voice. Approach In this pilot study, we made a first step toward scenarios in which no ground truth is available. We utilized a graph-based clustering approach to identify temporal segments of speech production from electrocorticographic (ECoG) signals alone. We then used the estimated speech segments to train a voice activity detection (VAD) model using only ECoG signals. We evaluated our approach using held-out open-loop recordings of a single dysarthric clinical trial participant living with ALS, and we compared the resulting performance to previous solutions trained with ground truth acoustic voice recordings. Main results Our approach achieves a median error rate of around 0.5 seconds with respect to the actual spoken speech. Embedded into a real-time BCI, our approach is capable of providing VAD results with a latency of only 10 ms. Significance To the best of our knowledge, our results show for the first time that speech activity can be predicted purely from unlabeled ECoG signals, a crucial step toward individuals who cannot provide this information anymore due to their neurological condition, such as patients with locked-in syndrome. Clinical Trial Information ClinicalTrials.gov, registration number NCT03567213.
Collapse
Affiliation(s)
- Miguel Angrick
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Shiyu Luo
- Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Qinwan Rabbani
- Department of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD, USA
| | - Shreya Joshi
- Department of Computer Science, The Johns Hopkins University, Baltimore, MD, USA
- Department of Cognitive Science, The Johns Hopkins University, Baltimore, MD, USA
| | - Daniel N Candrea
- Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Griffin W Milsap
- Research and Exploratory Development Department, Johns Hopkins Applied Physics Laboratory, Laurel, MD, USA
| | - Chad R Gordon
- Departments of Plastic and Reconstructive Surgery & Neurosurgery, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Kathryn Rosenblatt
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
- Department of Anesthesiology & Critical Care Medicine, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Lora Clawson
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Nicholas Maragakis
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Francesco V Tenore
- Research and Exploratory Development Department, Johns Hopkins Applied Physics Laboratory, Laurel, MD, USA
| | - Matthew S Fifer
- Research and Exploratory Development Department, Johns Hopkins Applied Physics Laboratory, Laurel, MD, USA
| | - Nick F Ramsey
- UMC Utrecht Brain Center, Department of Neurology and Neurosurgery, University Medical Center Utrecht, Utrecht, The Netherlands
| | - Nathan E Crone
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| |
Collapse
|
8
|
Zhang H, Jiao L, Yang S, Li H, Jiang X, Feng J, Zou S, Xu Q, Gu J, Wang X, Wei B. Brain-computer interfaces: the innovative key to unlocking neurological conditions. Int J Surg 2024; 110:5745-5762. [PMID: 39166947 PMCID: PMC11392146 DOI: 10.1097/js9.0000000000002022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2024] [Accepted: 07/30/2024] [Indexed: 08/23/2024]
Abstract
Neurological disorders such as Parkinson's disease, stroke, and spinal cord injury can pose significant threats to human mortality, morbidity, and functional independence. Brain-Computer Interface (BCI) technology, which facilitates direct communication between the brain and external devices, emerges as an innovative key to unlocking neurological conditions, demonstrating significant promise in this context. This comprehensive review uniquely synthesizes the latest advancements in BCI research across multiple neurological disorders, offering an interdisciplinary perspective on both clinical applications and emerging technologies. We explore the progress in BCI research and its applications in addressing various neurological conditions, with a particular focus on recent clinical studies and prospective developments. Initially, the review provides an up-to-date overview of BCI technology, encompassing its classification, operational principles, and prevalent paradigms. It then critically examines specific BCI applications in movement disorders, disorders of consciousness, cognitive and mental disorders, as well as sensory disorders, highlighting novel approaches and their potential impact on patient care. This review reveals emerging trends in BCI applications, such as the integration of artificial intelligence and the development of closed-loop systems, which represent significant advancements over previous technologies. The review concludes by discussing the prospects and directions of BCI technology, underscoring the need for interdisciplinary collaboration and ethical considerations. It emphasizes the importance of prioritizing bidirectional and high-performance BCIs, areas that have been underexplored in previous reviews. Additionally, we identify crucial gaps in current research, particularly in long-term clinical efficacy and the need for standardized protocols. The role of neurosurgery in spearheading the clinical translation of BCI research is highlighted. Our comprehensive analysis presents BCI technology as an innovative key to unlocking neurological disorders, offering a transformative approach to diagnosing, treating, and rehabilitating neurological conditions, with substantial potential to enhance patients' quality of life and advance the field of neurotechnology.
Collapse
Affiliation(s)
- Hongyu Zhang
- Department of Neurosurgery, The Fourth Affiliated Hospital of Harbin Medical University
- Harbin Medical University, Harbin
| | - Le Jiao
- Department of Neurosurgery, The First Hospital of Qiqihar, Qiqihar, Heilongjiang Province
| | | | | | | | - Jing Feng
- Department of Neurosurgery, The Fourth Affiliated Hospital of Harbin Medical University
- Harbin Medical University, Harbin
| | - Shuhuai Zou
- Department of Neurosurgery, The Fourth Affiliated Hospital of Harbin Medical University
- Harbin Medical University, Harbin
| | - Qiang Xu
- Department of Neurosurgery, The Fourth Affiliated Hospital of Harbin Medical University
- Harbin Medical University, Harbin
| | - Jianheng Gu
- Department of Neurosurgery, The Fourth Affiliated Hospital of Harbin Medical University
- Harbin Medical University, Harbin
| | - Xuefeng Wang
- Department of Neurosurgery, The Fourth Affiliated Hospital of Harbin Medical University
| | - Baojian Wei
- School of Nursing, Shandong First Medical University & Shandong Academy of Medical Sciences, Taian, Shandong, People's Republic of China
| |
Collapse
|
9
|
Rabbani Q, Shah S, Milsap G, Fifer M, Hermansky H, Crone N. Iterative alignment discovery of speech-associated neural activity. J Neural Eng 2024; 21:046056. [PMID: 39194182 DOI: 10.1088/1741-2552/ad663c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Accepted: 07/22/2024] [Indexed: 08/29/2024]
Abstract
Objective. Brain-computer interfaces (BCIs) have the potential to preserve or restore speech in patients with neurological disorders that weaken the muscles involved in speech production. However, successful training of low-latency speech synthesis and recognition models requires alignment of neural activity with intended phonetic or acoustic output with high temporal precision. This is particularly challenging in patients who cannot produce audible speech, as ground truth with which to pinpoint neural activity synchronized with speech is not available.Approach. In this study, we present a new iterative algorithm for neural voice activity detection (nVAD) called iterative alignment discovery dynamic time warping (IAD-DTW) that integrates DTW into the loss function of a deep neural network (DNN). The algorithm is designed to discover the alignment between a patient's electrocorticographic (ECoG) neural responses and their attempts to speak during collection of data for training BCI decoders for speech synthesis and recognition.Main results. To demonstrate the effectiveness of the algorithm, we tested its accuracy in predicting the onset and duration of acoustic signals produced by able-bodied patients with intact speech undergoing short-term diagnostic ECoG recordings for epilepsy surgery. We simulated a lack of ground truth by randomly perturbing the temporal correspondence between neural activity and an initial single estimate for all speech onsets and durations. We examined the model's ability to overcome these perturbations to estimate ground truth. IAD-DTW showed no notable degradation (<1% absolute decrease in accuracy) in performance in these simulations, even in the case of maximal misalignments between speech and silence.Significance. IAD-DTW is computationally inexpensive and can be easily integrated into existing DNN-based nVAD approaches, as it pertains only to the final loss computation. This approach makes it possible to train speech BCI algorithms using ECoG data from patients who are unable to produce audible speech, including those with Locked-In Syndrome.
Collapse
Affiliation(s)
- Qinwan Rabbani
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218, United States of America
| | - Samyak Shah
- Department of Neurology, Johns Hopkins Medicine, Baltimore, MD 21287, United States of America
| | - Griffin Milsap
- Research and Exploratory Development Department, Johns Hopkins University Applied Physics Laboratory, Laurel, MD 20723, United States of America
| | - Matthew Fifer
- Research and Exploratory Development Department, Johns Hopkins University Applied Physics Laboratory, Laurel, MD 20723, United States of America
| | - Hynek Hermansky
- Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD 21218, United States of America
| | - Nathan Crone
- Department of Neurology, Johns Hopkins Medicine, Baltimore, MD 21287, United States of America
| |
Collapse
|
10
|
Card NS, Wairagkar M, Iacobacci C, Hou X, Singer-Clark T, Willett FR, Kunz EM, Fan C, Nia MV, Deo DR, Srinivasan A, Choi EY, Glasser MF, Hochberg LR, Henderson JM, Shahlaie K, Stavisky SD, Brandman DM. An Accurate and Rapidly Calibrating Speech Neuroprosthesis. N Engl J Med 2024; 391:609-618. [PMID: 39141853 PMCID: PMC11328962 DOI: 10.1056/nejmoa2314132] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 08/16/2024]
Abstract
BACKGROUND Brain-computer interfaces can enable communication for people with paralysis by transforming cortical activity associated with attempted speech into text on a computer screen. Communication with brain-computer interfaces has been restricted by extensive training requirements and limited accuracy. METHODS A 45-year-old man with amyotrophic lateral sclerosis (ALS) with tetraparesis and severe dysarthria underwent surgical implantation of four microelectrode arrays into his left ventral precentral gyrus 5 years after the onset of the illness; these arrays recorded neural activity from 256 intracortical electrodes. We report the results of decoding his cortical neural activity as he attempted to speak in both prompted and unstructured conversational contexts. Decoded words were displayed on a screen and then vocalized with the use of text-to-speech software designed to sound like his pre-ALS voice. RESULTS On the first day of use (25 days after surgery), the neuroprosthesis achieved 99.6% accuracy with a 50-word vocabulary. Calibration of the neuroprosthesis required 30 minutes of cortical recordings while the participant attempted to speak, followed by subsequent processing. On the second day, after 1.4 additional hours of system training, the neuroprosthesis achieved 90.2% accuracy using a 125,000-word vocabulary. With further training data, the neuroprosthesis sustained 97.5% accuracy over a period of 8.4 months after surgical implantation, and the participant used it to communicate in self-paced conversations at a rate of approximately 32 words per minute for more than 248 cumulative hours. CONCLUSIONS In a person with ALS and severe dysarthria, an intracortical speech neuroprosthesis reached a level of performance suitable to restore conversational communication after brief training. (Funded by the Office of the Assistant Secretary of Defense for Health Affairs and others; BrainGate2 ClinicalTrials.gov number, NCT00912041.).
Collapse
Affiliation(s)
- Nicholas S. Card
- Department of Neurological Surgery, University of
California Davis, Davis, CA
| | - Maitreyee Wairagkar
- Department of Neurological Surgery, University of
California Davis, Davis, CA
| | - Carrina Iacobacci
- Department of Neurological Surgery, University of
California Davis, Davis, CA
| | - Xianda Hou
- Department of Neurological Surgery, University of
California Davis, Davis, CA
- Department of Computer Science, University of California
Davis, Davis, CA
| | - Tyler Singer-Clark
- Department of Neurological Surgery, University of
California Davis, Davis, CA
- Department of Biomedical Engineering, University of
California Davis, Davis, CA
| | - Francis R. Willett
- Department of Neurosurgery, Stanford University, Stanford,
CA
- Howard Hughes Medical Institute, Stanford University,
Stanford, CA
| | - Erin M. Kunz
- Department of Electrical Engineering, Stanford University,
Stanford, CA
- Wu Tsai Neurosciences Institute, Stanford University,
Stanford, CA
| | - Chaofei Fan
- Department of Computer Science, Stanford University,
Stanford, CA
| | - Maryam Vahdati Nia
- Department of Neurological Surgery, University of
California Davis, Davis, CA
- Department of Computer Science, University of California
Davis, Davis, CA
| | - Darrel R. Deo
- Department of Neurosurgery, Stanford University, Stanford,
CA
| | - Aparna Srinivasan
- Department of Neurological Surgery, University of
California Davis, Davis, CA
- Department of Biomedical Engineering, University of
California Davis, Davis, CA
| | - Eun Young Choi
- Department of Neurosurgery, Stanford University, Stanford,
CA
| | - Matthew F. Glasser
- Departments of Radiology and Neuroscience, Washington
University School of Medicine, Saint Louis, MO
| | - Leigh R Hochberg
- School of Engineering and Carney Institute for Brain
Sciences, Brown University, Providence, RI
- VA RR&D Center for Neurorestoration and
Neurotechnology, VA Providence Healthcare, Providence, RI
- Center for Neurotechnology and Neurorecovery, Department
of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston,
MA
| | - Jaimie M. Henderson
- Department of Neurosurgery, Stanford University, Stanford,
CA
- Wu Tsai Neurosciences Institute, Stanford University,
Stanford, CA
| | - Kiarash Shahlaie
- Department of Neurological Surgery, University of
California Davis, Davis, CA
| | - Sergey D. Stavisky
- Department of Neurological Surgery, University of
California Davis, Davis, CA
| | - David M. Brandman
- Department of Neurological Surgery, University of
California Davis, Davis, CA
| |
Collapse
|
11
|
Silva AB, Liu JR, Metzger SL, Bhaya-Grossman I, Dougherty ME, Seaton MP, Littlejohn KT, Tu-Chan A, Ganguly K, Moses DA, Chang EF. A bilingual speech neuroprosthesis driven by cortical articulatory representations shared between languages. Nat Biomed Eng 2024; 8:977-991. [PMID: 38769157 PMCID: PMC11554235 DOI: 10.1038/s41551-024-01207-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2023] [Accepted: 04/01/2024] [Indexed: 05/22/2024]
Abstract
Advancements in decoding speech from brain activity have focused on decoding a single language. Hence, the extent to which bilingual speech production relies on unique or shared cortical activity across languages has remained unclear. Here, we leveraged electrocorticography, along with deep-learning and statistical natural-language models of English and Spanish, to record and decode activity from speech-motor cortex of a Spanish-English bilingual with vocal-tract and limb paralysis into sentences in either language. This was achieved without requiring the participant to manually specify the target language. Decoding models relied on shared vocal-tract articulatory representations across languages, which allowed us to build a syllable classifier that generalized across a shared set of English and Spanish syllables. Transfer learning expedited training of the bilingual decoder by enabling neural data recorded in one language to improve decoding in the other language. Overall, our findings suggest shared cortical articulatory representations that persist after paralysis and enable the decoding of multiple languages without the need to train separate language-specific decoders.
Collapse
Affiliation(s)
- Alexander B Silva
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- University of California, Berkeley - University of California, San Francisco Graduate Program in Bioengineering, Berkeley, CA, USA
| | - Jessie R Liu
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- University of California, Berkeley - University of California, San Francisco Graduate Program in Bioengineering, Berkeley, CA, USA
| | - Sean L Metzger
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- University of California, Berkeley - University of California, San Francisco Graduate Program in Bioengineering, Berkeley, CA, USA
| | - Ilina Bhaya-Grossman
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- University of California, Berkeley - University of California, San Francisco Graduate Program in Bioengineering, Berkeley, CA, USA
| | - Maximilian E Dougherty
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
| | - Margaret P Seaton
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
| | - Kaylo T Littlejohn
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, Berkeley, CA, USA
| | - Adelyn Tu-Chan
- Department of Neurology, University of California, San Francisco, San Francisco, CA, USA
| | - Karunesh Ganguly
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- Department of Neurology, University of California, San Francisco, San Francisco, CA, USA
| | - David A Moses
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
| | - Edward F Chang
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA.
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA.
- University of California, Berkeley - University of California, San Francisco Graduate Program in Bioengineering, Berkeley, CA, USA.
| |
Collapse
|
12
|
Wyse-Sookoo K, Luo S, Candrea D, Schippers A, Tippett DC, Wester B, Fifer M, Vansteensel MJ, Ramsey NF, Crone NE. Stability of ECoG high gamma signals during speech and implications for a speech BCI system in an individual with ALS: a year-long longitudinal study. J Neural Eng 2024; 21:10.1088/1741-2552/ad5c02. [PMID: 38925110 PMCID: PMC11245360 DOI: 10.1088/1741-2552/ad5c02] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2023] [Accepted: 06/26/2024] [Indexed: 06/28/2024]
Abstract
Objective.Speech brain-computer interfaces (BCIs) have the potential to augment communication in individuals with impaired speech due to muscle weakness, for example in amyotrophic lateral sclerosis (ALS) and other neurological disorders. However, to achieve long-term, reliable use of a speech BCI, it is essential for speech-related neural signal changes to be stable over long periods of time. Here we study, for the first time, the stability of speech-related electrocorticographic (ECoG) signals recorded from a chronically implanted ECoG BCI over a 12 month period.Approach.ECoG signals were recorded by an ECoG array implanted over the ventral sensorimotor cortex in a clinical trial participant with ALS. Because ECoG-based speech decoding has most often relied on broadband high gamma (HG) signal changes relative to baseline (non-speech) conditions, we studied longitudinal changes of HG band power at baseline and during speech, and we compared these with residual high frequency noise levels at baseline. Stability was further assessed by longitudinal measurements of signal-to-noise ratio, activation ratio, and peak speech-related HG response magnitude (HG response peaks). Lastly, we analyzed the stability of the event-related HG power changes (HG responses) for individual syllables at each electrode.Main Results.We found that speech-related ECoG signal responses were stable over a range of syllables activating different articulators for the first year after implantation.Significance.Together, our results indicate that ECoG can be a stable recording modality for long-term speech BCI systems for those living with severe paralysis.Clinical Trial Information.ClinicalTrials.gov, registration number NCT03567213.
Collapse
Affiliation(s)
- Kimberley Wyse-Sookoo
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
| | - Shiyu Luo
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
| | - Daniel Candrea
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
| | - Anouck Schippers
- Department of Neurology and Neurosurgery, Brain Center, University Medical Center Utrecht, Utrecht, The Netherlands
| | - Donna C Tippett
- Department of Neurology, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
- Department of Otolaryngology-Head and Neck Surgery, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
- Department of Physical Medicine and Rehabilitation, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
| | - Brock Wester
- Research and Exploratory Development Department, Johns Hopkins University Applied Physics Laboratory, Laurel, MD, United States of America
| | - Matthew Fifer
- Research and Exploratory Development Department, Johns Hopkins University Applied Physics Laboratory, Laurel, MD, United States of America
| | - Mariska J Vansteensel
- Department of Neurology and Neurosurgery, Brain Center, University Medical Center Utrecht, Utrecht, The Netherlands
| | - Nick F Ramsey
- Department of Neurology and Neurosurgery, Brain Center, University Medical Center Utrecht, Utrecht, The Netherlands
| | - Nathan E Crone
- Department of Neurology, Johns Hopkins University School of Medicine, Baltimore, MD, United States of America
| |
Collapse
|
13
|
Silva AB, Littlejohn KT, Liu JR, Moses DA, Chang EF. The speech neuroprosthesis. Nat Rev Neurosci 2024; 25:473-492. [PMID: 38745103 PMCID: PMC11540306 DOI: 10.1038/s41583-024-00819-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/12/2024] [Indexed: 05/16/2024]
Abstract
Loss of speech after paralysis is devastating, but circumventing motor-pathway injury by directly decoding speech from intact cortical activity has the potential to restore natural communication and self-expression. Recent discoveries have defined how key features of speech production are facilitated by the coordinated activity of vocal-tract articulatory and motor-planning cortical representations. In this Review, we highlight such progress and how it has led to successful speech decoding, first in individuals implanted with intracranial electrodes for clinical epilepsy monitoring and subsequently in individuals with paralysis as part of early feasibility clinical trials to restore speech. We discuss high-spatiotemporal-resolution neural interfaces and the adaptation of state-of-the-art speech computational algorithms that have driven rapid and substantial progress in decoding neural activity into text, audible speech, and facial movements. Although restoring natural speech is a long-term goal, speech neuroprostheses already have performance levels that surpass communication rates offered by current assistive-communication technology. Given this accelerated rate of progress in the field, we propose key evaluation metrics for speed and accuracy, among others, to help standardize across studies. We finish by highlighting several directions to more fully explore the multidimensional feature space of speech and language, which will continue to accelerate progress towards a clinically viable speech neuroprosthesis.
Collapse
Affiliation(s)
- Alexander B Silva
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
| | - Kaylo T Littlejohn
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
- Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, Berkeley, CA, USA
| | - Jessie R Liu
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
| | - David A Moses
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA
| | - Edward F Chang
- Department of Neurological Surgery, University of California, San Francisco, San Francisco, CA, USA.
- Weill Institute for Neuroscience, University of California, San Francisco, San Francisco, CA, USA.
| |
Collapse
|
14
|
Angrick M, Luo S, Rabbani Q, Candrea DN, Shah S, Milsap GW, Anderson WS, Gordon CR, Rosenblatt KR, Clawson L, Tippett DC, Maragakis N, Tenore FV, Fifer MS, Hermansky H, Ramsey NF, Crone NE. Online speech synthesis using a chronically implanted brain-computer interface in an individual with ALS. Sci Rep 2024; 14:9617. [PMID: 38671062 PMCID: PMC11053081 DOI: 10.1038/s41598-024-60277-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Accepted: 04/21/2024] [Indexed: 04/28/2024] Open
Abstract
Brain-computer interfaces (BCIs) that reconstruct and synthesize speech using brain activity recorded with intracranial electrodes may pave the way toward novel communication interfaces for people who have lost their ability to speak, or who are at high risk of losing this ability, due to neurological disorders. Here, we report online synthesis of intelligible words using a chronically implanted brain-computer interface (BCI) in a man with impaired articulation due to ALS, participating in a clinical trial (ClinicalTrials.gov, NCT03567213) exploring different strategies for BCI communication. The 3-stage approach reported here relies on recurrent neural networks to identify, decode and synthesize speech from electrocorticographic (ECoG) signals acquired across motor, premotor and somatosensory cortices. We demonstrate a reliable BCI that synthesizes commands freely chosen and spoken by the participant from a vocabulary of 6 keywords previously used for decoding commands to control a communication board. Evaluation of the intelligibility of the synthesized speech indicates that 80% of the words can be correctly recognized by human listeners. Our results show that a speech-impaired individual with ALS can use a chronically implanted BCI to reliably produce synthesized words while preserving the participant's voice profile, and provide further evidence for the stability of ECoG for speech-based BCIs.
Collapse
Affiliation(s)
- Miguel Angrick
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA.
| | - Shiyu Luo
- Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Qinwan Rabbani
- Department of Electrical and Computer Engineering, The Johns Hopkins University, Baltimore, MD, USA
| | - Daniel N Candrea
- Department of Biomedical Engineering, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Samyak Shah
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Griffin W Milsap
- Research and Exploratory Development Department, Johns Hopkins Applied Physics Laboratory, Laurel, MD, USA
| | - William S Anderson
- Department of Neurosurgery, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Chad R Gordon
- Department of Neurosurgery, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
- Section of Neuroplastic and Reconstructive Surgery, Department of Plastic Surgery, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Kathryn R Rosenblatt
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
- Department of Anesthesiology & Critical Care Medicine, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Lora Clawson
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Donna C Tippett
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
- Department of Otolaryngology-Head and Neck Surgery, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
- Department of Physical Medicine and Rehabilitation, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Nicholas Maragakis
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA
| | - Francesco V Tenore
- Research and Exploratory Development Department, Johns Hopkins Applied Physics Laboratory, Laurel, MD, USA
| | - Matthew S Fifer
- Research and Exploratory Development Department, Johns Hopkins Applied Physics Laboratory, Laurel, MD, USA
| | - Hynek Hermansky
- Center for Language and Speech Processing, The Johns Hopkins University, Baltimore, MD, USA
- Human Language Technology Center of Excellence, The Johns Hopkins University, Baltimore, MD, USA
| | - Nick F Ramsey
- UMC Utrecht Brain Center, Department of Neurology and Neurosurgery, University Medical Center Utrecht, Utrecht, The Netherlands
| | - Nathan E Crone
- Department of Neurology, The Johns Hopkins University School of Medicine, Baltimore, MD, USA.
| |
Collapse
|
15
|
Card NS, Wairagkar M, Iacobacci C, Hou X, Singer-Clark T, Willett FR, Kunz EM, Fan C, Nia MV, Deo DR, Srinivasan A, Choi EY, Glasser MF, Hochberg LR, Henderson JM, Shahlaie K, Brandman DM, Stavisky SD. An accurate and rapidly calibrating speech neuroprosthesis. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2024:2023.12.26.23300110. [PMID: 38645254 PMCID: PMC11030484 DOI: 10.1101/2023.12.26.23300110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/23/2024]
Abstract
Brain-computer interfaces can enable rapid, intuitive communication for people with paralysis by transforming the cortical activity associated with attempted speech into text on a computer screen. Despite recent advances, communication with brain-computer interfaces has been restricted by extensive training data requirements and inaccurate word output. A man in his 40's with ALS with tetraparesis and severe dysarthria (ALSFRS-R = 23) was enrolled into the BrainGate2 clinical trial. He underwent surgical implantation of four microelectrode arrays into his left precentral gyrus, which recorded neural activity from 256 intracortical electrodes. We report a speech neuroprosthesis that decoded his neural activity as he attempted to speak in both prompted and unstructured conversational settings. Decoded words were displayed on a screen, then vocalized using text-to-speech software designed to sound like his pre-ALS voice. On the first day of system use, following 30 minutes of attempted speech training data, the neuroprosthesis achieved 99.6% accuracy with a 50-word vocabulary. On the second day, the size of the possible output vocabulary increased to 125,000 words, and, after 1.4 additional hours of training data, the neuroprosthesis achieved 90.2% accuracy. With further training data, the neuroprosthesis sustained 97.5% accuracy beyond eight months after surgical implantation. The participant has used the neuroprosthesis to communicate in self-paced conversations for over 248 hours. In an individual with ALS and severe dysarthria, an intracortical speech neuroprosthesis reached a level of performance suitable to restore naturalistic communication after a brief training period.
Collapse
Affiliation(s)
- Nicholas S Card
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Maitreyee Wairagkar
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Carrina Iacobacci
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Xianda Hou
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
- Departments of Computer Science, University of California Davis, Davis, CA, USA
| | - Tyler Singer-Clark
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
- Departments of Biomedical Engineering, University of California Davis, Davis, CA, USA
| | - Francis R Willett
- Departments of Neurosurgery, Stanford University, Stanford, CA, USA
- Departments of Electrical Engineering, Stanford University, Stanford, CA, USA
- Howard Hughes Medical Institute, Stanford University, Stanford, CA, USA
| | - Erin M Kunz
- Departments of Electrical Engineering, Stanford University, Stanford, CA, USA
- Departments of Mechanical Engineering, Stanford University, Stanford, CA, USA
| | - Chaofei Fan
- Departments of Computer Science, Stanford University, Stanford, CA, USA
| | - Maryam Vahdati Nia
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
- Departments of Computer Science, University of California Davis, Davis, CA, USA
| | - Darrel R Deo
- Departments of Neurosurgery, Stanford University, Stanford, CA, USA
| | - Aparna Srinivasan
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
- Departments of Biomedical Engineering, University of California Davis, Davis, CA, USA
| | - Eun Young Choi
- Departments of Neurosurgery, Stanford University, Stanford, CA, USA
| | - Matthew F Glasser
- Departments of Radiology and Neuroscience, Washington University School of Medicine, Saint Louis, MO, USA
| | - Leigh R Hochberg
- School of Engineering and Carney Institute for Brain Sciences, Brown University, Providence, RI, USA
- VA RR&D Center for Neurorestoration and Neurotechnology, VA Providence Healthcare, Providence, RI
- Center for Neurotechnology and Neurorecovery, Department of Neurology, Massachusetts General Hospital, Harvard Medical School, Boston, MA
| | - Jaimie M Henderson
- Departments of Neurosurgery, Stanford University, Stanford, CA, USA
- Wu Tsai Neurosciences Institute, Stanford University, Stanford, CA, USA
| | - Kiarash Shahlaie
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - David M Brandman
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
| | - Sergey D Stavisky
- Departments of Neurological Surgery, University of California Davis, Davis, CA, USA
| |
Collapse
|
16
|
Schippers A, Vansteensel MJ, Freudenburg ZV, Ramsey NF. Don't put words in my mouth: Speech perception can generate False Positive activation of a speech BCI. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2024:2024.01.21.23300437. [PMID: 38343801 PMCID: PMC10854295 DOI: 10.1101/2024.01.21.23300437] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 02/17/2024]
Abstract
Recent studies have demonstrated that speech can be decoded from brain activity and used for brain-computer interface (BCI)-based communication. It is however also known that the area often used as a signal source for speech decoding BCIs, the sensorimotor cortex (SMC), is also engaged when people perceive speech, thus making speech perception a potential source of false positive activation of the BCI. The current study investigated if and how speech perception may interfere with reliable speech BCI control. We recorded high-density electrocorticography (HD-ECoG) data from five subjects while they performed a speech perception and speech production task and trained a support-vector machine (SVM) on the produced speech data. Our results show that decoders that are highly reliable at detecting self-produced speech from brain signals also generate false positives during the perception of speech. We conclude that speech perception interferes with reliable BCI control, and that efforts to limit the occurrence of false positives during daily-life BCI use should be implemented in BCI design to increase the likelihood of successful adaptation by end users.
Collapse
Affiliation(s)
- A Schippers
- Department of Neurology and Neurosurgery, University Medical Center Utrecht Brain Center, Utrecht, The Netherlands
| | - M J Vansteensel
- Department of Neurology and Neurosurgery, University Medical Center Utrecht Brain Center, Utrecht, The Netherlands
| | - Z V Freudenburg
- Department of Neurology and Neurosurgery, University Medical Center Utrecht Brain Center, Utrecht, The Netherlands
| | - N F Ramsey
- Department of Neurology and Neurosurgery, University Medical Center Utrecht Brain Center, Utrecht, The Netherlands
| |
Collapse
|
17
|
Rosenfeld JV. Neurosurgery and the Brain-Computer Interface. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2024; 1462:513-527. [PMID: 39523287 DOI: 10.1007/978-3-031-64892-2_32] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
Brain-computer interfaces (BCIs) are devices that connect the human brain to an effector via a computer and electrode interface. BCIs may also transmit sensory data to the brain. We describe progress with the many types of surgically implanted BCIs, in which electrodes contact or penetrate the cerebral cortex. BCIs developed for restoration of movement in paralyzed limbs or control a robotic arm; restoration of somatic sensation, speech, vision, memory, hearing, and olfaction are also presented. Most devices remain experimental. Commercialization is costly, incurs financial risk, and is time consuming. There are many ethical principles that should be considered by neurosurgeons and by all those responsible for the care of people with serious neurological disability. These considerations are also paramount when the technology is used in for the purpose of enhancement of normal function and where commercial gain is a factor. A new regulatory and legislative framework is urgently required. The evolution of BCIs is occurring rapidly with advances in computer science, artificial intelligence, electronic engineering including wireless transmission, and materials science. The era of the brain-"cloud" interface is approaching.
Collapse
Affiliation(s)
- Jeffrey V Rosenfeld
- Department of Neurosurgery, Alfred Hospital, Melbourne, VIC, Australia.
- Department of Surgery, Monash University, Clayton, VIC, Australia.
- Department of Surgery, F. Edward Hébert School of Medicine, Uniformed Services University of The Health Sciences, Bethesda, MD, USA.
| |
Collapse
|