1
|
Kew SYN, Mok SY, Goh CH. Machine learning and brain-computer interface approaches in prognosis and individualized care strategies for individuals with amyotrophic lateral sclerosis: A systematic review. MethodsX 2024; 13:102765. [PMID: 39286440 PMCID: PMC11403252 DOI: 10.1016/j.mex.2024.102765] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Accepted: 05/15/2024] [Indexed: 09/19/2024] Open
Abstract
Amyotrophic lateral sclerosis (ALS) characterized by progressive degeneration of motor neurons is a debilitating disease, posing substantial challenges in both prognosis and daily life assistance. However, with the advancement of machine learning (ML) which is renowned for tackling many real-world settings, it can offer unprecedented opportunities in prognostic studies and facilitate individuals with ALS in motor-imagery tasks. ML models, such as random forests (RF), have emerged as the most common and effective algorithms for predicting disease progression and survival time in ALS. The findings revealed that RF models had an excellent predictive performance for ALS, with a testing R2 of 0.524 and minimal treatment effects of 0.0717 for patient survival time. Despite significant limitations in sample size, with a maximum of 18 participants, which may not adequately reflect the population diversity being studied, ML approaches have been effectively applied to ALS datasets, and numerous prognostic models have been tested using neuroimaging data, longitudinal datasets, and core clinical variables. In many literatures, the constraints of ML models are seldom explicitly enunciated. Therefore, the main objective of this research is to provide a review of the most significant studies on the usage of ML models for analyzing ALS. This review covers a variation of ML algorithms involved in applications in ALS prognosis besides, leveraging ML to improve the efficacy of brain-computer interfaces (BCIs) for ALS individuals in later stages with restricted voluntary muscular control. The key future advances in individualized care and ALS prognosis may include the advancement of more personalized care aids that enable real-time input and ongoing validation of ML in diverse healthcare contexts.
Collapse
Affiliation(s)
- Stephanie Yen Nee Kew
- Department of Mechatronics and Biomedical Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, 43000 Kajang, Selangor Darul Ehsan, Malaysia
| | - Siew-Ying Mok
- Department of Mechatronics and Biomedical Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, 43000 Kajang, Selangor Darul Ehsan, Malaysia
| | - Choon-Hian Goh
- Department of Mechatronics and Biomedical Engineering, Lee Kong Chian Faculty of Engineering and Science, Universiti Tunku Abdul Rahman, 43000 Kajang, Selangor Darul Ehsan, Malaysia
| |
Collapse
|
2
|
Lee J, Miri S, Bayro A, Kim M, Jeong H, Yeo WH. Biosignal-integrated robotic systems with emerging trends in visual interfaces: A systematic review. BIOPHYSICS REVIEWS 2024; 5:011301. [PMID: 38510371 PMCID: PMC10903439 DOI: 10.1063/5.0185568] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 01/29/2024] [Indexed: 03/22/2024]
Abstract
Human-machine interfaces (HMI) are currently a trendy and rapidly expanding area of research. Interestingly, the human user does not readily observe the interface between humans and machines. Instead, interactions between the machine and electrical signals from the user's body are obscured by complex control algorithms. The result is effectively a one-way street, wherein data is only transmitted from human to machine. Thus, a gap remains in the literature: how can information be effectively conveyed to the user to enable mutual understanding between humans and machines? Here, this paper reviews recent advancements in biosignal-integrated wearable robotics, with a particular emphasis on "visualization"-the presentation of relevant data, statistics, and visual feedback to the user. This review article covers various signals of interest, such as electroencephalograms and electromyograms, and explores novel sensor architectures and key materials. Recent developments in wearable robotics are examined from control and mechanical design perspectives. Additionally, we discuss current visualization methods and outline the field's future direction. While much of the HMI field focuses on biomedical and healthcare applications, such as rehabilitation of spinal cord injury and stroke patients, this paper also covers less common applications in manufacturing, defense, and other domains.
Collapse
Affiliation(s)
| | - Sina Miri
- Department of Mechanical and Industrial Engineering, The University of Illinois at Chicago, Chicago, Illinois 60607, USA
| | - Allison Bayro
- School of Biological and Health Systems Engineering, Ira A. Fulton Schools of Engineering, Arizona State University, Tempe, Arizona 85287, USA
| | - Myunghee Kim
- Department of Mechanical and Industrial Engineering, The University of Illinois at Chicago, Chicago, Illinois 60607, USA
| | - Heejin Jeong
- Authors to whom correspondence should be addressed:; ; and
| | - Woon-Hong Yeo
- Authors to whom correspondence should be addressed:; ; and
| |
Collapse
|
3
|
Yan S, Hu Y, Zhang R, Qi D, Hu Y, Yao D, Shi L, Zhang L. Multilayer network-based channel selection for motor imagery brain-computer interface. J Neural Eng 2024; 21:016029. [PMID: 38295419 DOI: 10.1088/1741-2552/ad2496] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Accepted: 01/31/2024] [Indexed: 02/02/2024]
Abstract
Objective. The number of electrode channels in a motor imagery-based brain-computer interface (MI-BCI) system influences not only its decoding performance, but also its convenience for use in applications. Although many channel selection methods have been proposed in the literature, they are usually based on the univariate features of a single channel. This leads to a loss of the interaction between channels and the exchange of information between networks operating at different frequency bands.Approach. We integrate brain networks containing four frequency bands into a multilayer network framework and propose a multilayer network-based channel selection (MNCS) method for MI-BCI systems. A graph learning-based method is used to estimate the multilayer network from electroencephalogram (EEG) data that are filtered by multiple frequency bands. The multilayer participation coefficient of the multilayer network is then computed to select EEG channels that do not contain redundant information. Furthermore, the common spatial pattern (CSP) method is used to extract effective features. Finally, a support vector machine classifier with a linear kernel is trained to accurately identify MI tasks.Main results. We used three publicly available datasets from the BCI Competition containing data on 12 healthy subjects and one dataset containing data on 15 stroke patients to validate the effectiveness of our proposed method. The results showed that the proposed MNCS method outperforms all channels (85.8% vs. 93.1%, 84.4% vs. 89.0%, 71.7% vs. 79.4%, and 72.7% vs. 84.0%). Moreover, it achieved significantly higher decoding accuracies on MI-BCI systems than state-of-the-art methods (pairedt-tests,p< 0.05).Significance. The experimental results showed that the proposed MNCS method can select appropriate channels to improve the decoding performance as well as the convenience of the application of MI-BCI systems.
Collapse
Affiliation(s)
- Shaoting Yan
- School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou, People's Republic of China
- Henan Key Laboratory of Brain Science and Brain-Computer Interface Technology, Zhengzhou, People's Republic of China
- Institute of Neuroscience, Zhengzhou University, Zhengzhou, People's Republic of China
| | - Yuxia Hu
- School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou, People's Republic of China
- Henan Key Laboratory of Brain Science and Brain-Computer Interface Technology, Zhengzhou, People's Republic of China
- Institute of Neuroscience, Zhengzhou University, Zhengzhou, People's Republic of China
| | - Rui Zhang
- School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou, People's Republic of China
- Henan Key Laboratory of Brain Science and Brain-Computer Interface Technology, Zhengzhou, People's Republic of China
- Institute of Neuroscience, Zhengzhou University, Zhengzhou, People's Republic of China
| | - Daowei Qi
- School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou, People's Republic of China
- Henan Key Laboratory of Brain Science and Brain-Computer Interface Technology, Zhengzhou, People's Republic of China
| | - Yubo Hu
- The No.3 Provincial People's Hospital of Henan Province, Zhengzhou, People's Republic of China
| | - Dezhong Yao
- School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou, People's Republic of China
- Henan Key Laboratory of Brain Science and Brain-Computer Interface Technology, Zhengzhou, People's Republic of China
| | - Li Shi
- Department of Automation, Tsinghua University, Beijing, People's Republic of China
- Beijing National Research Center for Information Science and Technology, Beijing, People's Republic of China
| | - Lipeng Zhang
- School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou, People's Republic of China
- Henan Key Laboratory of Brain Science and Brain-Computer Interface Technology, Zhengzhou, People's Republic of China
- Institute of Neuroscience, Zhengzhou University, Zhengzhou, People's Republic of China
| |
Collapse
|
4
|
Kosmyna N, Hauptmann E, Hmaidan Y. A Brain-Controlled Quadruped Robot: A Proof-of-Concept Demonstration. SENSORS (BASEL, SWITZERLAND) 2023; 24:80. [PMID: 38202942 PMCID: PMC10780665 DOI: 10.3390/s24010080] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Revised: 12/15/2023] [Accepted: 12/19/2023] [Indexed: 01/12/2024]
Abstract
Coupling brain-computer interfaces (BCIs) and robotic systems in the future can enable seamless personal assistant systems in everyday life, with the requests that can be performed in a discrete manner, using one's brain activity only. These types of systems might be of a particular interest for people with locked-in syndrome (LIS) or amyotrophic lateral sclerosis (ALS) because they can benefit from communicating with robotic assistants using brain sensing interfaces. In this proof-of-concept work, we explored how a wireless and wearable BCI device can control a quadruped robot-Boston Dynamics' Spot. The device measures the user's electroencephalography (EEG) and electrooculography (EOG) activity of the user from the electrodes embedded in the glasses' frame. The user responds to a series of questions with YES/NO answers by performing a brain-teaser activity of mental calculus. Each question-answer pair has a pre-configured set of actions for Spot. For instance, Spot was prompted to walk across a room, pick up an object, and retrieve it for the user (i.e., bring a bottle of water) when a sequence resolved to a YES response. Our system achieved at a success rate of 83.4%. To the best of our knowledge, this is the first integration of wireless, non-visual-based BCI systems with Spot in the context of personal assistant use cases. While this BCI quadruped robot system is an early prototype, future iterations may embody friendly and intuitive cues similar to regular service dogs. As such, this project aims to pave a path towards future developments in modern day personal assistant robots powered by wireless and wearable BCI systems in everyday living conditions.
Collapse
Affiliation(s)
- Nataliya Kosmyna
- Media Lab, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | | | - Yasmeen Hmaidan
- Psychology Department, University of Toronto, Toronto, ON M5S 3E4, Canada;
| |
Collapse
|
5
|
Fernandes F, Barbalho I, Bispo Júnior A, Alves L, Nagem D, Lins H, Arrais Júnior E, Coutinho KD, Morais AHF, Santos JPQ, Machado GM, Henriques J, Teixeira C, Dourado Júnior MET, Lindquist ARR, Valentim RAM. Digital Alternative Communication for Individuals with Amyotrophic Lateral Sclerosis: What We Have. J Clin Med 2023; 12:5235. [PMID: 37629277 PMCID: PMC10455505 DOI: 10.3390/jcm12165235] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2023] [Revised: 08/05/2023] [Accepted: 08/09/2023] [Indexed: 08/27/2023] Open
Abstract
Amyotrophic Lateral Sclerosis is a disease that compromises the motor system and the functional abilities of the person in an irreversible way, causing the progressive loss of the ability to communicate. Tools based on Augmentative and Alternative Communication are essential for promoting autonomy and improving communication, life quality, and survival. This Systematic Literature Review aimed to provide evidence on eye-image-based Human-Computer Interaction approaches for the Augmentative and Alternative Communication of people with Amyotrophic Lateral Sclerosis. The Systematic Literature Review was conducted and guided following a protocol consisting of search questions, inclusion and exclusion criteria, and quality assessment, to select primary studies published between 2010 and 2021 in six repositories: Science Direct, Web of Science, Springer, IEEE Xplore, ACM Digital Library, and PubMed. After the screening, 25 primary studies were evaluated. These studies showcased four low-cost, non-invasive Human-Computer Interaction strategies employed for Augmentative and Alternative Communication in people with Amyotrophic Lateral Sclerosis. The strategies included Eye-Gaze, which featured in 36% of the studies; Eye-Blink and Eye-Tracking, each accounting for 28% of the approaches; and the Hybrid strategy, employed in 8% of the studies. For these approaches, several computational techniques were identified. For a better understanding, a workflow containing the development phases and the respective methods used by each strategy was generated. The results indicate the possibility and feasibility of developing Human-Computer Interaction resources based on eye images for Augmentative and Alternative Communication in a control group. The absence of experimental testing in people with Amyotrophic Lateral Sclerosis reiterates the challenges related to the scalability, efficiency, and usability of these technologies for people with the disease. Although challenges still exist, the findings represent important advances in the fields of health sciences and technology, promoting a promising future with possibilities for better life quality.
Collapse
Affiliation(s)
- Felipe Fernandes
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Ingridy Barbalho
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Arnaldo Bispo Júnior
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Luca Alves
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Danilo Nagem
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Hertz Lins
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Ernano Arrais Júnior
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Karilany D. Coutinho
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Antônio H. F. Morais
- Advanced Nucleus of Technological Innovation (NAVI), Federal Institute of Rio Grande do Norte (IFRN), Natal 59015-000, Brazil; (A.H.F.M.); (J.P.Q.S.)
| | - João Paulo Q. Santos
- Advanced Nucleus of Technological Innovation (NAVI), Federal Institute of Rio Grande do Norte (IFRN), Natal 59015-000, Brazil; (A.H.F.M.); (J.P.Q.S.)
| | | | - Jorge Henriques
- Department of Informatics Engineering, Center for Informatics and Systems of the University of Coimbra, Universidade de Coimbra, 3030-788 Coimbra, Portugal; (J.H.); (C.T.)
| | - César Teixeira
- Department of Informatics Engineering, Center for Informatics and Systems of the University of Coimbra, Universidade de Coimbra, 3030-788 Coimbra, Portugal; (J.H.); (C.T.)
| | - Mário E. T. Dourado Júnior
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
- Department of Integrated Medicine, Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil
| | - Ana R. R. Lindquist
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| | - Ricardo A. M. Valentim
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal 59010-090, Brazil; (I.B.); (A.B.J.); (L.A.); (D.N.); (H.L.); (E.A.J.); (K.D.C.); (M.E.T.D.J.); (A.R.R.L.); (R.A.M.V.)
| |
Collapse
|
6
|
Driving Mode Selection through SSVEP-Based BCI and Energy Consumption Analysis. SENSORS 2022; 22:s22155631. [PMID: 35957188 PMCID: PMC9371069 DOI: 10.3390/s22155631] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/12/2022] [Revised: 07/20/2022] [Accepted: 07/25/2022] [Indexed: 12/04/2022]
Abstract
Background: The brain–computer interface (BCI) is a highly cross-discipline technology and its successful application in various domains has received increasing attention. However, the BCI-enabled automobile industry is has been comparatively less investigated. In particular, there are currently no studies focusing on brain-controlled driving mode selection. Specifically, different driving modes indicate different driving styles which can be selected according to the road condition or the preference of individual drivers. Methods: In this paper, a steady-state visual-evoked potential (SSVEP)-based driving mode selection system is proposed. Upon this system, drivers can select the intended driving modes by only gazing at the corresponding SSVEP stimuli. A novel EEG processing algorithm named inter-trial distance minimization analysis (ITDMA) is proposed to enhance SSVEP detection. Both offline and real-time experiments were carried out to validate the effectiveness of the proposed system. Conclusion: The results show that a high selection accuracy of up to 92.3% can be realized, although this depends on the specific choice of flickering duration, the number of EEG channels, and the number of training signals. Additionally, energy consumption is investigated in terms of which the proposed brain-controlled system considerably differs from a traditional driving mode selection system, and the main reason is shown to be the existence of a detection error.
Collapse
|
7
|
Liu K, Yu Y, Liu Y, Tang J, Liang X, Chu X, Zhou Z. A novel brain-controlled wheelchair combined with computer vision and augmented reality. Biomed Eng Online 2022; 21:50. [PMID: 35883092 PMCID: PMC9327337 DOI: 10.1186/s12938-022-01020-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2022] [Accepted: 07/11/2022] [Indexed: 11/10/2022] Open
Abstract
BACKGROUND Brain-controlled wheelchairs (BCWs) are important applications of brain-computer interfaces (BCIs). Currently, most BCWs are semiautomatic. When users want to reach a target of interest in their immediate environment, this semiautomatic interaction strategy is slow. METHODS To this end, we combined computer vision (CV) and augmented reality (AR) with a BCW and proposed the CVAR-BCW: a BCW with a novel automatic interaction strategy. The proposed CVAR-BCW uses a translucent head-mounted display (HMD) as the user interface, uses CV to automatically detect environments, and shows the detected targets through AR technology. Once a user has chosen a target, the CVAR-BCW can automatically navigate to it. For a few scenarios, the semiautomatic strategy might be useful. We integrated a semiautomatic interaction framework into the CVAR-BCW. The user can switch between the automatic and semiautomatic strategies. RESULTS We recruited 20 non-disabled subjects for this study and used the accuracy, information transfer rate (ITR), and average time required for the CVAR-BCW to reach each designated target as performance metrics. The experimental results showed that our CVAR-BCW performed well in indoor environments: the average accuracies across all subjects were 83.6% (automatic) and 84.1% (semiautomatic), the average ITRs were 8.2 bits/min (automatic) and 8.3 bits/min (semiautomatic), the average times required to reach a target were 42.4 s (automatic) and 93.4 s (semiautomatic), and the average workloads and degrees of fatigue for the two strategies were both approximately 20. CONCLUSIONS Our CVAR-BCW provides a user-centric interaction approach and a good framework for integrating more advanced artificial intelligence technologies, which may be useful in the field of disability assistance.
Collapse
Affiliation(s)
- Kaixuan Liu
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China
| | - Yang Yu
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China.
| | - Yadong Liu
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China
| | - Jingsheng Tang
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China
| | - Xinbin Liang
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China
| | - Xingxing Chu
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China
| | - Zongtan Zhou
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, 410073, Hunan, China
| |
Collapse
|
8
|
Zhao X, Jin J, Xu R, Li S, Sun H, Wang X, Cichocki A. A Regional Smoothing Block Sparse Bayesian Learning Method With Temporal Correlation for Channel Selection in P300 Speller. Front Hum Neurosci 2022; 16:875851. [PMID: 35754766 PMCID: PMC9231363 DOI: 10.3389/fnhum.2022.875851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Accepted: 05/18/2022] [Indexed: 11/13/2022] Open
Abstract
The P300-based brain-computer interfaces (BCIs) enable participants to communicate by decoding the electroencephalography (EEG) signal. Different regions of the brain correspond to various mental activities. Therefore, removing weak task-relevant and noisy channels through channel selection is necessary when decoding a specific type of activity from EEG. It can improve the recognition accuracy and reduce the training time of the subsequent models. This study proposes a novel block sparse Bayesian-based channel selection method for the P300 speller. In this method, we introduce block sparse Bayesian learning (BSBL) into the channel selection of P300 BCI for the first time and propose a regional smoothing BSBL (RSBSBL) by combining the spatial distribution properties of EEG. The RSBSBL can determine the number of channels adaptively. To ensure practicality, we design an automatic selection iteration strategy model to reduce the time cost caused by the inverse operation of the large-size matrix. We verified the proposed method on two public P300 datasets and on our collected datasets. The experimental results show that the proposed method can remove the inferior channels and work with the classifier to obtain high-classification accuracy. Hence, RSBSBL has tremendous potential for channel selection in P300 tasks.
Collapse
Affiliation(s)
- Xueqing Zhao
- The Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai, China
| | - Jing Jin
- The Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai, China
- Shenzhen Research Institute of East China University of Technology, Shenzhen, China
| | - Ren Xu
- g.tec medical engineering GmbH, Graz, Austria
| | - Shurui Li
- The Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai, China
| | - Hao Sun
- The Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai, China
| | - Xingyu Wang
- The Key Laboratory of Smart Manufacturing in Energy Chemical Process, Ministry of Education, East China University of Science and Technology, Shanghai, China
| | - Andrzej Cichocki
- Skolkovo Institute of Science and Technology, Moscow, Russia
- Systems Research Institute of Polish Academy of Science, Warsaw, Poland
- Department of Informatics, Nicolaus Copernicus University, Toruń, Poland
| |
Collapse
|
9
|
A Human-Computer Control System Based on Intelligent Recognition of Eye Movements and Its Application in Wheelchair Driving. MULTIMODAL TECHNOLOGIES AND INTERACTION 2021. [DOI: 10.3390/mti5090050] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
This paper presents a practical human-computer interaction system for wheelchair motion through eye tracking and eye blink detection. In this system, the pupil in the eye image has been extracted after binarization, and the center of the pupil was localized to capture the trajectory of eye movement and determine the direction of eye gaze. Meanwhile, convolutional neural networks for feature extraction and classification of open-eye and closed-eye images have been built, and machine learning was performed by extracting features from multiple individual images of open-eye and closed-eye states for input to the system. As an application of this human-computer interaction control system, experimental validation was carried out on a modified wheelchair and the proposed method proved to be effective and reliable based on the experimental results.
Collapse
|
10
|
Fernandes F, Barbalho I, Barros D, Valentim R, Teixeira C, Henriques J, Gil P, Dourado Júnior M. Biomedical signals and machine learning in amyotrophic lateral sclerosis: a systematic review. Biomed Eng Online 2021; 20:61. [PMID: 34130692 PMCID: PMC8207575 DOI: 10.1186/s12938-021-00896-2] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2020] [Accepted: 06/09/2021] [Indexed: 12/11/2022] Open
Abstract
INTRODUCTION The use of machine learning (ML) techniques in healthcare encompasses an emerging concept that envisages vast contributions to the tackling of rare diseases. In this scenario, amyotrophic lateral sclerosis (ALS) involves complexities that are yet not demystified. In ALS, the biomedical signals present themselves as potential biomarkers that, when used in tandem with smart algorithms, can be useful to applications within the context of the disease. METHODS This Systematic Literature Review (SLR) consists of searching for and investigating primary studies that use ML techniques and biomedical signals related to ALS. Following the definition and execution of the SLR protocol, 18 articles met the inclusion, exclusion, and quality assessment criteria, and answered the SLR research questions. DISCUSSIONS Based on the results, we identified three classes of ML applications combined with biomedical signals in the context of ALS: diagnosis (72.22%), communication (22.22%), and survival prediction (5.56%). CONCLUSIONS Distinct algorithmic models and biomedical signals have been reported and present promising approaches, regardless of their classes. In summary, this SLR provides an overview of the primary studies analyzed as well as directions for the construction and evolution of technology-based research within the scope of ALS.
Collapse
Affiliation(s)
- Felipe Fernandes
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal, RN Brazil
| | - Ingridy Barbalho
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal, RN Brazil
| | - Daniele Barros
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal, RN Brazil
| | - Ricardo Valentim
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal, RN Brazil
| | - César Teixeira
- Department of Informatics Engineering, Univ Coimbra, CISUC-Center for Informatics and Systems of the University of Coimbra, Coimbra, Portugal
| | - Jorge Henriques
- Department of Informatics Engineering, Univ Coimbra, CISUC-Center for Informatics and Systems of the University of Coimbra, Coimbra, Portugal
| | - Paulo Gil
- Department of Informatics Engineering, Univ Coimbra, CISUC-Center for Informatics and Systems of the University of Coimbra, Coimbra, Portugal
| | - Mário Dourado Júnior
- Laboratory of Technological Innovation in Health (LAIS), Federal University of Rio Grande do Norte (UFRN), Natal, RN Brazil
| |
Collapse
|
11
|
Jin J, Chen Z, Xu R, Miao Y, Wang X, Jung TP. Developing a Novel Tactile P300 Brain-Computer Interface With a Cheeks-Stim Paradigm. IEEE Trans Biomed Eng 2020; 67:2585-2593. [DOI: 10.1109/tbme.2020.2965178] [Citation(s) in RCA: 55] [Impact Index Per Article: 13.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
12
|
Brain-Computer Interface-Based Humanoid Control: A Review. SENSORS 2020; 20:s20133620. [PMID: 32605077 PMCID: PMC7374399 DOI: 10.3390/s20133620] [Citation(s) in RCA: 30] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/25/2020] [Revised: 06/12/2020] [Accepted: 06/17/2020] [Indexed: 11/17/2022]
Abstract
A Brain-Computer Interface (BCI) acts as a communication mechanism using brain signals to control external devices. The generation of such signals is sometimes independent of the nervous system, such as in Passive BCI. This is majorly beneficial for those who have severe motor disabilities. Traditional BCI systems have been dependent only on brain signals recorded using Electroencephalography (EEG) and have used a rule-based translation algorithm to generate control commands. However, the recent use of multi-sensor data fusion and machine learning-based translation algorithms has improved the accuracy of such systems. This paper discusses various BCI applications such as tele-presence, grasping of objects, navigation, etc. that use multi-sensor fusion and machine learning to control a humanoid robot to perform a desired task. The paper also includes a review of the methods and system design used in the discussed applications.
Collapse
|
13
|
Samsonovich AV. Socially emotional brain-inspired cognitive architecture framework for artificial intelligence. COGN SYST RES 2020. [DOI: 10.1016/j.cogsys.2019.12.002] [Citation(s) in RCA: 35] [Impact Index Per Article: 8.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
14
|
Vaughan TM. Brain-computer interfaces for people with amyotrophic lateral sclerosis. HANDBOOK OF CLINICAL NEUROLOGY 2020; 168:33-38. [DOI: 10.1016/b978-0-444-63934-9.00004-4] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
|
15
|
Tramonte S, Sorbello R, Guger C, Chella A. Acceptability Study of A3-K3 Robotic Architecture for a Neurorobotics Painting. Front Neurorobot 2019; 12:81. [PMID: 30687057 PMCID: PMC6336031 DOI: 10.3389/fnbot.2018.00081] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2017] [Accepted: 11/09/2018] [Indexed: 12/05/2022] Open
Abstract
In this paper, authors present a novel architecture for controlling an industrial robot via Brain Computer Interface. The robot used is a Series 2000 KR 210-2. The robotic arm was fitted with DI drawing devices that clamp, hold and manipulate various artistic media like brushes, pencils, pens. User selected a high-level task, for instance a shape or movement, using a human machine interface and the translation in robot movement was entirely demanded to the Robot Control Architecture defining a plan to accomplish user's task. The architecture was composed by a Human Machine Interface based on P300 Brain Computer Interface and a robotic architecture composed by a deliberative layer and a reactive layer to translate user's high-level command in a stream of movement for robots joints. To create a real-case scenario, the architecture was presented at Ars Electronica Festival, where the A3-K3 architecture has been used for painting. Visitors completed a survey to address 4 self-assessed different dimensions related to human-robot interaction: the technology knowledge, the personal attitude, the innovativeness and the satisfaction. The obtained results have led to further exploring the border of human-robot interaction, highlighting the possibilities of human expression in the interaction process with a machine to create art.
Collapse
Affiliation(s)
- Salvatore Tramonte
- RoboticsLab, Department of Industrial and Digital Innovation (DIID), Universitá di Palermo, Palermo, Italy
| | - Rosario Sorbello
- RoboticsLab, Department of Industrial and Digital Innovation (DIID), Universitá di Palermo, Palermo, Italy
| | | | - Antonio Chella
- RoboticsLab, Department of Industrial and Digital Innovation (DIID), Universitá di Palermo, Palermo, Italy.,ICAR-CNR, Palermo, Italy
| |
Collapse
|
16
|
Robinson N, Thomas KP, Vinod AP. Neurophysiological predictors and spectro-spatial discriminative features for enhancing SMR-BCI. J Neural Eng 2018; 15:066032. [PMID: 30277219 DOI: 10.1088/1741-2552/aae597] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
Neural engineering research is actively engaged in optimizing the robustness of sensorimotor rhythms (SMR)-brain-computer interface (BCI) to boost its potential real-world use. OBJECTIVE This paper investigates two vital factors in efficient and robust SMR-BCI design-algorithms that address subject-variability of optimal features and neurophysiological factors that correlate with BCI performance. Existing SMR-BCI research using electroencephalogram (EEG) to classify bilateral motor imagery (MI) focus on identifying subject-specific frequency bands with most discriminative motor patterns localized to sensorimotor region. APPROACH A novel strategy to further optimize BCI performance by taking into account the variability of discriminative spectral regions across various EEG channels is proposed in this paper. MAIN RESULTS The proposed technique results in a significant ([Formula: see text]) increase in average ([Formula: see text]) classification accuracy by [Formula: see text] accompanied by a considerable reduction in number of channels and bands. The session-to-session transfer variation in spectro-spatial patterns using proposed algorithm is investigated offline and classification performance of the optimized BCI model is successfully evaluated in an online SMR-BCI. Further, the effective prediction of SMR-BCI performance with physiological indicators derived from multi-channel resting-state EEG is demonstrated. The results indicate that the resting state activation patterns such as entropy and gamma power from pre-motor (fronto-central) and posterior (parietal and centro-parietal) areas, and beta power from posterior (centro-parietal) areas estimate BCI performance with minimum error. These patterns, strongly related to BCI performance, may represent certain cognitive states during rest. SIGNIFICANCE Findings reported in this paper imply the need for subject-specific modelling of BCI and the prediction of BCI performance using multi-channel rest-state parameters, to ensure enhanced BCI performance.
Collapse
Affiliation(s)
- Neethu Robinson
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Avenue, Singapore 639798, Singapore
| | | | | |
Collapse
|