1
|
Chowdhury RS, Bose S, Ghosh S, Konar A. Attention Induced Dual Convolutional-Capsule Network (AIDC-CN): A deep learning framework for motor imagery classification. Comput Biol Med 2024; 183:109260. [PMID: 39426071 DOI: 10.1016/j.compbiomed.2024.109260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Revised: 09/08/2024] [Accepted: 10/06/2024] [Indexed: 10/21/2024]
Abstract
In recent times, Electroencephalography (EEG)-based motor imagery (MI) decoding has garnered significant attention due to its extensive applicability in healthcare, including areas such as assistive robotics and rehabilitation engineering. Nevertheless, the decoding of EEG signals presents considerable challenges owing to their inherent complexity, non-stationary characteristics, and low signal-to-noise ratio. Notably, deep learning-based classifiers have emerged as a prominent focus for addressing the EEG signal decoding process. This study introduces a novel deep learning classifier named the Attention Induced Dual Convolutional-Capsule Network (AIDC-CN) with the specific aim of accurately categorizing various motor imagination class labels. To enhance the classifier's performance, a dual feature extraction approach leveraging spectrogram and brain connectivity networks has been employed, diversifying the feature set in the classification task. The main highlights of the proposed AIDC-CN classifier includes the introduction of a dual convolution layer to handle the brain connectivity and spectrogram features, addition of a novel self-attention module (SAM) to accentuate the relevant parts of the convolved spectrogram features, introduction of a new cross-attention module (CAM) to refine the outputs obtained from the dual convolution layers and incorporation of a Gaussian Error Linear Unit (GELU) based dynamic routing algorithm to strengthen the coupling among the primary and secondary capsule layers. Performance analysis undertaken on four public data sets depict the superior performance of the proposed model with respect to the state-of-the-art techniques. The code for this model is available at https://github.com/RiteshSurChowdhury/AIDC-CN.
Collapse
Affiliation(s)
- Ritesh Sur Chowdhury
- Artificial Intelligence Laboratory, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata, 700032, West Bengal, India
| | - Shirsha Bose
- Department of Informatics, Technical University of Munich, Munich, Bavaria 85748, Germany
| | - Sayantani Ghosh
- Artificial Intelligence Laboratory, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata, 700032, West Bengal, India
| | - Amit Konar
- Artificial Intelligence Laboratory, Department of Electronics and Telecommunication Engineering, Jadavpur University, Kolkata, 700032, West Bengal, India.
| |
Collapse
|
2
|
An S, Kim S, Chikontwe P, Park SH. Dual Attention Relation Network With Fine-Tuning for Few-Shot EEG Motor Imagery Classification. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:15479-15493. [PMID: 37379192 DOI: 10.1109/tnnls.2023.3287181] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/30/2023]
Abstract
Recently, motor imagery (MI) electroencephalography (EEG) classification techniques using deep learning have shown improved performance over conventional techniques. However, improving the classification accuracy on unseen subjects is still challenging due to intersubject variability, scarcity of labeled unseen subject data, and low signal-to-noise ratio (SNR). In this context, we propose a novel two-way few-shot network able to efficiently learn how to learn representative features of unseen subject categories and classify them with limited MI EEG data. The pipeline includes an embedding module that learns feature representations from a set of signals, a temporal-attention module to emphasize important temporal features, an aggregation-attention module for key support signal discovery, and a relation module for final classification based on relation scores between a support set and a query signal. In addition to the unified learning of feature similarity and a few-shot classifier, our method can emphasize informative features in support data relevant to the query, which generalizes better on unseen subjects. Furthermore, we propose to fine-tune the model before testing by arbitrarily sampling a query signal from the provided support set to adapt to the distribution of the unseen subject. We evaluate our proposed method with three different embedding modules on cross-subject and cross-dataset classification tasks using brain-computer interface (BCI) competition IV 2a, 2b, and GIST datasets. Extensive experiments show that our model significantly improves over the baselines and outperforms existing few-shot approaches.
Collapse
|
3
|
Keutayeva A, Fakhrutdinov N, Abibullaev B. Compact convolutional transformer for subject-independent motor imagery EEG-based BCIs. Sci Rep 2024; 14:25775. [PMID: 39468119 PMCID: PMC11519587 DOI: 10.1038/s41598-024-73755-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/05/2024] [Accepted: 09/20/2024] [Indexed: 10/30/2024] Open
Abstract
Motor imagery electroencephalography (EEG) analysis is crucial for the development of effective brain-computer interfaces (BCIs), yet it presents considerable challenges due to the complexity of the data and inter-subject variability. This paper introduces EEGCCT, an application of compact convolutional transformers designed specifically to improve the analysis of motor imagery tasks in EEG. Unlike traditional approaches, EEGCCT model significantly enhances generalization from limited data, effectively addressing a common limitation in EEG datasets. We validate and test our models using the open-source BCI Competition IV datasets 2a and 2b, employing a Leave-One-Subject-Out (LOSO) strategy to ensure subject-independent performance. Our findings demonstrate that EEGCCT not only outperforms conventional models like EEGNet in standard evaluations but also achieves better performance compared to other advanced models such as Conformer, Hybrid s-CViT, and Hybrid t-CViT, while utilizing fewer parameters and achieving an accuracy of 70.12%. Additionally, the paper presents a comprehensive ablation study that includes targeted data augmentation, hyperparameter optimization, and architectural improvements.
Collapse
Affiliation(s)
- Aigerim Keutayeva
- Institute of Smart Systems and Artificial Intelligence (ISSAI), Nazarbayev University, Astana, 010000, Kazakhstan.
| | - Nail Fakhrutdinov
- Department of Computer Science, Nazarbayev University, Astana, 010000, Kazakhstan
| | - Berdakh Abibullaev
- Department of Robotics Engineering, Nazarbayev University, Astana, 010000, Kazakhstan
| |
Collapse
|
4
|
Liao L, Lu J, Wang L, Zhang Y, Gao D, Wang M. CT-Net: an interpretable CNN-Transformer fusion network for fNIRS classification. Med Biol Eng Comput 2024; 62:3233-3247. [PMID: 38816665 DOI: 10.1007/s11517-024-03138-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 05/23/2024] [Indexed: 06/01/2024]
Abstract
Functional near-infrared spectroscopy (fNIRS), an optical neuroimaging technique, has been widely used in the field of brain activity recognition and brain-computer interface. Existing works have proposed deep learning-based algorithms for the fNIRS classification problem. In this paper, a novel approach based on convolutional neural network and Transformer, named CT-Net, is established to guide the deep modeling for the classification of mental arithmetic (MA) tasks. We explore the effect of data representations, and design a temporal-level combination of two raw chromophore signals to improve the data utilization and enrich the feature learning of the model. We evaluate our model on two open-access datasets and achieve the classification accuracy of 98.05% and 77.61%, respectively. Moreover, we explain our model by the gradient-weighted class activation mapping, which presents a high consistent between the contributing value of features learned by the model and the mapping of brain activity in the MA task. The results suggest the feasibility and interpretability of CT-Net for decoding MA tasks.
Collapse
Affiliation(s)
- Lingxiang Liao
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Jingqing Lu
- School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Lutao Wang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Yongqing Zhang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Dongrui Gao
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China
| | - Manqing Wang
- School of Computer Science, Chengdu University of Information Technology, Chengdu, 610225, China.
- School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, 611731, China.
| |
Collapse
|
5
|
Xia X, Shi Y, Li P, Liu X, Liu J, Men H. FBANet: An Effective Data Mining Method for Food Olfactory EEG Recognition. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:13550-13560. [PMID: 37220050 DOI: 10.1109/tnnls.2023.3269949] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
At present, the sensory evaluation of food mostly depends on artificial sensory evaluation and machine perception, but artificial sensory evaluation is greatly interfered with by subjective factors, and machine perception is difficult to reflect human feelings. In this article, a frequency band attention network (FBANet) for olfactory electroencephalogram (EEG) was proposed to distinguish the difference in food odor. First, the olfactory EEG evoked experiment was designed to collect the olfactory EEG, and the preprocessing of olfactory EEG, such as frequency division, was completed. Second, the FBANet consisted of frequency band feature mining and frequency band feature self-attention, in which frequency band feature mining can effectively mine multiband features of olfactory EEG with different scales, and frequency band feature self-attention can integrate the extracted multiband features and realize classification. Finally, compared with other advanced models, the performance of the FBANet was evaluated. The results show that FBANet was better than the state-of-the-art techniques. In conclusion, FBANet effectively mined the olfactory EEG data information and distinguished the differences between the eight food odors, which proposed a new idea for food sensory evaluation based on multiband olfactory EEG analysis.
Collapse
|
6
|
Li D, Wang J, Xu J, Fang X, Ji Y. Cross-Channel Specific-Mutual Feature Transfer Learning for Motor Imagery EEG Signals Decoding. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:13472-13482. [PMID: 37220058 DOI: 10.1109/tnnls.2023.3269512] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
In recent years, with the rapid development of deep learning, various deep learning frameworks have been widely used in brain-computer interface (BCI) research for decoding motor imagery (MI) electroencephalogram (EEG) signals to understand brain activity accurately. The electrodes, however, record the mixed activities of neurons. If different features are directly embedded in the same feature space, the specific and mutual features of different neuron regions are not considered, which will reduce the expression ability of the feature itself. We propose a cross-channel specific-mutual feature transfer learning (CCSM-FT) network model to solve this problem. The multibranch network extracts the specific and mutual features of brain's multiregion signals. Effective training tricks are used to maximize the distinction between the two kinds of features. Suitable training tricks can also improve the effectiveness of the algorithm compared with novel models. Finally, we transfer two kinds of features to explore the potential of mutual and specific features to enhance the expressive power of the feature and use the auxiliary set to improve identification performance. The experimental results show that the network has a better classification effect in the BCI Competition IV-2a and the HGD datasets.
Collapse
|
7
|
Mahalungkar SP, Shrivastava R, Angadi S. A brief survey on human activity recognition using motor imagery of EEG signals. Electromagn Biol Med 2024; 43:312-327. [PMID: 39425602 DOI: 10.1080/15368378.2024.2415089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 10/06/2024] [Indexed: 10/21/2024]
Abstract
Human being's biological processes and psychological activities are jointly connected to the brain. So, the examination of human activity is more significant for the well-being of humans. There are various models for brain activity detection considering neuroimaging for attaining decreased time requirement, increased control commands, and enhanced accuracy. Motor Imagery (MI)-based Brain-Computer Interface (BCI) systems create a way in which the brain can interact with the environment by processing Electroencephalogram (EEG) signals. Human Activity Recognition (HAR) deals with identifying the physiological activities of human beings based on sensory signals. This survey reviews the different methods available for HAR based on MI-EEG signals. A total of 50 research articles based on HAR from EEG signals are considered in this survey. This survey discusses the challenges faced by various techniques for HAR. Moreover, the papers are assessed considering various parameters, techniques, publication year, performance metrics, utilized tools, employed databases, etc. There were many techniques developed to solve the problem of HAR and they are classified as Machine Learning (ML) and Deep Learning (DL)models. At last, the research gaps and limitations of the techniques were discussed that contribute to developing an effective HAR.
Collapse
Affiliation(s)
- Seema Pankaj Mahalungkar
- Department of Computer Science and Engineering, Mansarovar Global University, Bhopal, Madhya Pradesh, India
- Computer Science and Engineering, Nutan College of Engineering and Research, Talegaon Dabhade, Pune, India
| | - Rahul Shrivastava
- School of Computer Science and Engineering, VIT Bhopal University, Bhopal, Madhya Pradesh, India
| | - Sanjeevkumar Angadi
- Computer Science and Engineering, Nutan College of Engineering and Research, Talegaon Dabhade, Pune, India
| |
Collapse
|
8
|
Lee MH, Shomanov A, Begim B, Kabidenova Z, Nyssanbay A, Yazici A, Lee SW. EAV: EEG-Audio-Video Dataset for Emotion Recognition in Conversational Contexts. Sci Data 2024; 11:1026. [PMID: 39300129 PMCID: PMC11413008 DOI: 10.1038/s41597-024-03838-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 08/29/2024] [Indexed: 09/22/2024] Open
Abstract
Understanding emotional states is pivotal for the development of next-generation human-machine interfaces. Human behaviors in social interactions have resulted in psycho-physiological processes influenced by perceptual inputs. Therefore, efforts to comprehend brain functions and human behavior could potentially catalyze the development of AI models with human-like attributes. In this study, we introduce a multimodal emotion dataset comprising data from 30-channel electroencephalography (EEG), audio, and video recordings from 42 participants. Each participant engaged in a cue-based conversation scenario, eliciting five distinct emotions: neutral, anger, happiness, sadness, and calmness. Throughout the experiment, each participant contributed 200 interactions, which encompassed both listening and speaking. This resulted in a cumulative total of 8,400 interactions across all participants. We evaluated the baseline performance of emotion recognition for each modality using established deep neural network (DNN) methods. The Emotion in EEG-Audio-Visual (EAV) dataset represents the first public dataset to incorporate three primary modalities for emotion recognition within a conversational context. We anticipate that this dataset will make significant contributions to the modeling of the human emotional process, encompassing both fundamental neuroscience and machine learning viewpoints.
Collapse
Affiliation(s)
- Min-Ho Lee
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Adai Shomanov
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Balgyn Begim
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Zhuldyz Kabidenova
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Aruna Nyssanbay
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Adnan Yazici
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Seong-Whan Lee
- Korea University, Department of Artificial Intelligence, Seoul, 02841, Republic of Korea.
| |
Collapse
|
9
|
Sultana M, Perdikis S. Automatic Feature Selection for Sensorimotor Rhythms Brain-Computer Interface Fusing Expert and Data-Driven Knowledge. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3422-3431. [PMID: 39250352 DOI: 10.1109/tnsre.2024.3456591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/11/2024]
Abstract
Early brain-computer interface (BCI) systems were mainly based on prior neurophysiological knowledge coupled with feedback training, while state-of-the-art interfaces rely on data-driven, machine learning (ML)-oriented methods. Despite the advances in BCI that ML can be credited with, the performance of BCI solutions is still not up to the mark, posing a major barrier to the widespread use of this technology. This paper proposes a novel, automatic feature selection method for BCI able to leverage both data-dependent and expert knowledge to suppress noisy features and highlight the most relevant ones thanks to a fuzzy logic (FL) system. Our approach exploits the capability of FL to increase the reliability of decision-making by fusing heterogeneous information channels while maintaining transparency and simplicity. We show that our method leads to significant improvement in classification accuracy, feature stability and class bias when applied to large motor imagery or attempt datasets including end-users with motor disabilities. We postulate that combining data-driven methods with knowledge derived from neuroscience literature through FL can enhance the performance, explainability, and learnability of BCIs.
Collapse
|
10
|
Rong F, Yang B, Guan C. Decoding Multi-Class Motor Imagery From Unilateral Limbs Using EEG Signals. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3399-3409. [PMID: 39236133 DOI: 10.1109/tnsre.2024.3454088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/07/2024]
Abstract
The EEG is a widely utilized neural signal source, particularly in motor imagery-based brain-computer interface (MI-BCI), offering distinct advantages in applications like stroke rehabilitation. Current research predominantly concentrates on the bilateral limbs paradigm and decoding, but the use scenarios for stroke rehabilitation are typically for unilateral upper limbs. There is a significant challenge to decoding unilateral MI of multitasks due to the overlapped spatial neural activities of the tasks. This study aims to formulate a novel MI-BCI experimental paradigm for unilateral limbs with multitasks. The paradigm encompasses four imagined movement directions: top-bottom, left-right, top right-bottom left, and top left-bottom right. Forty-six healthy subjects participated in this experiment. Commonly used machine learning techniques, such as FBCSP, EEGNet, deepConvNet, and FBCNet, were employed for evaluation. To improve decoding accuracy, we propose an MVCA method that introduces temporal convolution and attention mechanism to effectively capture temporal features from multiple perspectives. With the MVCA model, we have achieved 40.6% and 64.89% classification accuracies for the four-class and two-class scenarios (top right-bottom left and top left-bottom right), respectively. Conclusion: This is the first study demonstrating that motor imagery of multiple directions in unilateral limbs can be decoded. In particular, decoding two directions, right top to left bottom and left top to right bottom, provides the best accuracy, which sheds light on future studies. This study advances the development of the MI-BCI paradigm, offering preliminary evidence for the feasibility of decoding multiple directional information from EEG. This, in turn, enhances the dimensions of MI control commands.
Collapse
|
11
|
Rajpura P, Cecotti H, Kumar Meena Y. Explainable artificial intelligence approaches for brain-computer interfaces: a review and design space. J Neural Eng 2024; 21:041003. [PMID: 39029500 DOI: 10.1088/1741-2552/ad6593] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Accepted: 07/19/2024] [Indexed: 07/21/2024]
Abstract
Objective.This review paper provides an integrated perspective of Explainable Artificial Intelligence (XAI) techniques applied to Brain-Computer Interfaces (BCIs). BCIs use predictive models to interpret brain signals for various high-stake applications. However, achieving explainability in these complex models is challenging as it compromises accuracy. Trust in these models can be established by incorporating reasoning or causal relationships from domain experts. The field of XAI has emerged to address the need for explainability across various stakeholders, but there is a lack of an integrated perspective in XAI for BCI (XAI4BCI) literature. It is necessary to differentiate key concepts like explainability, interpretability, and understanding, often used interchangeably in this context, and formulate a comprehensive framework.Approach.To understand the need of XAI for BCI, we pose six key research questions for a systematic review and meta-analysis, encompassing its purposes, applications, usability, and technical feasibility. We employ the PRISMA methodology-preferred reporting items for systematic reviews and meta-analyses to review (n = 1246) and analyse (n = 84) studies published in 2015 and onwards for key insights.Main results.The results highlight that current research primarily focuses on interpretability for developers and researchers, aiming to justify outcomes and enhance model performance. We discuss the unique approaches, advantages, and limitations of XAI4BCI from the literature. We draw insights from philosophy, psychology, and social sciences. We propose a design space for XAI4BCI, considering the evolving need to visualise and investigate predictive model outcomes customised for various stakeholders in the BCI development and deployment lifecycle.Significance.This paper is the first to focus solely on reviewing XAI4BCI research articles. This systematic review and meta-analysis findings with the proposed design space prompt important discussions on establishing standards for BCI explanations, highlighting current limitations, and guiding the future of XAI in BCI.
Collapse
Affiliation(s)
- Param Rajpura
- Human-AI Interaction (HAIx) Lab, Indian Institute of Technology Gandhinagar, Gandhinagar, India
| | - Hubert Cecotti
- Department of Computer Science, California State University, Fresno, CA, United States of America
| | - Yogesh Kumar Meena
- Human-AI Interaction (HAIx) Lab, Indian Institute of Technology Gandhinagar, Gandhinagar, India
| |
Collapse
|
12
|
Luo Y, Mu W, Wang L, Wang J, Wang P, Gan Z, Zhang L, Kang X. An EEG channel selection method for motor imagery based on Fisher score and local optimization. J Neural Eng 2024; 21:036030. [PMID: 38842111 DOI: 10.1088/1741-2552/ad504a] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 05/24/2024] [Indexed: 06/07/2024]
Abstract
Objective. Multi-channel electroencephalogram (EEG) technology in brain-computer interface (BCI) research offers the advantage of enhanced spatial resolution and system performance. However, this also implies that more time is needed in the data processing stage, which is not conducive to the rapid response of BCI. Hence, it is a necessary and challenging task to reduce the number of EEG channels while maintaining decoding effectiveness.Approach. In this paper, we propose a local optimization method based on the Fisher score for within-subject EEG channel selection. Initially, we extract the common spatial pattern characteristics of EEG signals in different bands, calculate Fisher scores for each channel based on these characteristics, and rank them accordingly. Subsequently, we employ a local optimization method to finalize the channel selection.Main results. On the BCI Competition IV Dataset IIa, our method selects an average of 11 channels across four bands, achieving an average accuracy of 79.37%. This represents a 6.52% improvement compared to using the full set of 22 channels. On our self-collected dataset, our method similarly achieves a significant improvement of 24.20% with less than half of the channels, resulting in an average accuracy of 76.95%.Significance. This research explores the importance of channel combinations in channel selection tasks and reveals that appropriately combining channels can further enhance the quality of channel selection. The results indicate that the model selected a small number of channels with higher accuracy in two-class motor imagery EEG classification tasks. Additionally, it improves the portability of BCI systems through channel selection and combinations, offering the potential for the development of portable BCI systems.
Collapse
Affiliation(s)
- Yangjie Luo
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Wei Mu
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Lu Wang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Junkongshuai Wang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Pengchao Wang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
| | - Zhongxue Gan
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
- Ji Hua Laboratory, Foshan, People's Republic of China
| | - Lihua Zhang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
- Ji Hua Laboratory, Foshan, People's Republic of China
| | - Xiaoyang Kang
- Laboratory for Neural Interface and Brain Computer Interface, Engineering Research Center of AI & Robotics, Ministry of Education, Shanghai Engineering Research Center of AI & Robotics, MOE Frontiers Center for Brain Science, State Key Laboratory of Medical Neurobiology, Institute of AI & Robotics, Academy for Engineering & Technology, Fudan University, Shanghai, People's Republic of China
- Ji Hua Laboratory, Foshan, People's Republic of China
- Yiwu Research Institute of Fudan University, Yiwu City, People's Republic of China
- Research Center for Intelligent Sensing, Zhejiang Lab, Hangzhou, People's Republic of China
| |
Collapse
|
13
|
Miao M, Yang Z, Sheng Z, Xu B, Zhang W, Cheng X. Multi-source deep domain adaptation ensemble framework for cross-dataset motor imagery EEG transfer learning. Physiol Meas 2024; 45:055024. [PMID: 38772402 DOI: 10.1088/1361-6579/ad4e95] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2023] [Accepted: 05/21/2024] [Indexed: 05/23/2024]
Abstract
Objective. Electroencephalography (EEG) is an important kind of bioelectric signal for measuring physiological activities of the brain, and motor imagery (MI) EEG has significant clinical application prospects. Convolutional neural network has become a mainstream algorithm for MI EEG classification, however lack of subject-specific data considerably restricts its decoding accuracy and generalization performance. To address this challenge, a novel transfer learning (TL) framework using auxiliary dataset to improve the MI EEG classification performance of target subject is proposed in this paper.Approach. We developed a multi-source deep domain adaptation ensemble framework (MSDDAEF) for cross-dataset MI EEG decoding. The proposed MSDDAEF comprises three main components: model pre-training, deep domain adaptation, and multi-source ensemble. Moreover, for each component, different designs were examined to verify the robustness of MSDDAEF.Main results. Bidirectional validation experiments were performed on two large public MI EEG datasets (openBMI and GIST). The highest average classification accuracy of MSDDAEF reaches 74.28% when openBMI serves as target dataset and GIST serves as source dataset. While the highest average classification accuracy of MSDDAEF is 69.85% when GIST serves as target dataset and openBMI serves as source dataset. In addition, the classification performance of MSDDAEF surpasses several well-established studies and state-of-the-art algorithms.Significance. The results of this study show that cross-dataset TL is feasible for left/right-hand MI EEG decoding, and further indicate that MSDDAEF is a promising solution for addressing MI EEG cross-dataset variability.
Collapse
Affiliation(s)
- Minmin Miao
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou, People's Republic of China
| | - Zhong Yang
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
| | - Zhenzhen Sheng
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou, People's Republic of China
| | - Baoguo Xu
- School of Instrument Science and Engineering, Southeast University, Nanjing, People's Republic of China
| | - Wenbin Zhang
- College of Computer Science and Software Engineering, Hohai University, Nanjing, Jiangsu Province, People's Republic of China
| | - Xinmin Cheng
- School of Information Engineering, Huzhou University, Huzhou, People's Republic of China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural Resources, Huzhou University, Huzhou, People's Republic of China
| |
Collapse
|
14
|
Akuthota S, K R, Ravichander J. Artifact removal and motor imagery classification in EEG using advanced algorithms and modified DNN. Heliyon 2024; 10:e27198. [PMID: 38560190 PMCID: PMC10980936 DOI: 10.1016/j.heliyon.2024.e27198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Revised: 02/21/2024] [Accepted: 02/26/2024] [Indexed: 04/04/2024] Open
Abstract
This paper presents an advanced approach for EEG artifact removal and motor imagery classification using a combination of Four Class Iterative Filtering and Filter Bank Common Spatial Pattern Algorithm with a Modified Deep Neural Network (DNN) classifier. The research aims to enhance the accuracy and reliability of BCI systems by addressing the challenges posed by EEG artifacts and complex motor imagery tasks. The methodology begins by introducing FCIF, a novel technique for ocular artifact removal, utilizing iterative filtering and filter banks. FCIF's mathematical formulation allows for effective artifact mitigation, thereby improving the quality of EEG data. In tandem, the FC-FBCSP algorithm is introduced, extending the Filter Bank Common Spatial Pattern approach to handle four-class motor imagery classification. The Modified DNN classifier enhances the discriminatory power of the FC-FBCSP features, optimizing the classification process. The paper showcases a comprehensive experimental setup, featuring the utilization of BCI Competition IV Dataset 2a & 2b. Detailed preprocessing steps, including filtering and feature extraction, are presented with mathematical rigor. Results demonstrate the remarkable artifact removal capabilities of FCIF and the classification prowess of FC-FBCSP combined with the Modified DNN classifier. Comparative analysis highlights the superiority of the proposed approach over baseline methods and the method achieves the mean accuracy of 98.575%.
Collapse
Affiliation(s)
- Srinath Akuthota
- Department of Electronics & Communication Engineering, SR University, Warangal-506371, Telangana, India
| | - RajKumar K
- Department of Electronics & Communication Engineering, SR University, Warangal-506371, Telangana, India
| | - Janapati Ravichander
- Department of Electronics & Communication Engineering, SR University, Warangal-506371, Telangana, India
| |
Collapse
|
15
|
Li W, Li H, Sun X, Kang H, An S, Wang G, Gao Z. Self-supervised contrastive learning for EEG-based cross-subject motor imagery recognition. J Neural Eng 2024; 21:026038. [PMID: 38565100 DOI: 10.1088/1741-2552/ad3986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 03/20/2024] [Indexed: 04/04/2024]
Abstract
Objective. The extensive application of electroencephalography (EEG) in brain-computer interfaces (BCIs) can be attributed to its non-invasive nature and capability to offer high-resolution data. The acquisition of EEG signals is a straightforward process, but the datasets associated with these signals frequently exhibit data scarcity and require substantial resources for proper labeling. Furthermore, there is a significant limitation in the generalization performance of EEG models due to the substantial inter-individual variability observed in EEG signals.Approach. To address these issues, we propose a novel self-supervised contrastive learning framework for decoding motor imagery (MI) signals in cross-subject scenarios. Specifically, we design an encoder combining convolutional neural network and attention mechanism. In the contrastive learning training stage, the network undergoes training with the pretext task of data augmentation to minimize the distance between pairs of homologous transformations while simultaneously maximizing the distance between pairs of heterologous transformations. It enhances the amount of data utilized for training and improves the network's ability to extract deep features from original signals without relying on the true labels of the data.Main results. To evaluate our framework's efficacy, we conduct extensive experiments on three public MI datasets: BCI IV IIa, BCI IV IIb, and HGD datasets. The proposed method achieves cross-subject classification accuracies of 67.32%, 82.34%, and 81.13%on the three datasets, demonstrating superior performance compared to existing methods.Significance. Therefore, this method has great promise for improving the performance of cross-subject transfer learning in MI-based BCI systems.
Collapse
Affiliation(s)
- Wenjie Li
- Tianjin International Engineering Institute, Tianjin University, Tianjin 300072, People's Republic of China
| | - Haoyu Li
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Xinlin Sun
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Huicong Kang
- Department of Neurology, Shanxi Bethune Hospital, Shanxi Academy of Medical Science, Tongji Shanxi Hospital, Third Hospital of Shanxi Medical University, Taiyuan 030000, People's Republic of China
- Department of Neurology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430000, People's Republic of China
| | - Shan An
- JD Health International Inc., Beijing 100176, People's Republic of China
| | - Guoxin Wang
- JD Health International Inc., Beijing 100176, People's Republic of China
| | - Zhongke Gao
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| |
Collapse
|
16
|
Srimadumathi V, Ramasubba Reddy M. Classification of Motor Imagery EEG signals using high resolution time-frequency representations and convolutional neural network. Biomed Phys Eng Express 2024; 10:035025. [PMID: 38513274 DOI: 10.1088/2057-1976/ad3647] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2023] [Accepted: 03/21/2024] [Indexed: 03/23/2024]
Abstract
A Motor Imagery (MI) based Brain Computer Interface (BCI) system aims to provide neuro-rehabilitation for the motor disabled people and patients with brain injuries (e.g., stroke patients) etc. The aim of this work is to classify the left and right hand MI tasks by utilizing the occurrence of event related desynchronization and synchronization (ERD\ERS) in the Electroencephalogram (EEG) during these tasks. This study proposes to use a set of Complex Morlet Wavelets (CMW) having frequency dependent widths to generate high-resolution time-frequency representations (TFR) of the MI EEG signals present in the channels C3 and C4. A novel method for the selection of the value of number of cycles relative to the center frequency of the CMW is studied here for extracting the MI task features. The generated TFRs are given as input to a Convolutional neural network (CNN) for classifying them into left or right hand MI tasks. The proposed framework attains a classification accuracy of 82.2% on the BCI Competition IV dataset 2a, showing that the TFRs generated in this work give a higher classification accuracy than the baseline methods and other existing algorithms.
Collapse
Affiliation(s)
- V Srimadumathi
- Department of Applied Mechanics and Biomedical Engineering, Indian Institute of Technology, Madras, 600036, India
| | - M Ramasubba Reddy
- Department of Applied Mechanics and Biomedical Engineering, Indian Institute of Technology, Madras, 600036, India
| |
Collapse
|
17
|
Han J, Gu X, Yang GZ, Lo B. Noise-Factorized Disentangled Representation Learning for Generalizable Motor Imagery EEG Classification. IEEE J Biomed Health Inform 2024; 28:765-776. [PMID: 38010934 DOI: 10.1109/jbhi.2023.3337072] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2023]
Abstract
Motor Imagery (MI) Electroencephalography (EEG) is one of the most common Brain-Computer Interface (BCI) paradigms that has been widely used in neural rehabilitation and gaming. Although considerable research efforts have been dedicated to developing MI EEG classification algorithms, they are mostly limited in handling scenarios where the training and testing data are not from the same subject or session. Such poor generalization capability significantly limits the realization of BCI in real-world applications. In this paper, we proposed a novel framework to disentangle the representation of raw EEG data into three components, subject/session-specific, MI-task-specific, and random noises, so that the subject/session-specific feature extends the generalization capability of the system. This is realized by a joint discriminative and generative framework, supported by a series of fundamental training losses and training strategies. We evaluated our framework on three public MI EEG datasets, and detailed experimental results show that our method can achieve superior performance by a large margin compared to current state-of-the-art benchmark algorithms.
Collapse
|
18
|
Park D, Park H, Kim S, Choo S, Lee S, Nam CS, Jung JY. Spatio-Temporal Explanation of 3D-EEGNet for Motor Imagery EEG Classification Using Permutation and Saliency. IEEE Trans Neural Syst Rehabil Eng 2023; 31:4504-4513. [PMID: 37934650 DOI: 10.1109/tnsre.2023.3330922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2023]
Abstract
Recently, convolutional neural network (CNN)-based classification models have shown good performance for motor imagery (MI) brain-computer interfaces (BCI) using electroencephalogram (EEG) in end-to-end learning. Although a few explainable artificial intelligence (XAI) techniques have been developed, it is still challenging to interpret the CNN models for EEG-based BCI classification effectively. In this research, we propose 3D-EEGNet as a 3D CNN model to improve both the explainability and performance of MI EEG classification. The proposed approach exhibited better performances on two MI EEG datasets than the existing EEGNet, which uses a 2D input shape. The MI classification accuracies are improved around 1.8% and 6.1% point in average on the datasets, respectively. The permutation-based XAI method is first applied for the reliable explanation of the 3D-EEGNet. Next, to find a faster XAI method for spatio-temporal explanation, we design a novel technique based on the normalized discounted cumulative gain (NDCG) for selecting the best among a few saliency-based methods due to their higher time complexity than the permutation-based method. Among the saliency-based methods, DeepLIFT was selected because the NDCG scores indicated its results are the most similar to the permutation-based results. Finally, the fast spatio-temporal explanation using DeepLIFT provides deeper understanding for the classification results of the 3D-EEGNet and the important properties in the MI EEG experiments.
Collapse
|
19
|
Deny P, Cheon S, Son H, Choi KW. Hierarchical Transformer for Motor Imagery-Based Brain Computer Interface. IEEE J Biomed Health Inform 2023; 27:5459-5470. [PMID: 37578918 DOI: 10.1109/jbhi.2023.3304646] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/16/2023]
Abstract
In this paper, we propose a novel transformer-based classification algorithm for the brain computer interface (BCI) using a motor imagery (MI) electroencephalogram (EEG) signal. To design the MI classification algorithm, we apply an up-to-date deep learning model, the transformer, that has revolutionized the natural language processing (NLP) and successfully widened its application to many other domains such as the computer vision. Within a long MI trial spanning a few seconds, the classification algorithm should give more attention to the time periods during which the intended motor task is imagined by the subject without any artifact. To achieve this goal, we propose a hierarchical transformer architecture that consists of a high-level transformer (HLT) and a low-level transformer (LLT). We break down a long MI trial into a number of short-term intervals. The LLT extracts a feature from each short-term interval, and the HLT pays more attention to the features from more relevant short-term intervals by using the self-attention mechanism of the transformer. We have done extensive tests of the proposed scheme on four open MI datasets, and shown that the proposed hierarchical transformer excels in both the subject-dependent and subject-independent tests.
Collapse
|
20
|
Guo L, Yu T, Zhao S, Li X, Liao X, Li Y. CLEP: Contrastive Learning for Epileptic Seizure Prediction Using a Spatio-Temporal-Spectral Network. IEEE Trans Neural Syst Rehabil Eng 2023; 31:3915-3926. [PMID: 37796668 DOI: 10.1109/tnsre.2023.3322275] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/07/2023]
Abstract
Seizure prediction of epileptic preictal period through electroencephalogram (EEG) signals is important for clinical epilepsy diagnosis. However, recent deep learning-based methods commonly employ intra-subject training strategy and need sufficient data, which are laborious and time-consuming for a practical system and pose a great challenge for seizure predicting. Besides, multi-domain characterizations, including spatio-temporal-spectral dependencies in an epileptic brain are generally neglected or not considered simultaneously in current approaches, and this insufficiency commonly leads to suboptimal seizure prediction performance. To tackle the above issues, in this paper, we propose Contrastive Learning for Epileptic seizure Prediction (CLEP) using a Spatio-Temporal-Spectral Network (STS-Net). Specifically, the CLEP learns intrinsic epileptic EEG patterns across subjects by contrastive learning. The STS-Net extracts multi-scale temporal and spectral representations under different rhythms from raw EEG signals. Then, a novel triple attention layer (TAL) is employed to construct inter-dimensional interaction among multi-domain features. Moreover, a spatio dynamic graph convolution network (sdGCN) is proposed to dynamically model the spatial relationships between electrodes and aggregate spatial information. The proposed CLEP-STS-Net achieves a sensitivity of 96.7% and a false prediction rate of 0.072/h on the CHB-MIT scalp EEG database. We also validate the proposed method on clinical intracranial EEG (iEEG) database from our Xuanwu Hospital of Capital Medical University, and the predicting system yielded a sensitivity of 95%, a false prediction rate of 0.087/h. The experimental results outperform the state-of-the-art studies which validate the efficacy of our method. Our code is available at https://github.com/LianghuiGuo/CLEP-STS-Net.
Collapse
|
21
|
Zhang D, Li H, Xie J, Li D. MI-DAGSC: A domain adaptation approach incorporating comprehensive information from MI-EEG signals. Neural Netw 2023; 167:183-198. [PMID: 37659115 DOI: 10.1016/j.neunet.2023.08.008] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2023] [Revised: 07/24/2023] [Accepted: 08/06/2023] [Indexed: 09/04/2023]
Abstract
Non-stationarity of EEG signals leads to high variability between subjects, making it challenging to directly use data from other subjects (source domain) for the classifier in the current subject (target domain). In this study, we propose MI-DAGSC to address domain adaptation challenges in EEG-based motor imagery (MI) decoding. By combining domain-level information, class-level information, and inter-sample structure information, our model effectively aligns the feature distributions of source and target domains. This work is an extension of our previous domain adaptation work MI-DABAN (Li et al., 2023). Based on MI-DABAN, MI-DAGSC designs Sample-Feature Blocks (SFBs) and Graph Convolution Blocks (GCBs) to focus on intra-sample and inter-sample information. The synergistic integration of SFBs and GCBs enable the model to capture comprehensive information and understand the relationship between samples, thus improving representation learning. Furthermore, we introduce a triplet loss to enhance the alignment and compactness of feature representations. Extensive experiments on real EEG datasets demonstrate the effectiveness of MI-DAGSC, confirming that our method makes a valuable contribution to the MI-EEG decoding. Moreover, it holds great potential for various applications in brain-computer interface systems and neuroscience research. And the code of the proposed architecture in this study is available under https://github.com/zhangdx21/MI-DAGSC.
Collapse
Affiliation(s)
- Dongxue Zhang
- Jilin University, College of Computer Science and Technology, Changchun, Jilin Province, China; Key Laboratory of Symbol Computation and Knowledge Engineering, Jilin University, Changchun 130012, China.
| | - Huiying Li
- Jilin University, College of Computer Science and Technology, Changchun, Jilin Province, China; Key Laboratory of Symbol Computation and Knowledge Engineering, Jilin University, Changchun 130012, China.
| | - Jingmeng Xie
- Xi'an Jiaotong University, College of Electronic information, Xi'an, Shanxi Province, China.
| | - Dajun Li
- Jilin Provincial People's Hospital, Changchun, Jilin Province, China
| |
Collapse
|
22
|
Cui J, Yuan L, Wang Z, Li R, Jiang T. Towards best practice of interpreting deep learning models for EEG-based brain computer interfaces. Front Comput Neurosci 2023; 17:1232925. [PMID: 37663037 PMCID: PMC10470463 DOI: 10.3389/fncom.2023.1232925] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 07/24/2023] [Indexed: 09/05/2023] Open
Abstract
Introduction As deep learning has achieved state-of-the-art performance for many tasks of EEG-based BCI, many efforts have been made in recent years trying to understand what have been learned by the models. This is commonly done by generating a heatmap indicating to which extent each pixel of the input contributes to the final classification for a trained model. Despite the wide use, it is not yet understood to which extent the obtained interpretation results can be trusted and how accurate they can reflect the model decisions. Methods We conduct studies to quantitatively evaluate seven different deep interpretation techniques across different models and datasets for EEG-based BCI. Results The results reveal the importance of selecting a proper interpretation technique as the initial step. In addition, we also find that the quality of the interpretation results is inconsistent for individual samples despite when a method with an overall good performance is used. Many factors, including model structure and dataset types, could potentially affect the quality of the interpretation results. Discussion Based on the observations, we propose a set of procedures that allow the interpretation results to be presented in an understandable and trusted way. We illustrate the usefulness of our method for EEG-based BCI with instances selected from different scenarios.
Collapse
Affiliation(s)
- Jian Cui
- Research Center for Augmented Intelligence, Research Institute of Artificial Intelligence, Zhejiang Lab, Hangzhou, China
| | - Liqiang Yuan
- School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Zhaoxiang Wang
- Research Center for Augmented Intelligence, Research Institute of Artificial Intelligence, Zhejiang Lab, Hangzhou, China
| | - Ruilin Li
- School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Tianzi Jiang
- Research Center for Augmented Intelligence, Research Institute of Artificial Intelligence, Zhejiang Lab, Hangzhou, China
- Brainnetome Center, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
23
|
Dong Y, Wen X, Gao F, Gao C, Cao R, Xiang J, Cao R. Subject-Independent EEG Classification of Motor Imagery Based on Dual-Branch Feature Fusion. Brain Sci 2023; 13:1109. [PMID: 37509039 PMCID: PMC10377689 DOI: 10.3390/brainsci13071109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2023] [Revised: 07/13/2023] [Accepted: 07/19/2023] [Indexed: 07/30/2023] Open
Abstract
A brain computer interface (BCI) system helps people with motor dysfunction interact with the external environment. With the advancement of technology, BCI systems have been applied in practice, but their practicability and usability are still greatly challenged. A large amount of calibration time is often required before BCI systems are used, which can consume the patient's energy and easily lead to anxiety. This paper proposes a novel motion-assisted method based on a novel dual-branch multiscale auto encoder network (MSAENet) to decode human brain motion imagery intentions, while introducing a central loss function to compensate for the shortcomings of traditional classifiers that only consider inter-class differences and ignore intra-class coupling. The effectiveness of the method is validated on three datasets, namely BCIIV2a, SMR-BCI and OpenBMI, to achieve zero calibration of the MI-BCI system. The results show that our proposed network displays good results on all three datasets. In the case of subject-independence, the MSAENet outperformed the other four comparison methods on the BCIIV2a and SMR-BCI datasets, while achieving F1_score values as high as 69.34% on the OpenBMI dataset. Our method maintains better classification accuracy with a small number of parameters and short prediction times, and the method achieves zero calibration of the MI-BCI system.
Collapse
Affiliation(s)
- Yanqing Dong
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Xin Wen
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Fang Gao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Chengxin Gao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Ruochen Cao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| | - Jie Xiang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan 030024, China
| | - Rui Cao
- School of Software, Taiyuan University of Technology, Taiyuan 030024, China
| |
Collapse
|
24
|
Nam H, Kim JM, Choi W, Bak S, Kam TE. The effects of layer-wise relevance propagation-based feature selection for EEG classification: a comparative study on multiple datasets. Front Hum Neurosci 2023; 17:1205881. [PMID: 37342822 PMCID: PMC10277566 DOI: 10.3389/fnhum.2023.1205881] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Accepted: 05/17/2023] [Indexed: 06/23/2023] Open
Abstract
Introduction The brain-computer interface (BCI) allows individuals to control external devices using their neural signals. One popular BCI paradigm is motor imagery (MI), which involves imagining movements to induce neural signals that can be decoded to control devices according to the user's intention. Electroencephalography (EEG) is frequently used for acquiring neural signals from the brain in the fields of MI-BCI due to its non-invasiveness and high temporal resolution. However, EEG signals can be affected by noise and artifacts, and patterns of EEG signals vary across different subjects. Therefore, selecting the most informative features is one of the essential processes to enhance classification performance in MI-BCI. Methods In this study, we design a layer-wise relevance propagation (LRP)-based feature selection method which can be easily integrated into deep learning (DL)-based models. We assess its effectiveness for reliable class-discriminative EEG feature selection on two different publicly available EEG datasets with various DL-based backbone models in the subject-dependent scenario. Results and discussion The results show that LRP-based feature selection enhances the performance for MI classification on both datasets for all DL-based backbone models. Based on our analysis, we believe that it can broad its capability to different research domains.
Collapse
Affiliation(s)
| | | | | | | | - Tae-Eui Kam
- Department of Artificial Intelligence, Korea University, Seoul, Republic of Korea
| |
Collapse
|
25
|
García-Murillo DG, Álvarez-Meza AM, Castellanos-Dominguez CG. KCS-FCnet: Kernel Cross-Spectral Functional Connectivity Network for EEG-Based Motor Imagery Classification. Diagnostics (Basel) 2023; 13:diagnostics13061122. [PMID: 36980430 PMCID: PMC10046910 DOI: 10.3390/diagnostics13061122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2023] [Revised: 02/25/2023] [Accepted: 03/02/2023] [Indexed: 03/18/2023] Open
Abstract
This paper uses EEG data to introduce an approach for classifying right and left-hand classes in Motor Imagery (MI) tasks. The Kernel Cross-Spectral Functional Connectivity Network (KCS-FCnet) method addresses these limitations by providing richer spatial-temporal-spectral feature maps, a simpler architecture, and a more interpretable approach for EEG-driven MI discrimination. In particular, KCS-FCnet uses a single 1D-convolutional-based neural network to extract temporal-frequency features from raw EEG data and a cross-spectral Gaussian kernel connectivity layer to model channel functional relationships. As a result, the functional connectivity feature map reduces the number of parameters, improving interpretability by extracting meaningful patterns related to MI tasks. These patterns can be adapted to the subject’s unique characteristics. The validation results prove that introducing KCS-FCnet shallow architecture is a promising approach for EEG-based MI classification with the potential for real-world use in brain–computer interface systems.
Collapse
|
26
|
Chen J, Wang D, Yi W, Xu M, Tan X. Filter bank sinc-convolutional network with channel self-attention for high performance motor imagery decoding. J Neural Eng 2023; 20. [PMID: 36763992 DOI: 10.1088/1741-2552/acbb2c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2022] [Accepted: 02/10/2023] [Indexed: 02/12/2023]
Abstract
Objective.Motor Imagery Brain-Computer Interface (MI-BCI) is an active Brain-Computer Interface (BCI) paradigm focusing on the identification of motor intention, which is one of the most important non-invasive BCI paradigms. In MI-BCI studies, deep learning-based methods (especially lightweight networks) have attracted more attention in recent years, but the decoding performance still needs further improving.Approach.To solve this problem, we designed a filter bank structure with sinc-convolutional layers for spatio-temporal feature extraction of MI-electroencephalography in four motor rhythms. The Channel Self-Attention method was introduced for feature selection based on both global and local information, so as to build a model called Filter Bank Sinc-convolutional Network with Channel Self-Attention for high performance MI-decoding. Also, we proposed a data augmentation method based on multivariate empirical mode decomposition to improve the generalization capability of the model.Main results.We performed an intra-subject evaluation experiment on unseen data of three open MI datasets. The proposed method achieved mean accuracy of 78.20% (4-class scenario) on BCI Competition IV IIa, 87.34% (2-class scenario) on BCI Competition IV IIb, and 72.03% (2-class scenario) on Open Brain Machine Interface (OpenBMI) dataset, which are significantly higher than those of compared deep learning-based methods by at least 3.05% (p= 0.0469), 3.18% (p= 0.0371), and 2.27% (p= 0.0024) respectively.Significance.This work provides a new option for deep learning-based MI decoding, which can be employed for building BCI systems for motor rehabilitation.
Collapse
Affiliation(s)
- Jiaming Chen
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Dan Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Weibo Yi
- Beijing Machine and Equipment Institute, Beijing, People's Republic of China
| | - Meng Xu
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Xiyue Tan
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| |
Collapse
|
27
|
Jaipriya D, Sriharipriya KC. A comparative analysis of masking empirical mode decomposition and a neural network with feed-forward and back propagation along with masking empirical mode decomposition to improve the classification performance for a reliable brain-computer interface. Front Comput Neurosci 2022; 16:1010770. [PMID: 36405787 PMCID: PMC9672820 DOI: 10.3389/fncom.2022.1010770] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Accepted: 10/03/2022] [Indexed: 02/25/2024] Open
Abstract
In general, extraction and classification are used in various fields like image processing, pattern recognition, signal processing, and so on. Extracting effective characteristics from raw electroencephalogram (EEG) signals is a crucial role of the brain-computer interface for motor imagery. Recently, there has been a great deal of focus on motor imagery in the EEG signals since they encode a person's intent to do an action. Researchers have been using MI signals to assist paralyzed people and even move them on their own with certain equipment, like wheelchairs. As a result, proper decoding is an important step required for the interconnection of the brain and the computer. EEG decoding is a challenging process because of poor SNR, complexity, and other reasons. However, choosing an appropriate method to extract the features to improve the performance of motor imagery recognition is still a research hotspot. To extract the features of the EEG signal in the classification task, this paper proposes a Masking Empirical Mode Decomposition (MEMD) based Feed Forward Back Propagation Neural Network (MEMD-FFBPNN). The dataset consists of EEG signals which are first normalized using the minimax method and given as input to the MEMD to extract the features and then given to the FFBPNN to classify the tasks. The accuracy of the proposed method MEMD-FFBPNN has been measured using the confusion matrix, mean square error and which has been recorded up to 99.9%. Thus, the proposed method gives better accuracy than the other conventional methods.
Collapse
Affiliation(s)
| | - K. C. Sriharipriya
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, India
| |
Collapse
|
28
|
Image-Based Learning Using Gradient Class Activation Maps for Enhanced Physiological Interpretability of Motor Imagery Skills. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12031695] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Brain activity stimulated by the motor imagery paradigm (MI) is measured by Electroencephalography (EEG), which has several advantages to be implemented with the widely used Brain–Computer Interfaces (BCIs) technology. However, the substantial inter/intra variability of recorded data significantly influences individual skills on the achieved performance. This study explores the ability to distinguish between MI tasks and the interpretability of the brain’s ability to produce elicited mental responses with improved accuracy. We develop a Deep and Wide Convolutional Neuronal Network fed by a set of topoplots extracted from the multichannel EEG data. Further, we perform a visualization technique based on gradient-based class activation maps (namely, GradCam++) at different intervals along the MI paradigm timeline to account for intra-subject variability in neural responses over time. We also cluster the dynamic spatial representation of the extracted maps across the subject set to come to a deeper understanding of MI-BCI coordination skills. According to the results obtained from the evaluated GigaScience Database of motor-evoked potentials, the developed approach enhances the physiological explanation of motor imagery in aspects such as neural synchronization between rhythms, brain lateralization, and the ability to predict the MI onset responses and their evolution during training sessions.
Collapse
|
29
|
Lee DY, Jeong JH, Lee BH, Lee SW. Motor Imagery Classification Using Inter-Task Transfer Learning via A Channel-Wise Variational Autoencoder-based Convolutional Neural Network. IEEE Trans Neural Syst Rehabil Eng 2022; 30:226-237. [PMID: 35041605 DOI: 10.1109/tnsre.2022.3143836] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Highly sophisticated control based on a brain-computer interface (BCI) requires decoding kinematic information from brain signals. The forearm is a region of the upper limb that is often used in everyday life, but intuitive movements within the same limb have rarely been investigated in previous BCI studies. In this study, we focused on various forearm movement decoding from electroencephalography (EEG) signals using a small number of samples. Ten healthy participants took part in an experiment and performed motor execution (ME) and motor imagery (MI) of the intuitive movement tasks (Dataset I). We propose a convolutional neural network using a channel-wise variational autoencoder (CVNet) based on inter-task transfer learning. We approached that training the reconstructed ME-EEG signals together will also achieve more sufficient classification performance with only a small amount of MI-EEG signals. The proposed CVNet was validated on our own Dataset I and a public dataset, BNCI Horizon 2020 (Dataset II). The classification accuracies of various movements are confirmed to be 0.83 (±0.04) and 0.69 (±0.04) for Dataset I and II, respectively. The results show that the proposed method exhibits performance increases of approximately 0.09~0.27 and 0.08~0.24 compared with the conventional models for Dataset I and II, respectively. The outcomes suggest that the training model for decoding imagined movements can be performed using data from ME and a small number of data samples from MI. Hence, it is presented the feasibility of BCI learning strategies that can sufficiently learn deep learning with a few amount of calibration dataset and time only, with stable performance.
Collapse
|
30
|
Wang X, Cavigelli L, Schneider T, Benini L. Sub-100 μW Multispectral Riemannian Classification for EEG-Based Brain-Machine Interfaces. IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS 2021; 15:1149-1160. [PMID: 34932486 DOI: 10.1109/tbcas.2021.3137290] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Motor imagery (MI) brain-machine interfaces (BMIs) enable us to control machines by merely thinking of performing a motor action. Practical use cases require a wearable solution where the classification of the brain signals is done locally near the sensor using machine learning models embedded on energy-efficient microcontroller units (MCUs), for assured privacy, user comfort, and long-term usage. In this work, we provide practical insights on the accuracy-cost trade-off for embedded BMI solutions. Our multispectral Riemannian classifier reaches 75.1% accuracy on a 4-class MI task. The accuracy is further improved by tuning different types of classifiers to each subject, achieving 76.4%. We further scale down the model by quantizing it to mixed-precision representations with a minimal accuracy loss of 1% and 1.4%, respectively, which is still up to 4.1% more accurate than the state-of-the-art embedded convolutional neural network. We implement the model on a low-power MCU within an energy budget of merely 198 μJ and taking only 16.9 ms per classification. Classifying samples continuously, overlapping the 3.5 s samples by 50% to avoid missing user inputs allows for operation at just 85 μW. Compared to related works in embedded MI-BMIs, our solution sets the new state-of-the-art in terms of accuracy-energy trade-off for near-sensor classification.
Collapse
|
31
|
|
32
|
Altaheri H, Muhammad G, Alsulaiman M, Amin SU, Altuwaijri GA, Abdul W, Bencherif MA, Faisal M. Deep learning techniques for classification of electroencephalogram (EEG) motor imagery (MI) signals: a review. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06352-5] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/06/2023]
|
33
|
Singh A, Hussain AA, Lal S, Guesgen HW. A Comprehensive Review on Critical Issues and Possible Solutions of Motor Imagery Based Electroencephalography Brain-Computer Interface. SENSORS 2021; 21:s21062173. [PMID: 33804611 PMCID: PMC8003721 DOI: 10.3390/s21062173] [Citation(s) in RCA: 36] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/28/2020] [Revised: 03/15/2021] [Accepted: 03/16/2021] [Indexed: 01/16/2023]
Abstract
Motor imagery (MI) based brain–computer interface (BCI) aims to provide a means of communication through the utilization of neural activity generated due to kinesthetic imagination of limbs. Every year, a significant number of publications that are related to new improvements, challenges, and breakthrough in MI-BCI are made. This paper provides a comprehensive review of the electroencephalogram (EEG) based MI-BCI system. It describes the current state of the art in different stages of the MI-BCI (data acquisition, MI training, preprocessing, feature extraction, channel and feature selection, and classification) pipeline. Although MI-BCI research has been going for many years, this technology is mostly confined to controlled lab environments. We discuss recent developments and critical algorithmic issues in MI-based BCI for commercial deployment.
Collapse
|