1
|
Lopes F, Pinto MF, Dourado A, Schulze-Bonhage A, Dümpelmann M, Teixeira C. Addressing data limitations in seizure prediction through transfer learning. Sci Rep 2024; 14:14169. [PMID: 38898066 PMCID: PMC11187122 DOI: 10.1038/s41598-024-64802-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2024] [Accepted: 06/13/2024] [Indexed: 06/21/2024] Open
Abstract
According to the literature, seizure prediction models should be developed following a patient-specific approach. However, seizures are usually very rare events, meaning the number of events that may be used to optimise seizure prediction approaches is limited. To overcome such constraint, we analysed the possibility of using data from patients from an external database to improve patient-specific seizure prediction models. We present seizure prediction models trained using a transfer learning procedure. We trained a deep convolutional autoencoder using electroencephalogram data from 41 patients collected from the EPILEPSIAE database. Then, a bidirectional long short-term memory and a classifier layers were added on the top of the encoder part and were optimised for 24 patients from the Universitätsklinikum Freiburg individually. The encoder was used as a feature extraction module. Therefore, its weights were not changed during the patient-specific training. Experimental results showed that seizure prediction models optimised using pretrained weights present about four times fewer false alarms while maintaining the same ability to predict seizures and achieved more 13% validated patients. Therefore, results evidenced that the optimisation using transfer learning was more stable and faster, saving computational resources. In summary, adopting transfer learning for seizure prediction models represents a significant advancement. It addresses the data limitation seen in the seizure prediction field and offers more efficient and stable training, conserving computational resources. Additionally, despite the compact size, transfer learning allows to easily share data knowledge due to fewer ethical restrictions and lower storage requirements. The convolutional autoencoder developed in this study will be shared with the scientific community, promoting further research.
Collapse
Affiliation(s)
- Fábio Lopes
- Department of Informatics Engineering, Center for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, Portugal.
- Department Neurosurgery, Epilepsy Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany.
| | - Mauro F Pinto
- Department of Informatics Engineering, Center for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, Portugal
| | - António Dourado
- Department of Informatics Engineering, Center for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, Portugal
| | - Andreas Schulze-Bonhage
- Department Neurosurgery, Epilepsy Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
| | - Matthias Dümpelmann
- Department Neurosurgery, Epilepsy Center, Medical Center - University of Freiburg, Faculty of Medicine, University of Freiburg, Freiburg, Germany
- Department of Microsystems Engineering (IMTEK), University of Freiburg, Freiburg, Germany
| | - César Teixeira
- Department of Informatics Engineering, Center for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, Portugal
| |
Collapse
|
2
|
Junqueira B, Aristimunha B, Chevallier S, de Camargo RY. A systematic evaluation of Euclidean alignment with deep learning for EEG decoding. J Neural Eng 2024; 21:036038. [PMID: 38776898 DOI: 10.1088/1741-2552/ad4f18] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 05/22/2024] [Indexed: 05/25/2024]
Abstract
Objective:Electroencephalography signals are frequently used for various Brain-Computer interface (BCI) tasks. While deep learning (DL) techniques have shown promising results, they are hindered by the substantial data requirements. By leveraging data from multiple subjects, transfer learning enables more effective training of DL models. A technique that is gaining popularity is Euclidean alignment (EA) due to its ease of use, low computational complexity, and compatibility with DL models. However, few studies evaluate its impact on the training performance of shared and individual DL models. In this work, we systematically evaluate the effect of EA combined with DL for decoding BCI signals.Approach:We used EA as a pre-processing step to train shared DL models with data from multiple subjects and evaluated their transferability to new subjects.Main results:Our experimental results show that it improves decoding in the target subject by 4.33% and decreases convergence time by more than 70%. We also trained individual models for each subject to use as a majority-voting ensemble classifier. In this scenario, using EA improved the 3-model ensemble accuracy by 3.71%. However, when compared to the shared model with EA, the ensemble accuracy was 3.62% lower.Significance:EA succeeds in the task of improving transfer learning performance with DL models and, could be used as a standard pre-processing technique.
Collapse
Affiliation(s)
- Bruna Junqueira
- University of São Paulo, Sao Paulo, Brazil
- Université Paris-Saclay, Inria TAU team, LISN-CNRS, Orsay, France
| | - Bruno Aristimunha
- Université Paris-Saclay, Inria TAU team, LISN-CNRS, Orsay, France
- Federal University of ABC, Santo Andre, Brazil
| | | | | |
Collapse
|
3
|
Pfeffer MA, Ling SSH, Wong JKW. Exploring the frontier: Transformer-based models in EEG signal analysis for brain-computer interfaces. Comput Biol Med 2024; 178:108705. [PMID: 38865781 DOI: 10.1016/j.compbiomed.2024.108705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2023] [Revised: 05/01/2024] [Accepted: 06/01/2024] [Indexed: 06/14/2024]
Abstract
This review systematically explores the application of transformer-based models in EEG signal processing and brain-computer interface (BCI) development, with a distinct focus on ensuring methodological rigour and adhering to empirical validations within the existing literature. By examining various transformer architectures, such as the Temporal Spatial Transformer Network (TSTN) and EEG Conformer, this review delineates their capabilities in mitigating challenges intrinsic to EEG data, such as noise and artifacts, and their subsequent implications on decoding and classification accuracies across disparate mental tasks. The analytical scope extends to a meticulous examination of attention mechanisms within transformer models, delineating their role in illuminating critical temporal and spatial EEG features and facilitating interpretability in model decision-making processes. The discourse additionally encapsulates emerging works that substantiate the efficacy of transformer models in noise reduction of EEG signals and diversifying applications beyond the conventional motor imagery paradigm. Furthermore, this review elucidates evident gaps and propounds exploratory avenues in the applications of pre-trained transformers in EEG analysis and the potential expansion into real-time and multi-task BCI applications. Collectively, this review distils extant knowledge, navigates through the empirical findings, and puts forward a structured synthesis, thereby serving as a conduit for informed future research endeavours in transformer-enhanced, EEG-based BCI systems.
Collapse
Affiliation(s)
- Maximilian Achim Pfeffer
- Faculty of Engineering and Information Technology, University of Technology Sydney, CB11 81-113, Broadway, Ultimo, 2007, New South Wales, Australia.
| | - Steve Sai Ho Ling
- Faculty of Engineering and Information Technology, University of Technology Sydney, CB11 81-113, Broadway, Ultimo, 2007, New South Wales, Australia.
| | - Johnny Kwok Wai Wong
- Faculty of Design, Architecture and Building, University of Technology Sydney, 15 Broadway, Ultimo, 2007, New South Wales, Australia.
| |
Collapse
|
4
|
Yuen B, Dong X, Lu T. A 3D ray traced biological neural network learning model. Nat Commun 2024; 15:4693. [PMID: 38824154 DOI: 10.1038/s41467-024-48747-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Accepted: 05/13/2024] [Indexed: 06/03/2024] Open
Abstract
Training large neural networks on big datasets requires significant computational resources and time. Transfer learning reduces training time by pre-training a base model on one dataset and transferring the knowledge to a new model for another dataset. However, current choices of transfer learning algorithms are limited because the transferred models always have to adhere to the dimensions of the base model and can not easily modify the neural architecture to solve other datasets. On the other hand, biological neural networks (BNNs) are adept at rearranging themselves to tackle completely different problems using transfer learning. Taking advantage of BNNs, we design a dynamic neural network that is transferable to any other network architecture and can accommodate many datasets. Our approach uses raytracing to connect neurons in a three-dimensional space, allowing the network to grow into any shape or size. In the Alcala dataset, our transfer learning algorithm trains the fastest across changing environments and input sizes. In addition, we show that our algorithm also outperformance the state of the art in EEG dataset. In the future, this network may be considered for implementation on real biological neural networks to decrease power consumption.
Collapse
Affiliation(s)
- Brosnan Yuen
- Department of Electrical and Computer Engineering, University of Victoria, 3800 Finnerty Road, Victoria, V8P 5C2, BC, Canada
| | - Xiaodai Dong
- Department of Electrical and Computer Engineering, University of Victoria, 3800 Finnerty Road, Victoria, V8P 5C2, BC, Canada.
| | - Tao Lu
- Department of Electrical and Computer Engineering, University of Victoria, 3800 Finnerty Road, Victoria, V8P 5C2, BC, Canada.
| |
Collapse
|
5
|
Li J, She Q, Meng M, Du S, Zhang Y. Three-stage transfer learning for motor imagery EEG recognition. Med Biol Eng Comput 2024; 62:1689-1701. [PMID: 38342784 DOI: 10.1007/s11517-024-03036-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 01/17/2024] [Indexed: 02/13/2024]
Abstract
Motor imagery (MI) paradigms have been widely used in neural rehabilitation and drowsiness state assessment. The progress in brain-computer interface (BCI) technology has emphasized the importance of accurately and efficiently detecting motor imagery intentions from electroencephalogram (EEG). Despite the recent breakthroughs made in developing EEG-based algorithms for decoding MI, the accuracy and efficiency of these models remain limited by technical challenges posed by cross-subject heterogeneity in EEG data processing and the scarcity of EEG data for training. Inspired by the optimal transport theory, this study aims to develop a novel three-stage transfer learning (TSTL) method, which uses the existing labeled data from a source domain to improve classification performance on an unlabeled target domain. Notably, the proposed method comprises three components, namely, the Riemannian tangent space mapping (RTSM), source domain transformer (SDT), and optimal subspace mapping (OSM). The RTSM maps a symmetric positive definite matrix from the Riemannian space to the tangent space to minimize the marginal probability distribution drift. The SDT transforms the source domain to a target domain by finding the optimal transport mapping matrix to reduce the joint probability distribution differences. The OSM finally maps the transformed source domain and original target domain to the same subspace to further mitigate the distribution discrepancy. The performance of the proposed method was validated on two public BCI datasets, and the average accuracy of the algorithm on two datasets was 72.24% and 69.29%. Our results demonstrated the improved performance of EEG-based MI detection in comparison with state-of-the-art algorithms.
Collapse
Affiliation(s)
- Junhao Li
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang, China
| | - Qingshan She
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang, China.
| | - Ming Meng
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang, China
| | - Shengzhi Du
- Department of Electrical Engineering, Tshwane University of Technology, Pretoria, 0001, South Africa
| | - Yingchun Zhang
- Department of Biomedical Engineering, University of Houston, Houston, TX, 77204, USA
| |
Collapse
|
6
|
Shuqfa Z, Lakas A, Belkacem AN. Increasing accessibility to a large brain-computer interface dataset: Curation of physionet EEG motor movement/imagery dataset for decoding and classification. Data Brief 2024; 54:110181. [PMID: 38586146 PMCID: PMC10998040 DOI: 10.1016/j.dib.2024.110181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2023] [Revised: 01/22/2024] [Accepted: 02/05/2024] [Indexed: 04/09/2024] Open
Abstract
A reliable motor imagery (MI) brain-computer interface (BCI) requires accurate decoding, which in turn requires model calibration using electroencephalography (EEG) signals from subjects executing or imagining the execution of movements. Although the PhysioNet EEG Motor Movement/Imagery Dataset is currently the largest EEG dataset in the literature, relatively few studies have used it to decode MI trials. In the present study, we curated and cleaned this dataset to store it in an accessible format that is convenient for quick exploitation, decoding, and classification using recent integrated development environments. We dropped six subjects owing to anomalies in EEG recordings and pre-possessed the rest, resulting in 103 subjects spanning four MI and four motor execution tasks. The annotations were coded to correspond to different tasks using numerical values. The resulting dataset is stored in both MATLAB structure and CSV files to ensure ease of access and organization. We believe that improving the accessibility of this dataset will help EEG-based MI-BCI decoding and classification, enabling more reliable real-life applications. The convenience and ease of access of this dataset may therefore lead to improvements in cross-subject classification and transfer learning.
Collapse
Affiliation(s)
- Zaid Shuqfa
- Connected Autonomous Intelligent Systems Lab, Department of Computer and Network Engineering, College of IT (CIT), United Arab Emirates University (UAEU), Al Ain City 15551, the United Arab Emirates
- Rabdan Academy, P.O. Box 114646, Abu Dhabi, the United Arab Emirates
| | - Abderrahmane Lakas
- Connected Autonomous Intelligent Systems Lab, Department of Computer and Network Engineering, College of IT (CIT), United Arab Emirates University (UAEU), Al Ain City 15551, the United Arab Emirates
| | - Abdelkader Nasreddine Belkacem
- Connected Autonomous Intelligent Systems Lab, Department of Computer and Network Engineering, College of IT (CIT), United Arab Emirates University (UAEU), Al Ain City 15551, the United Arab Emirates
| |
Collapse
|
7
|
Nia AF, Tang V, Talou GM, Billinghurst M. Synthesizing affective neurophysiological signals using generative models: A review paper. J Neurosci Methods 2024; 406:110129. [PMID: 38614286 DOI: 10.1016/j.jneumeth.2024.110129] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 01/04/2024] [Accepted: 04/03/2024] [Indexed: 04/15/2024]
Abstract
The integration of emotional intelligence in machines is an important step in advancing human-computer interaction. This demands the development of reliable end-to-end emotion recognition systems. However, the scarcity of public affective datasets presents a challenge. In this literature review, we emphasize the use of generative models to address this issue in neurophysiological signals, particularly Electroencephalogram (EEG) and Functional Near-Infrared Spectroscopy (fNIRS). We provide a comprehensive analysis of different generative models used in the field, examining their input formulation, deployment strategies, and methodologies for evaluating the quality of synthesized data. This review serves as a comprehensive overview, offering insights into the advantages, challenges, and promising future directions in the application of generative models in emotion recognition systems. Through this review, we aim to facilitate the progression of neurophysiological data augmentation, thereby supporting the development of more efficient and reliable emotion recognition systems.
Collapse
Affiliation(s)
- Alireza F Nia
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand.
| | - Vanessa Tang
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand
| | - Gonzalo Maso Talou
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand
| | - Mark Billinghurst
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand
| |
Collapse
|
8
|
Hamzah HA, Abdalla KK. EEG-based emotion recognition systems; comprehensive study. Heliyon 2024; 10:e31485. [PMID: 38818173 PMCID: PMC11137547 DOI: 10.1016/j.heliyon.2024.e31485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 05/16/2024] [Indexed: 06/01/2024] Open
Abstract
Emotion recognition technology through EEG signal analysis is currently a fundamental concept in artificial intelligence. This recognition has major practical implications in emotional health care, human-computer interaction, and so on. This paper provides a comprehensive study of different methods for extracting electroencephalography (EEG) features for emotion recognition from four different perspectives, including time domain features, frequency domain features, time-frequency features, and nonlinear features. We summarize the current pattern recognition methods adopted in most related works, and with the rapid development of deep learning (DL) attracting the attention of researchers in this field, we pay more attention to deep learning-based studies and analyse the characteristics, advantages, disadvantages, and applicable scenarios. Finally, the current challenges and future development directions in this field were summarized. This paper can help novice researchers in this field gain a systematic understanding of the current status of emotion recognition research based on EEG signals and provide ideas for subsequent related research.
Collapse
Affiliation(s)
- Hussein Ali Hamzah
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| | - Kasim K. Abdalla
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| |
Collapse
|
9
|
Chen SY, Chang CM, Chiang KJ, Wei CS. SSVEP-DAN: Cross-Domain Data Alignment for SSVEP-Based Brain-Computer Interfaces. IEEE Trans Neural Syst Rehabil Eng 2024; 32:2027-2037. [PMID: 38781061 DOI: 10.1109/tnsre.2024.3404432] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/25/2024]
Abstract
Steady-state visual-evoked potential (SSVEP)-based brain-computer interfaces (BCIs) offer a non-invasive means of communication through high-speed speller systems. However, their efficiency is highly dependent on individual training data acquired during time-consuming calibration sessions. To address the challenge of data insufficiency in SSVEP-based BCIs, we introduce SSVEP-DAN, the first dedicated neural network model designed to align SSVEP data across different domains, encompassing various sessions, subjects, or devices. Our experimental results demonstrate the ability of SSVEP-DAN to transform existing source SSVEP data into supplementary calibration data. This results in a significant improvement in SSVEP decoding accuracy while reducing the calibration time. We envision SSVEP-DAN playing a crucial role in future applications of high-performance SSVEP-based BCIs. The source code for this work is available at: https://github.com/CECNL/SSVEP-DAN.
Collapse
|
10
|
Liang Y, Zhang C, An S, Wang Z, Shi K, Peng T, Ma Y, Xie X, He J, Zheng K. FetchEEG: a hybrid approach combining feature extraction and temporal-channel joint attention for EEG-based emotion classification. J Neural Eng 2024; 21:036011. [PMID: 38701773 DOI: 10.1088/1741-2552/ad4743] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Accepted: 05/03/2024] [Indexed: 05/05/2024]
Abstract
Objective. Electroencephalogram (EEG) analysis has always been an important tool in neural engineering, and the recognition and classification of human emotions are one of the important tasks in neural engineering. EEG data, obtained from electrodes placed on the scalp, represent a valuable resource of information for brain activity analysis and emotion recognition. Feature extraction methods have shown promising results, but recent trends have shifted toward end-to-end methods based on deep learning. However, these approaches often overlook channel representations, and their complex structures pose certain challenges to model fitting.Approach. To address these challenges, this paper proposes a hybrid approach named FetchEEG that combines feature extraction and temporal-channel joint attention. Leveraging the advantages of both traditional feature extraction and deep learning, the FetchEEG adopts a multi-head self-attention mechanism to extract representations between different time moments and channels simultaneously. The joint representations are then concatenated and classified using fully-connected layers for emotion recognition. The performance of the FetchEEG is verified by comparison experiments on a self-developed dataset and two public datasets.Main results. In both subject-dependent and subject-independent experiments, the FetchEEG demonstrates better performance and stronger generalization ability than the state-of-the-art methods on all datasets. Moreover, the performance of the FetchEEG is analyzed for different sliding window sizes and overlap rates in the feature extraction module. The sensitivity of emotion recognition is investigated for three- and five-frequency-band scenarios.Significance. FetchEEG is a novel hybrid method based on EEG for emotion classification, which combines EEG feature extraction with Transformer neural networks. It has achieved state-of-the-art performance on both self-developed datasets and multiple public datasets, with significantly higher training efficiency compared to end-to-end methods, demonstrating its effectiveness and feasibility.
Collapse
Affiliation(s)
- Yu Liang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Chenlong Zhang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Shan An
- JD Health International Inc., Beijing, People's Republic of China
| | - Zaitian Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Kaize Shi
- University of Technology Sydney, Sydney, Australia
| | - Tianhao Peng
- Beihang University, Beijing, People's Republic of China
| | - Yuqing Ma
- Beihang University, Beijing, People's Republic of China
| | - Xiaoyang Xie
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Jian He
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Kun Zheng
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| |
Collapse
|
11
|
Arulkumaran K, Di Vincenzo M, Dossa RFJ, Akiyama S, Ogawa Lillrank D, Sato M, Tomeoka K, Sasai S. A comparison of visual and auditory EEG interfaces for robot multi-stage task control. Front Robot AI 2024; 11:1329270. [PMID: 38783889 PMCID: PMC11111866 DOI: 10.3389/frobt.2024.1329270] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2023] [Accepted: 04/23/2024] [Indexed: 05/25/2024] Open
Abstract
Shared autonomy holds promise for assistive robotics, whereby physically-impaired people can direct robots to perform various tasks for them. However, a robot that is capable of many tasks also introduces many choices for the user, such as which object or location should be the target of interaction. In the context of non-invasive brain-computer interfaces for shared autonomy-most commonly electroencephalography-based-the two most common choices are to provide either auditory or visual stimuli to the user-each with their respective pros and cons. Using the oddball paradigm, we designed comparable auditory and visual interfaces to speak/display the choices to the user, and had users complete a multi-stage robotic manipulation task involving location and object selection. Users displayed differing competencies-and preferences-for the different interfaces, highlighting the importance of considering modalities outside of vision when constructing human-robot interfaces.
Collapse
|
12
|
Song M, Gwon D, Jun SC, Ahn M. Signal alignment for cross-datasets in P300 brain-computer interfaces. J Neural Eng 2024; 21:036007. [PMID: 38657615 DOI: 10.1088/1741-2552/ad430d] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2023] [Accepted: 04/24/2024] [Indexed: 04/26/2024]
Abstract
Objective.Transfer learning has become an important issue in the brain-computer interface (BCI) field, and studies on subject-to-subject transfer within the same dataset have been performed. However, few studies have been performed on dataset-to-dataset transfer, including paradigm-to-paradigm transfer. In this study, we propose a signal alignment (SA) for P300 event-related potential (ERP) signals that is intuitive, simple, computationally less expensive, and can be used for cross-dataset transfer learning.Approach.We proposed a linear SA that uses the P300's latency, amplitude scale, and reverse factor to transform signals. For evaluation, four datasets were introduced (two from conventional P300 Speller BCIs, one from a P300 Speller with face stimuli, and the last from a standard auditory oddball paradigm).Results.Although the standard approach without SA had an average precision (AP) score of 25.5%, the approach demonstrated a 35.8% AP score, and we observed that the number of subjects showing improvement was 36.0% on average. Particularly, we confirmed that the Speller dataset with face stimuli was more comparable with other datasets.Significance.We proposed a simple and intuitive way to align ERP signals that uses the characteristics of ERP signals. The results demonstrated the feasibility of cross-dataset transfer learning even between datasets with different paradigms.
Collapse
Affiliation(s)
- Minseok Song
- Department of Computer Science and Electrical Engineering, Handong Global University, Pohang, Republic of Korea
| | - Daeun Gwon
- Department of Computer Science and Electrical Engineering, Handong Global University, Pohang, Republic of Korea
| | - Sung Chan Jun
- School of Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology, Gwangju, Republic of Korea
- AI Graduate School, Gwangju Institute of Science and Technology, Gwangju, Republic of Korea
| | - Minkyu Ahn
- Department of Computer Science and Electrical Engineering, Handong Global University, Pohang, Republic of Korea
- School of Computer Science and Electrical Engineering, Handong Global University, Pohang, Republic of Korea
| |
Collapse
|
13
|
Mirjalili S, Duarte A. More than the sum of its parts: investigating episodic memory as a multidimensional cognitive process. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2024:2024.04.22.590651. [PMID: 38712266 PMCID: PMC11071378 DOI: 10.1101/2024.04.22.590651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2024]
Abstract
Why do we remember some events but forget others? Previous studies attempting to decode successful vs. unsuccessful brain states to investigate this question have met with limited success, potentially due, in part, to assessing episodic memory as a unidimensional process, despite evidence that multiple domains contribute to episodic encoding. Using a novel machine learning algorithm known as "transfer learning", we leveraged visual perception, sustained attention, and selective attention brain states to better predict episodic memory performance from trial-to-trial encoding electroencephalography (EEG) activity. We found that this multidimensional treatment of memory decoding improved prediction performance compared to traditional, unidimensional, methods, with each cognitive domain explaining unique variance in decoding of successful encoding-related neural activity. Importantly, this approach could be applied to cognitive domains outside of memory. Overall, this study provides critical insight into the underlying reasons why some events are remembered while others are not.
Collapse
|
14
|
Li W, Li H, Sun X, Kang H, An S, Wang G, Gao Z. Self-supervised contrastive learning for EEG-based cross-subject motor imagery recognition. J Neural Eng 2024; 21:026038. [PMID: 38565100 DOI: 10.1088/1741-2552/ad3986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2023] [Accepted: 03/20/2024] [Indexed: 04/04/2024]
Abstract
Objective. The extensive application of electroencephalography (EEG) in brain-computer interfaces (BCIs) can be attributed to its non-invasive nature and capability to offer high-resolution data. The acquisition of EEG signals is a straightforward process, but the datasets associated with these signals frequently exhibit data scarcity and require substantial resources for proper labeling. Furthermore, there is a significant limitation in the generalization performance of EEG models due to the substantial inter-individual variability observed in EEG signals.Approach. To address these issues, we propose a novel self-supervised contrastive learning framework for decoding motor imagery (MI) signals in cross-subject scenarios. Specifically, we design an encoder combining convolutional neural network and attention mechanism. In the contrastive learning training stage, the network undergoes training with the pretext task of data augmentation to minimize the distance between pairs of homologous transformations while simultaneously maximizing the distance between pairs of heterologous transformations. It enhances the amount of data utilized for training and improves the network's ability to extract deep features from original signals without relying on the true labels of the data.Main results. To evaluate our framework's efficacy, we conduct extensive experiments on three public MI datasets: BCI IV IIa, BCI IV IIb, and HGD datasets. The proposed method achieves cross-subject classification accuracies of 67.32%, 82.34%, and 81.13%on the three datasets, demonstrating superior performance compared to existing methods.Significance. Therefore, this method has great promise for improving the performance of cross-subject transfer learning in MI-based BCI systems.
Collapse
Affiliation(s)
- Wenjie Li
- Tianjin International Engineering Institute, Tianjin University, Tianjin 300072, People's Republic of China
| | - Haoyu Li
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Xinlin Sun
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| | - Huicong Kang
- Department of Neurology, Shanxi Bethune Hospital, Shanxi Academy of Medical Science, Tongji Shanxi Hospital, Third Hospital of Shanxi Medical University, Taiyuan 030000, People's Republic of China
- Department of Neurology, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan 430000, People's Republic of China
| | - Shan An
- JD Health International Inc., Beijing 100176, People's Republic of China
| | - Guoxin Wang
- JD Health International Inc., Beijing 100176, People's Republic of China
| | - Zhongke Gao
- School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, People's Republic of China
| |
Collapse
|
15
|
Lim RY, Lew WCL, Ang KK. Review of EEG Affective Recognition with a Neuroscience Perspective. Brain Sci 2024; 14:364. [PMID: 38672015 PMCID: PMC11048077 DOI: 10.3390/brainsci14040364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2024] [Revised: 04/02/2024] [Accepted: 04/06/2024] [Indexed: 04/28/2024] Open
Abstract
Emotions are a series of subconscious, fleeting, and sometimes elusive manifestations of the human innate system. They play crucial roles in everyday life-influencing the way we evaluate ourselves, our surroundings, and how we interact with our world. To date, there has been an abundance of research on the domains of neuroscience and affective computing, with experimental evidence and neural network models, respectively, to elucidate the neural circuitry involved in and neural correlates for emotion recognition. Recent advances in affective computing neural network models often relate closely to evidence and perspectives gathered from neuroscience to explain the models. Specifically, there has been growing interest in the area of EEG-based emotion recognition to adopt models based on the neural underpinnings of the processing, generation, and subsequent collection of EEG data. In this respect, our review focuses on providing neuroscientific evidence and perspectives to discuss how emotions potentially come forth as the product of neural activities occurring at the level of subcortical structures within the brain's emotional circuitry and the association with current affective computing models in recognizing emotions. Furthermore, we discuss whether such biologically inspired modeling is the solution to advance the field in EEG-based emotion recognition and beyond.
Collapse
Affiliation(s)
- Rosary Yuting Lim
- Institute for Infocomm Research, Agency for Science, Technology and Research, A*STAR, 1 Fusionopolis Way, #21-01 Connexis, Singapore 138632, Singapore; (R.Y.L.); (W.-C.L.L.)
| | - Wai-Cheong Lincoln Lew
- Institute for Infocomm Research, Agency for Science, Technology and Research, A*STAR, 1 Fusionopolis Way, #21-01 Connexis, Singapore 138632, Singapore; (R.Y.L.); (W.-C.L.L.)
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave., 32 Block N4 02a, Singapore 639798, Singapore
| | - Kai Keng Ang
- Institute for Infocomm Research, Agency for Science, Technology and Research, A*STAR, 1 Fusionopolis Way, #21-01 Connexis, Singapore 138632, Singapore; (R.Y.L.); (W.-C.L.L.)
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave., 32 Block N4 02a, Singapore 639798, Singapore
| |
Collapse
|
16
|
Ng HW, Guan C. Subject-independent meta-learning framework towards optimal training of EEG-based classifiers. Neural Netw 2024; 172:106108. [PMID: 38219680 DOI: 10.1016/j.neunet.2024.106108] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2023] [Revised: 11/13/2023] [Accepted: 01/05/2024] [Indexed: 01/16/2024]
Abstract
Advances in deep learning have shown great promise towards the application of performing high-accuracy Electroencephalography (EEG) signal classification in a variety of tasks. However, many EEG-based datasets are often plagued by the issue of high inter-subject signal variability. Robust deep learning models are notoriously difficult to train under such scenarios, often leading to subpar or widely varying performance across subjects under the leave-one-subject-out paradigm. Recently, the model agnostic meta-learning framework was introduced as a way to increase the model's ability to generalize towards new tasks. While the original framework focused on task-based meta-learning, this research aims to show that the meta-learning methodology can be modified towards subject-based signal classification while maintaining the same task objectives and achieve state-of-the-art performance. Namely, we propose the novel implementation of a few/zero-shot subject-independent meta-learning framework towards multi-class inner speech and binary class motor imagery classification. Compared to current subject-adaptive methods which utilize large number of labels from the target, the proposed framework shows its effectiveness in training zero-calibration and few-shot models for subject-independent EEG classification. The proposed few/zero-shot subject-independent meta-learning mechanism performs well on both small and large datasets and achieves robust, generalized performance across subjects. The results obtained shows a significant improvement over the current state-of-the-art, with the binary class motor imagery achieving 88.70% and the accuracy of multi-class inner speech achieving an average of 31.15%. Codes will be made available to public upon publication.
Collapse
Affiliation(s)
- Han Wei Ng
- Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore; AI Singapore, 3 Research Link, 117602, Singapore.
| | - Cuntai Guan
- Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore
| |
Collapse
|
17
|
Liu R, Chen Y, Li A, Ding Y, Yu H, Guan C. Aggregating intrinsic information to enhance BCI performance through federated learning. Neural Netw 2024; 172:106100. [PMID: 38232427 DOI: 10.1016/j.neunet.2024.106100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Revised: 11/20/2023] [Accepted: 01/03/2024] [Indexed: 01/19/2024]
Abstract
Insufficient data is a long-standing challenge for Brain-Computer Interface (BCI) to build a high-performance deep learning model. Though numerous research groups and institutes collect a multitude of EEG datasets for the same BCI task, sharing EEG data from multiple sites is still challenging due to the heterogeneity of devices. The significance of this challenge cannot be overstated, given the critical role of data diversity in fostering model robustness. However, existing works rarely discuss this issue, predominantly centering their attention on model training within a single dataset, often in the context of inter-subject or inter-session settings. In this work, we propose a hierarchical personalized Federated Learning EEG decoding (FLEEG) framework to surmount this challenge. This innovative framework heralds a new learning paradigm for BCI, enabling datasets with disparate data formats to collaborate in the model training process. Each client is assigned a specific dataset and trains a hierarchical personalized model to manage diverse data formats and facilitate information exchange. Meanwhile, the server coordinates the training procedure to harness knowledge gleaned from all datasets, thus elevating overall performance. The framework has been evaluated in Motor Imagery (MI) classification with nine EEG datasets collected by different devices but implementing the same MI task. Results demonstrate that the proposed framework can boost classification performance up to 8.4% by enabling knowledge sharing between multiple datasets, especially for smaller datasets. Visualization results also indicate that the proposed framework can empower the local models to put a stable focus on task-related areas, yielding better performance. To the best of our knowledge, this is the first end-to-end solution to address this important challenge.
Collapse
Affiliation(s)
- Rui Liu
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore.
| | - Yuanyuan Chen
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore.
| | - Anran Li
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore.
| | - Yi Ding
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore.
| | - Han Yu
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore.
| | - Cuntai Guan
- School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Ave, 639798, Singapore.
| |
Collapse
|
18
|
Han Y, Ke Y, Wang R, Wang T, Ming D. Enhancing SSVEP-BCI Performance Under Fatigue State Using Dynamic Stopping Strategy. IEEE Trans Neural Syst Rehabil Eng 2024; 32:1407-1415. [PMID: 38517720 DOI: 10.1109/tnsre.2024.3380635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/24/2024]
Abstract
Steady-state visual evoked potential (SSVEP)-based brain-computer interfaces (BCIs) have emerged as a prominent technology due to their high information transfer rate, rapid calibration time, and robust signal-to-noise ratio. However, a critical challenge for practical applications is performance degradation caused by user fatigue during prolonged use. This work proposes novel methods to address this challenge by dynamically adjusting data acquisition length and updating detection models based on a fatigue-aware stopping strategy. Two 16-target SSVEP-BCIs were employed, one using low-frequency and the other using high-frequency stimulation. A self-recorded fatigue dataset from 24 subjects was utilized for extensive evaluation. A simulated online experiment demonstrated that the proposed methods outperform the conventional fixed stopping strategy in terms of classification accuracy, information transfer rate, and selection time, irrespective of stimulation frequency. These findings suggest that the proposed approach can significantly improve SSVEP-BCI performance under fatigue conditions, leading to superior performance during extended use.
Collapse
|
19
|
Wang Z, Wang R, Chu Y, Chen G, Lin T, Jiang R, Wang J. A method to assess industrial paraffin contamination levels in rice and its transferability analysis based on transfer component analysis. Food Chem 2024; 436:137682. [PMID: 37837682 DOI: 10.1016/j.foodchem.2023.137682] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 09/19/2023] [Accepted: 10/04/2023] [Indexed: 10/16/2023]
Abstract
Accurate assessment of industrial paraffin contamination levels (IPCLs) in rice is critical for food safety. However, time-consuming and labor-intensive experiments to produce labels for targeted adulterated rice have hindered the development of IPCL estimation methods. In this paper, a transfer learning method (TCA-LSSVR) has been developed. The algorithm integrates transfer component analysis (TCA) with domain adaptive capabilities to produce accurate estimates. Rice from 7 different regions and 3 industrial paraffins were used to generate 4,680 samples from 9 datasets for benchmarking. The test results showed that the established algorithm achieved good estimation performance in various modelling strategies, and only 20 % of off-site samples were needed to supplement the source dataset, the average determination coefficient R2 reached 0.7045, the average RMSE reached 0.140 %, and the average RPD reached 2.023. This work highlights the prospect of rapidly developing a new generation of adulteration detection algorithms using only previous trial data.
Collapse
Affiliation(s)
- Zhentao Wang
- College of Engineering, Northeast Agricultural University, Harbin 150030, China; Heilongjiang Provincial Key Laboratory of Modern Agricultural Equipment Technology in Northern Cold Regions, Harbin 150030, China
| | - Ruidong Wang
- College of Engineering, Northeast Agricultural University, Harbin 150030, China
| | - Yuhang Chu
- College of Engineering, Northeast Agricultural University, Harbin 150030, China; Heilongjiang Provincial Key Laboratory of Modern Agricultural Equipment Technology in Northern Cold Regions, Harbin 150030, China
| | - Guoqing Chen
- College of Engineering, Northeast Agricultural University, Harbin 150030, China; Heilongjiang Provincial Key Laboratory of Modern Agricultural Equipment Technology in Northern Cold Regions, Harbin 150030, China
| | - Tenghui Lin
- College of Engineering, Northeast Agricultural University, Harbin 150030, China
| | - Rui Jiang
- College of Engineering, Northeast Agricultural University, Harbin 150030, China
| | - Jinfeng Wang
- College of Engineering, Northeast Agricultural University, Harbin 150030, China; Heilongjiang Provincial Key Laboratory of Modern Agricultural Equipment Technology in Northern Cold Regions, Harbin 150030, China.
| |
Collapse
|
20
|
Deng H, Li M, Zuo H, Zhou H, Qi E, Wu X, Xu G. Personalized motor imagery prediction model based on individual difference of ERP. J Neural Eng 2024; 21:016027. [PMID: 38359457 DOI: 10.1088/1741-2552/ad29d6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Accepted: 02/15/2024] [Indexed: 02/17/2024]
Abstract
Objective. Motor imagery-based brain-computer interaction (MI-BCI) is a novel method of achieving human and external environment interaction that can assist individuals with motor disorders to rehabilitate. However, individual differences limit the utility of the MI-BCI. In this study, a personalized MI prediction model based on the individual difference of event-related potential (ERP) is proposed to solve the MI individual difference.Approach.A novel paradigm named action observation-based multi-delayed matching posture task evokes ERP during a delayed matching posture task phase by retrieving picture stimuli and videos, and generates MI electroencephalogram through action observation and autonomous imagery in an action observation-based motor imagery phase. Based on the correlation between the ERP and MI, a logistic regression-based personalized MI prediction model is built to predict each individual's suitable MI action. 32 subjects conducted the MI task with or without the help of the prediction model to select the MI action. Then classification accuracy of the MI task is used to evaluate the proposed model and three traditional MI methods.Main results.The personalized MI prediction model successfully predicts suitable action among 3 sets of daily actions. Under suitable MI action, the individual's ERP amplitude and event-related desynchronization (ERD) intensity are the largest, which helps to improve the accuracy by 14.25%.Significance.The personalized MI prediction model that uses the temporal ERP features to predict the classification accuracy of MI is feasible for improving the individual's MI-BCI performance, providing a new personalized solution for the individual difference and practical BCI application.
Collapse
Affiliation(s)
- Haodong Deng
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, 300132 Tianjin, People's Republic of China
- School of Health Sciences and Biomedical Engineering, Hebei University of Technology, 300132 Tianjin, People's Republic of China
- Hebei Key Laboratory of Bioelectromagnetics and Neuroengineering, 300132 Tianjin, People's Republic of China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, 300132 Tianjin, People's Republic of China
| | - Mengfan Li
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, 300132 Tianjin, People's Republic of China
- School of Health Sciences and Biomedical Engineering, Hebei University of Technology, 300132 Tianjin, People's Republic of China
- Hebei Key Laboratory of Bioelectromagnetics and Neuroengineering, 300132 Tianjin, People's Republic of China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, 300132 Tianjin, People's Republic of China
| | - Haoxin Zuo
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, 300132 Tianjin, People's Republic of China
- School of Health Sciences and Biomedical Engineering, Hebei University of Technology, 300132 Tianjin, People's Republic of China
- Hebei Key Laboratory of Bioelectromagnetics and Neuroengineering, 300132 Tianjin, People's Republic of China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, 300132 Tianjin, People's Republic of China
| | - Huihui Zhou
- Peng Cheng Laboratory, 518000 Shenzhen, People's Republic of China
| | - Enming Qi
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, 300132 Tianjin, People's Republic of China
- School of Health Sciences and Biomedical Engineering, Hebei University of Technology, 300132 Tianjin, People's Republic of China
- Hebei Key Laboratory of Bioelectromagnetics and Neuroengineering, 300132 Tianjin, People's Republic of China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, 300132 Tianjin, People's Republic of China
| | - Xue Wu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, 300132 Tianjin, People's Republic of China
- School of Health Sciences and Biomedical Engineering, Hebei University of Technology, 300132 Tianjin, People's Republic of China
- Hebei Key Laboratory of Bioelectromagnetics and Neuroengineering, 300132 Tianjin, People's Republic of China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, 300132 Tianjin, People's Republic of China
| | - Guizhi Xu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, 300132 Tianjin, People's Republic of China
- School of Health Sciences and Biomedical Engineering, Hebei University of Technology, 300132 Tianjin, People's Republic of China
- Hebei Key Laboratory of Bioelectromagnetics and Neuroengineering, 300132 Tianjin, People's Republic of China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, 300132 Tianjin, People's Republic of China
| |
Collapse
|
21
|
Wang X, Li B, Lin Y, Gao X. Multi-source domain adaptation based tempo-spatial convolution network for cross-subject EEG classification in RSVP task. J Neural Eng 2024; 21:016025. [PMID: 38324909 DOI: 10.1088/1741-2552/ad2710] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Accepted: 02/07/2024] [Indexed: 02/09/2024]
Abstract
Objective.Many subject-dependent methods were proposed for electroencephalogram (EEG) classification in rapid serial visual presentation (RSVP) task, which required a large amount of data from new subject and were time-consuming to calibrate system. Cross-subject classification can realize calibration reduction or zero calibration. However, cross-subject classification in RSVP task is still a challenge.Approach.This study proposed a multi-source domain adaptation based tempo-spatial convolution (MDA-TSC) network for cross-subject RSVP classification. The proposed network consisted of three modules. First, the common feature extraction with multi-scale tempo-spatial convolution was constructed to extract domain-invariant features across all subjects, which could improve generalization of the network. Second, the multi-branch domain-specific feature extraction and alignment was conducted to extract and align domain-specific feature distributions of source and target domains in pairs, which could consider feature distribution differences among source domains. Third, the domain-specific classifier was exploited to optimize the network through loss functions and obtain prediction for the target domain.Main results.The proposed network was evaluated on the benchmark RSVP dataset, and the cross-subject classification results showed that the proposed MDA-TSC network outperformed the reference methods. Moreover, the effectiveness of the MDA-TSC network was verified through both ablation studies and visualization.Significance.The proposed network could effectively improve cross-subject classification performance in RSVP task, and was helpful to reduce system calibration time.
Collapse
Affiliation(s)
- Xuepu Wang
- School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing 100081, People's Republic of China
| | - Bowen Li
- School of Medicine, Tsinghua University, Beijing 100084, People's Republic of China
| | - Yanfei Lin
- School of Integrated Circuits and Electronics, Beijing Institute of Technology, Beijing 100081, People's Republic of China
| | - Xiaorong Gao
- School of Medicine, Tsinghua University, Beijing 100084, People's Republic of China
| |
Collapse
|
22
|
Zhong Y, Yao L, Pan G, Wang Y. Cross-Subject Motor Imagery Decoding by Transfer Learning of Tactile ERD. IEEE Trans Neural Syst Rehabil Eng 2024; 32:662-671. [PMID: 38271166 DOI: 10.1109/tnsre.2024.3358491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2024]
Abstract
For Brain-Computer Interface (BCI) based on motor imagery (MI), the MI task is abstract and spontaneous, presenting challenges in measurement and control and resulting in a lower signal-to-noise ratio. The quality of the collected MI data significantly impacts the cross-subject calibration results. To address this challenge, we introduce a novel cross-subject calibration method based on passive tactile afferent stimulation, in which data induced by tactile stimulation is utilized to calibrate transfer learning models for cross-subject decoding. During the experiments, tactile stimulation was applied to either the left or right hand, with subjects only required to sense tactile stimulation. Data from these tactile tasks were used to train or fine-tune models and subsequently applied to decode pure MI data. We evaluated BCI performance using both the classical Common Spatial Pattern (CSP) combined with the Linear Discriminant Analysis (LDA) algorithm and a state-of-the-art deep transfer learning model. The results demonstrate that the proposed calibration method achieved decoding performance at an equivalent level to traditional MI calibration, with the added benefit of outperforming traditional MI calibration with fewer trials. The simplicity and effectiveness of the proposed cross-subject tactile calibration method make it valuable for practical applications of BCI, especially in clinical settings.
Collapse
|
23
|
Darvishi-Bayazi MJ, Ghaemi MS, Lesort T, Arefin MR, Faubert J, Rish I. Amplifying pathological detection in EEG signaling pathways through cross-dataset transfer learning. Comput Biol Med 2024; 169:107893. [PMID: 38183700 DOI: 10.1016/j.compbiomed.2023.107893] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2023] [Revised: 11/28/2023] [Accepted: 12/20/2023] [Indexed: 01/08/2024]
Abstract
Pathology diagnosis based on EEG signals and decoding brain activity holds immense importance in understanding neurological disorders. With the advancement of artificial intelligence methods and machine learning techniques, the potential for accurate data-driven diagnoses and effective treatments has grown significantly. However, applying machine learning algorithms to real-world datasets presents diverse challenges at multiple levels. The scarcity of labeled data, especially in low regime scenarios with limited availability of real patient cohorts due to high costs of recruitment, underscores the vital deployment of scaling and transfer learning techniques. In this study, we explore a real-world pathology classification task to highlight the effectiveness of data and model scaling and cross-dataset knowledge transfer. As such, we observe varying performance improvements through data scaling, indicating the need for careful evaluation and labeling. Additionally, we identify the challenges of possible negative transfer and emphasize the significance of some key components to overcome distribution shifts and potential spurious correlations and achieve positive transfer. We see improvement in the performance of the target model on the target (NMT) datasets by using the knowledge from the source dataset (TUAB) when a low amount of labeled data was available. Our findings demonstrated that a small and generic model (e.g. ShallowNet) performs well on a single dataset, however, a larger model (e.g. TCN) performs better in transfer learning when leveraging a larger and more diverse dataset.
Collapse
Affiliation(s)
- Mohammad-Javad Darvishi-Bayazi
- Mila, Québec AI Institute, Montréal, QC, Canada; Faubert Lab, Montréal, QC, Canada; Université de Montréal, Montréal, QC, Canada.
| | | | - Timothee Lesort
- Mila, Québec AI Institute, Montréal, QC, Canada; Université de Montréal, Montréal, QC, Canada
| | - Md Rifat Arefin
- Mila, Québec AI Institute, Montréal, QC, Canada; Université de Montréal, Montréal, QC, Canada
| | - Jocelyn Faubert
- Faubert Lab, Montréal, QC, Canada; Université de Montréal, Montréal, QC, Canada
| | - Irina Rish
- Mila, Québec AI Institute, Montréal, QC, Canada; Université de Montréal, Montréal, QC, Canada
| |
Collapse
|
24
|
Zhang J, Liu D, Chen W, Pei Z, Wang J. Boosting lower-limb motor imagery performance through an ensemble method for gait rehabilitation. Comput Biol Med 2024; 169:107910. [PMID: 38183703 DOI: 10.1016/j.compbiomed.2023.107910] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2023] [Revised: 12/05/2023] [Accepted: 12/24/2023] [Indexed: 01/08/2024]
Abstract
Lower-limb exoskeletons have been used extensively in many rehabilitation applications to assist disabled people with their therapies. Brain-machine interfaces (BMIs) further provide effective and natural control schemes. However, the limited performance of brain signal decoding from lower-limb kinematics restricts the broad growth of both BMI and rehabilitation industry. To address these challenges, we propose an ensemble method for lower-limb motor imagery (MI) classification. The proposed model employs multiple techniques to boost performance, including deep and shallow parts. Traditional wavelet transformation followed by filter-bank common spatial pattern (CSP) employs neurophysiologically reasonable patterns, while multi-head self-attention (MSA) followed by temporal convolutional network (TCN) extracts deeper encoded generalized patterns. Experimental results in a customized lower-limb exoskeleton on 8 subjects in 3 consecutive sessions showed that the proposed method achieved 60.27% and 64.20% for three (MI of left leg, MI of right leg, and rest) and two classes (lower-limb MI vs. rest), respectively. Besides, the proposed model achieves improvements of up to 4% and 2% accuracy for the subject-specific and subject-independent modes compared to the current state-of-the-art (SOTA) techniques, respectively. Finally, feature analysis was conducted to show discriminative brain patterns in each MI task and sessions with different feedback modalities. The proposed models integrated in the brain-actuated lower-limb exoskeleton established a potential BMI for gait training and neuroprosthesis.
Collapse
Affiliation(s)
- Jing Zhang
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, China; Hangzhou Innovation Institute, Beihang University, Hangzhou, Zhejiang 310052, China.
| | - Dong Liu
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, China.
| | - Weihai Chen
- School of Electrical Engineering and Automation, Anhui University, Hefei, Anhui 230601, China.
| | - Zhongcai Pei
- Hangzhou Innovation Institute, Beihang University, Hangzhou, Zhejiang 310052, China.
| | - Jianhua Wang
- Hangzhou Innovation Institute, Beihang University, Hangzhou, Zhejiang 310052, China.
| |
Collapse
|
25
|
Chen Z, Yang R, Huang M, Li F, Lu G, Wang Z. EEGProgress: A fast and lightweight progressive convolution architecture for EEG classification. Comput Biol Med 2024; 169:107901. [PMID: 38159400 DOI: 10.1016/j.compbiomed.2023.107901] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Revised: 12/11/2023] [Accepted: 12/23/2023] [Indexed: 01/03/2024]
Abstract
Because of the intricate topological structure and connection of the human brain, extracting deep spatial features from electroencephalograph (EEG) signals is a challenging and time-consuming task. The extraction of topological spatial information plays a crucial role in EEG classification, and the architecture of the spatial convolution greatly affects the performance and complexity of convolutional neural network (CNN) based EEG classification models. In this study, a progressive convolution CNN architecture named EEGProgress is proposed, aiming to efficiently extract the topological spatial information of EEG signals from multi-scale levels (electrode, brain region, hemisphere, global) with superior speed. To achieve this, the raw EEG data is permuted using the empirical topological permutation rule, integrating the EEG data with numerous topological properties. Subsequently, the spatial features are extracted by a progressive feature extractor including prior, electrode, region, and hemisphere convolution blocks, progressively extracting the deep spatial features with reduced parameters and speed. Finally, the comparison and ablation experiments under both cross-subject and within-subject scenarios are conducted on a public dataset to verify the performance of the proposed EEGProgress and the effectiveness of the topological permutation. The results demonstrate the superior feature extraction ability of the proposed EEGProgress, with an average increase of 4.02% compared to other CNN-based EEG classification models under both cross-subject and within-subject scenarios. Furthermore, with the obtained average testing time, FLOPs, and parameters, the proposed EEGProgress outperforms other comparison models in terms of model complexity.
Collapse
Affiliation(s)
- Zhige Chen
- School of Advanced Technology, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China; School of Electrical Engineering, Electronics and Computer Science, University of Liverpool, Liverpool L69 3BX, United Kingdom
| | - Rui Yang
- School of Advanced Technology, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China
| | - Mengjie Huang
- Design School, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China.
| | - Fumin Li
- School of Advanced Technology, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China; School of Electrical Engineering, Electronics and Computer Science, University of Liverpool, Liverpool L69 3BX, United Kingdom
| | - Guoping Lu
- School of Electrical Engineering, Nantong University, Nantong 226019, China
| | - Zidong Wang
- Department of Computer Science, Brunel University London, Uxbridge, Middlesex UB8 3PH, United Kingdom
| |
Collapse
|
26
|
Rafiei MH, Gauthier LV, Adeli H, Takabi D. Self-Supervised Learning for Electroencephalography. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:1457-1471. [PMID: 35867362 DOI: 10.1109/tnnls.2022.3190448] [Citation(s) in RCA: 25] [Impact Index Per Article: 25.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Decades of research have shown machine learning superiority in discovering highly nonlinear patterns embedded in electroencephalography (EEG) records compared with conventional statistical techniques. However, even the most advanced machine learning techniques require relatively large, labeled EEG repositories. EEG data collection and labeling are costly. Moreover, combining available datasets to achieve a large data volume is usually infeasible due to inconsistent experimental paradigms across trials. Self-supervised learning (SSL) solves these challenges because it enables learning from EEG records across trials with variable experimental paradigms, even when the trials explore different phenomena. It aggregates multiple EEG repositories to increase accuracy, reduce bias, and mitigate overfitting in machine learning training. In addition, SSL could be employed in situations where there is limited labeled training data, and manual labeling is costly. This article: 1) provides a brief introduction to SSL; 2) describes some SSL techniques employed in recent studies, including EEG; 3) proposes current and potential SSL techniques for future investigations in EEG studies; 4) discusses the cons and pros of different SSL techniques; and 5) proposes holistic implementation tips and potential future directions for EEG SSL practices.
Collapse
|
27
|
Rakhmatulin I, Dao MS, Nassibi A, Mandic D. Exploring Convolutional Neural Network Architectures for EEG Feature Extraction. SENSORS (BASEL, SWITZERLAND) 2024; 24:877. [PMID: 38339594 PMCID: PMC10856895 DOI: 10.3390/s24030877] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Revised: 01/12/2024] [Accepted: 01/20/2024] [Indexed: 02/12/2024]
Abstract
The main purpose of this paper is to provide information on how to create a convolutional neural network (CNN) for extracting features from EEG signals. Our task was to understand the primary aspects of creating and fine-tuning CNNs for various application scenarios. We considered the characteristics of EEG signals, coupled with an exploration of various signal processing and data preparation techniques. These techniques include noise reduction, filtering, encoding, decoding, and dimension reduction, among others. In addition, we conduct an in-depth analysis of well-known CNN architectures, categorizing them into four distinct groups: standard implementation, recurrent convolutional, decoder architecture, and combined architecture. This paper further offers a comprehensive evaluation of these architectures, covering accuracy metrics, hyperparameters, and an appendix that contains a table outlining the parameters of commonly used CNN architectures for feature extraction from EEG signals.
Collapse
Affiliation(s)
- Ildar Rakhmatulin
- Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK; (A.N.)
| | - Minh-Son Dao
- National Institute of Information and Communications Technology (NICT), Tokyo 184-0015, Japan
| | - Amir Nassibi
- Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK; (A.N.)
| | - Danilo Mandic
- Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK; (A.N.)
| |
Collapse
|
28
|
Ellis CA, Miller RL, Calhoun VD. Evaluating Augmentation Approaches for Deep Learning-based Major Depressive Disorder Diagnosis with Raw Electroencephalogram Data. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.12.15.571938. [PMID: 38187601 PMCID: PMC10769199 DOI: 10.1101/2023.12.15.571938] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/09/2024]
Abstract
While deep learning methods are increasingly applied in research contexts for neuropsychiatric disorder diagnosis, small dataset size limits their potential for clinical translation. Data augmentation (DA) could address this limitation, but the utility of EEG DA methods remains relatively underexplored in neuropsychiatric disorder diagnosis. In this study, we train a model for major depressive disorder diagnosis. We then evaluate the utility of 6 EEG DA approaches. Importantly, to remove the bias that could be introduced by comparing performance for models trained on larger augmented training sets to models trained on smaller baseline sets, we also introduce a new baseline trained on duplicate training data to better. We lastly examine the effects of the DA approaches upon representations learned by the model with a pair of explainability analyses. We find that while most approaches boost model performance, they do not improve model performance beyond that of simply using a duplicate training set without DA. The exception to this is channel dropout augmentation, which does improve model performance. These findings suggest the importance of comparing EEG DA methods to a baseline with a duplicate training set of equal size to the augmented training set. We also found that some DA methods increased model robustness to frequency (Fourier transform surrogates) and channel (channel dropout) perturbation. While our findings on EEG DA efficacy are restricted to our dataset and model, we hope that future studies on deep learning for small EEG datasets and on new EEG DA methods will find our findings helpful.
Collapse
Affiliation(s)
- Charles A Ellis
- Center for Translational Research in Neuroimaging and Data Science Georgia State University, Emory University, Georgia Institute of Technology Atlanta, USA
| | - Robyn L Miller
- Center for Translational Research in Neuroimaging and Data Science Georgia State University, Emory University, Georgia Institute of Technology Atlanta, USA
| | - Vince D Calhoun
- Center for Translational Research in Neuroimaging and Data Science Georgia State University, Emory University, Georgia Institute of Technology Atlanta, USA
| |
Collapse
|
29
|
De Miguel-Rubio A, Gallego-Aguayo I, De Miguel-Rubio MD, Arias-Avila M, Lucena-Anton D, Alba-Rueda A. Effectiveness of the Combined Use of a Brain-Machine Interface System and Virtual Reality as a Therapeutic Approach in Patients with Spinal Cord Injury: A Systematic Review. Healthcare (Basel) 2023; 11:3189. [PMID: 38132079 PMCID: PMC10742447 DOI: 10.3390/healthcare11243189] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 11/30/2023] [Accepted: 12/14/2023] [Indexed: 12/23/2023] Open
Abstract
Spinal cord injury has a major impact on both the individual and society. This damage can cause permanent loss of sensorimotor functions, leading to structural and functional changes in somatotopic regions of the spinal cord. The combined use of a brain-machine interface and virtual reality offers a therapeutic alternative to be considered in the treatment of this pathology. This systematic review aimed to evaluate the effectiveness of the combined use of virtual reality and the brain-machine interface in the treatment of spinal cord injuries. A search was performed in PubMed, Web of Science, PEDro, Cochrane Central Register of Controlled Trials, CINAHL, Scopus, and Medline, including articles published from the beginning of each database until January 2023. Articles were selected based on strict inclusion and exclusion criteria. The Cochrane Collaboration's tool was used to assess the risk of bias and the PEDro scale and SCIRE systems were used to evaluate the methodological quality of the studies. Eleven articles were selected from a total of eighty-two. Statistically significant changes were found in the upper limb, involving improvements in shoulder and upper arm mobility, and weaker muscles were strengthened. In conclusion, most of the articles analyzed used the electroencephalogram as a measurement instrument for the assessment of various parameters, and most studies have shown improvements. Nonetheless, further research is needed with a larger sample size and long-term follow-up to establish conclusive results regarding the effect size of these interventions.
Collapse
Affiliation(s)
- Amaranta De Miguel-Rubio
- Department of Nursing, Pharmacology and Physiotherapy, University of Cordoba, 14004 Cordoba, Spain; (I.G.-A.); (A.A.-R.)
| | - Ignacio Gallego-Aguayo
- Department of Nursing, Pharmacology and Physiotherapy, University of Cordoba, 14004 Cordoba, Spain; (I.G.-A.); (A.A.-R.)
| | | | - Mariana Arias-Avila
- Physical Therapy Department, Universidade Federal de São Carlos, São Paulo 13565-905, Brazil;
| | - David Lucena-Anton
- Department of Nursing and Physiotherapy, University of Cadiz, 11009 Cadiz, Spain;
| | - Alvaro Alba-Rueda
- Department of Nursing, Pharmacology and Physiotherapy, University of Cordoba, 14004 Cordoba, Spain; (I.G.-A.); (A.A.-R.)
| |
Collapse
|
30
|
Han J, Wei X, Faisal AA. EEG decoding for datasets with heterogenous electrode configurations using transfer learning graph neural networks. J Neural Eng 2023; 20:066027. [PMID: 37931308 DOI: 10.1088/1741-2552/ad09ff] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Accepted: 11/06/2023] [Indexed: 11/08/2023]
Abstract
Objective. Brain-machine interfacing (BMI) has greatly benefited from adopting machine learning methods for feature learning that require extensive data for training, which are often unavailable from a single dataset. Yet, it is difficult to combine data across labs or even data within the same lab collected over the years due to the variation in recording equipment and electrode layouts resulting in shifts in data distribution, changes in data dimensionality, and altered identity of data dimensions. Our objective is to overcome this limitation and learn from many different and diverse datasets across labs with different experimental protocols.Approach. To tackle the domain adaptation problem, we developed a novel machine learning framework combining graph neural networks (GNNs) and transfer learning methodologies for non-invasive motor imagery (MI) EEG decoding, as an example of BMI. Empirically, we focus on the challenges of learning from EEG data with different electrode layouts and varying numbers of electrodes. We utilize three MI EEG databases collected using very different numbers of EEG sensors (from 22 channels to 64) and layouts (from custom layouts to 10-20).Main results. Our model achieved the highest accuracy with lower standard deviations on the testing datasets. This indicates that the GNN-based transfer learning framework can effectively aggregate knowledge from multiple datasets with different electrode layouts, leading to improved generalization in subject-independent MI EEG classification.Significance. The findings of this study have important implications for brain-computer-interface research, as they highlight a promising method for overcoming the limitations posed by non-unified experimental setups. By enabling the integration of diverse datasets with varying electrode layouts, our proposed approach can help advance the development and application of BMI technologies.
Collapse
Affiliation(s)
- Jinpei Han
- Brain & Behaviour Lab, Department of Computing, Imperial College London, London SW7 2AZ, United Kingdom
| | - Xiaoxi Wei
- Brain & Behaviour Lab, Department of Computing, Imperial College London, London SW7 2AZ, United Kingdom
| | - A Aldo Faisal
- Brain & Behaviour Lab, Department of Computing, Imperial College London, London SW7 2AZ, United Kingdom
- Chair in Digital Health & Data Science, University of Bayreuth, 95447 Bayreuth, Germany
| |
Collapse
|
31
|
Zhang B, Wei D, Yan G, Li X, Su Y, Cai H. Spatial-Temporal EEG Fusion Based on Neural Network for Major Depressive Disorder Detection. Interdiscip Sci 2023; 15:542-559. [PMID: 37140772 PMCID: PMC10158716 DOI: 10.1007/s12539-023-00567-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2023] [Revised: 04/12/2023] [Accepted: 04/13/2023] [Indexed: 05/05/2023]
Abstract
In view of the major depressive disorder characteristics such as high mortality as well as high recurrence, it is important to explore an objective and effective detection method for major depressive disorder. Considering the advantages complementary of different machine learning algorithms in information mining process, as well as the fusion complementary of different information, in this study, the spatial-temporal electroencephalography fusion framework using neural network is proposed for major depressive disorder detection. Since electroencephalography is a typical time series signal, we introduce recurrent neural network embedded in long short-term memory unit for extract temporal domain features to solve the problem of long-distance information dependence. To reduce the volume conductor effect, the temporal electroencephalography data are mapping into a spatial brain functional network using phase lag index, then the spatial domain features were extracted from brain functional network using 2D convolutional neural networks. Considering the complementarity between different types of features, the spatial-temporal electroencephalography features are fused to achieve data diversity. The experimental results show that spatial-temporal features fusion can improve the detection accuracy of major depressive disorder with a highest of 96.33%. In addition, our research also found that theta, alpha, and full frequency band in brain regions of left frontal, left central, right temporal are closely related to MDD detection, especially theta frequency band in left frontal region. Only using single-dimension EEG data as decision basis, it is difficult to fully explore the valuable information hidden in the data, which affects the overall detection performance of MDD. Meanwhile, different algorithms have their own advantages for different application scenarios. Ideally, different algorithms should use their respective advantages to jointly address complex problems in engineering fields. To this end, we propose a computer-aided MDD detection framework based on spatial-temporal EEG fusion using neural network, as shown in Fig. 1. The simplified process is as follows: (1) Raw EEG data acquisition and preprocessing. (2) The time series EEG data of each channel are input as recurrent neural network (RNN), and RNN is used to process and extract temporal domain (TD) features. (3) The BFN among different EEG channels is constructed, and CNN is used to process and extract the spatial domain (SD) features of the BFN. (4) Based on the theory of information complementarity, the spatial-temporal information is fused to realize efficient MDD detection. Fig. 1 MDD detection framework based on spatial-temporal EEG fusion.
Collapse
Affiliation(s)
- Bingtao Zhang
- School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou, 730070, China.
- Key Laboratory of Opto-Technology and Intelligent Control Ministry of Education, Lanzhou Jiaotong University, Lanzhou, 730070, China.
- School of Information Science and Engineering, Lanzhou University, Lanzhou, 730000, China.
| | - Dan Wei
- School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou, 730070, China
| | - Guanghui Yan
- School of Electronic and Information Engineering, Lanzhou Jiaotong University, Lanzhou, 730070, China
| | - Xiulan Li
- Gansu Province Big Data Center, Lanzhou, 730000, China.
| | - Yun Su
- School of Information Science and Engineering, Lanzhou University, Lanzhou, 730000, China
- College of Computer Science and Engineering, Northwest Normal University, Lanzhou, 730070, China
| | - Hanshu Cai
- School of Information Science and Engineering, Lanzhou University, Lanzhou, 730000, China
| |
Collapse
|
32
|
Ellis CA, Miller RL, Calhoun VD. Cross-Sampling Rate Transfer Learning for Enhanced Raw EEG Deep Learning Classifier Performance in Major Depressive Disorder Diagnosis. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.11.13.566915. [PMID: 38014293 PMCID: PMC10680672 DOI: 10.1101/2023.11.13.566915] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/29/2023]
Abstract
Transfer learning offers a route for developing robust deep learning models on small raw electroencephalography (EEG) datasets. Nevertheless, the utility of applying representations learned from large datasets with a lower sampling rate to smaller datasets with higher sampling rates remains relatively unexplored. In this study, we transfer representations learned by a convolutional neural network on a large, publicly available sleep dataset with a 100 Hertz sampling rate to a major depressive disorder (MDD) diagnosis task at a sampling rate of 200 Hertz. Importantly, we find that the early convolutional layers contain representations that are generalizable across tasks. Moreover, our approach significantly increases mean model accuracy from 82.33% to 86.99%, increases the model's use of lower frequencies, (θ-band), and increases its robustness to channel loss. We expect this analysis to provide useful guidance and enable more widespread use of transfer learning in EEG deep learning studies.
Collapse
|
33
|
Jin H, Chen S. Biometric Recognition Based on Recurrence Plot and InceptionV3 Model Using Eye Movements. IEEE J Biomed Health Inform 2023; 27:5554-5563. [PMID: 37682647 DOI: 10.1109/jbhi.2023.3313261] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/10/2023]
Abstract
The ability to use eye movement signals as a feature in biometric recognition is a novel characteristic of biometric recognition technology. However, present technologies have not fully exploited the correlation features between eye movement signals. To address this, we propose an eye movement biometric recognition model that is based on recurrence plot encoding and the InceptionV3 model. We first encode the original eye movement signal using the recurrence plot to obtain a 2-D image that is then used as input for the InceptionV3 model to perform biometric recognition. Our experimental results using the GazeBaseV2.0 eye movement dataset demonstrate that our proposed model achieved a high biometric recognition accuracy of 96.58% ± 0.66% using the recurrence plot transformation of the horizontal gaze position signals and the InceptionV3 model, surpassing the accuracy achieved by other models. The use of horizontal gaze position eye movement signals for biometric recognition outperforms the use of vertical gaze position signals when using our proposed methods. Furthermore, the biometric recognition that is achieved through recurrent plot encoding is superior to that achieved using Markov transition fields and Gramian angular field transformations.
Collapse
|
34
|
Altindis F, Banerjee A, Phlypo R, Yilmaz B, Congedo M. Transfer Learning for P300 Brain-Computer Interfaces by Joint Alignment of Feature Vectors. IEEE J Biomed Health Inform 2023; 27:4696-4706. [PMID: 37506011 DOI: 10.1109/jbhi.2023.3299837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/30/2023]
Abstract
This article presents a new transfer learning method named group learning, that jointly aligns multiple domains (many-to-many) and an extension named fast alignment that aligns any further domain to previously aligned group of domains (many-to-one). The proposed group alignment algorithm (GALIA) is evaluated on brain-computer interface (BCI) data and optimal hyper-parameter values of the algorithm are studied for classification performance and computational cost. Six publicly available P300 databases comprising 333 sessions from 177 subjects are used. As compared to the conventional subject-specific train/test pipeline, both group learning and fast alignment significantly improve the classification accuracy except for the database with clinical subjects (average improvement: 2.12±1.88%). GALIA utilizes cyclic approximate joint diagonalization (AJD) to find a set of linear transformations, one for each domain, jointly aligning the feature vectors of all domains. Group learning achieves a many-to-many transfer learning without compromising the classification performance on non-clinical BCI data. Fast alignment further extends the group learning for any unseen domains, allowing a many-to-one transfer learning with the same properties. The former method creates a single machine learning model using data from previous subjects and/or sessions, whereas the latter exploits the trained model for an unseen domain requiring no further training of the classifier.
Collapse
|
35
|
Sun Y, Chen X. Epileptic EEG Signal Detection Using Variational Modal Decomposition and Improved Grey Wolf Algorithm. SENSORS (BASEL, SWITZERLAND) 2023; 23:8078. [PMID: 37836909 PMCID: PMC10575143 DOI: 10.3390/s23198078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 09/16/2023] [Accepted: 09/18/2023] [Indexed: 10/15/2023]
Abstract
Epilepsy does great harm to the human body, and even threatens human life when it is serious. Therefore, research focused on the diagnosis and treatment of epilepsy holds paramount clinical significance. In this paper, we utilized variational modal decomposition (VMD) and an enhanced grey wolf algorithm to detect epileptic electroencephalogram (EEG) signals. Data were extracted from each patient's preseizure period and seizure period of 200 s each, with every 2 s as a segment, meaning 100 data points could be obtained for each patient's health period as well as 100 data points for each patient's epilepsy period. Variational modal decomposition (VMD) was used to obtain the corresponding intrinsic modal function (VMF) of the data. Then, the differential entropy (DE) and high frequency detection (HFD) of each VMF were extracted as features. The improved grey wolf algorithm is adopted for a selected channel to improve the maximum value of the channel. Finally, the EEG signal samples were classified using a support vector machine (SVM) classifier to achieve the accurate detection of epilepsy EEG signals. Experimental results show that the accuracy, sensitivity and specificity of the proposed method can reach 98.3%, 98.9% and 98.5%, respectively. The proposed algorithm in this paper can be used as an index to detect epileptic seizures and has certain guiding significance for the early diagnosis and effective treatment of epileptic patients.
Collapse
Affiliation(s)
- Yongxin Sun
- College of Electronic Information Engineering, Changchun University of Science and Technology, Changchun 130022, China
- College of Physics and Electronic Information, Baicheng Normal University, Baicheng 137099, China
| | - Xiaojuan Chen
- College of Electronic Information Engineering, Changchun University of Science and Technology, Changchun 130022, China
| |
Collapse
|
36
|
Chowdhury RR, Muhammad Y, Adeel U. Enhancing Cross-Subject Motor Imagery Classification in EEG-Based Brain-Computer Interfaces by Using Multi-Branch CNN. SENSORS (BASEL, SWITZERLAND) 2023; 23:7908. [PMID: 37765965 PMCID: PMC10536894 DOI: 10.3390/s23187908] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/11/2023] [Revised: 08/23/2023] [Accepted: 09/11/2023] [Indexed: 09/29/2023]
Abstract
A brain-computer interface (BCI) is a computer-based system that allows for communication between the brain and the outer world, enabling users to interact with computers using neural activity. This brain signal is obtained from electroencephalogram (EEG) signals. A significant obstacle to the development of BCIs based on EEG is the classification of subject-independent motor imagery data since EEG data are very individualized. Deep learning techniques such as the convolutional neural network (CNN) have illustrated their influence on feature extraction to increase classification accuracy. In this paper, we present a multi-branch (five branches) 2D convolutional neural network that employs several hyperparameters for every branch. The proposed model achieved promising results for cross-subject classification and outperformed EEGNet, ShallowConvNet, DeepConvNet, MMCNN, and EEGNet_Fusion on three public datasets. Our proposed model, EEGNet Fusion V2, achieves 89.6% and 87.8% accuracy for the actual and imagined motor activity of the eegmmidb dataset and scores of 74.3% and 84.1% for the BCI IV-2a and IV-2b datasets, respectively. However, the proposed model has a bit higher computational cost, i.e., it takes around 3.5 times more computational time per sample than EEGNet_Fusion.
Collapse
Affiliation(s)
- Radia Rayan Chowdhury
- Department of Computing & Games, School of Computing, Engineering & Digital Technologies, Teesside University, Middlesbrough TS1 3BX, UK
| | - Yar Muhammad
- Department of Computing & Games, School of Computing, Engineering & Digital Technologies, Teesside University, Middlesbrough TS1 3BX, UK
- Department of Computer Science, School of Physics, Engineering & Computer Science, University of Hertfordshire, Hatfield AL10 9AB, UK
| | - Usman Adeel
- Department of Computing & Games, School of Computing, Engineering & Digital Technologies, Teesside University, Middlesbrough TS1 3BX, UK
| |
Collapse
|
37
|
Huang J, Zhang ZQ, Xiong B, Wang Q, Wan B, Li F, Yang P. Cross-Subject Transfer Method Based on Domain Generalization for Facilitating Calibration of SSVEP-Based BCIs. IEEE Trans Neural Syst Rehabil Eng 2023; 31:3307-3319. [PMID: 37578926 DOI: 10.1109/tnsre.2023.3305202] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/16/2023]
Abstract
In steady-state visual evoked potential (SSVEP)-based brain-computer interfaces (BCIs), various spatial filtering methods based on individual calibration data have been proposed to alleviate the interference of spontaneous activities in SSVEP signals for enhancing the SSVEP detection performance. However, the time-consuming calibration session would increase the visual fatigue of subjects and reduce the usability of the BCI system. The key idea of this study is to propose a cross-subject transfer method based on domain generalization, which transfers the domain-invariant spatial filters and templates learned from source subjects to the target subject with no access to the EEG data from the target subject. The transferred spatial filters and templates are obtained by maximizing the intra- and inter-subject correlations using the SSVEP data corresponding to the target and its neighboring stimuli. For SSVEP detection of the target subject, four types of correlation coefficients are calculated to construct the feature vector. Experimental results estimated with three SSVEP datasets show that the proposed cross-subject transfer method improves the SSVEP detection performance compared to state-of-art methods. The satisfactory results demonstrate that the proposed method provides an effective transfer learning strategy requiring no tedious data collection process for new users, holding the potential of promoting practical applications of SSVEP-based BCI.
Collapse
|
38
|
Yap HY, Choo YH, Mohd Yusoh ZI, Khoh WH. An evaluation of transfer learning models in EEG-based authentication. Brain Inform 2023; 10:19. [PMID: 37535168 PMCID: PMC10400490 DOI: 10.1186/s40708-023-00198-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2022] [Accepted: 07/01/2023] [Indexed: 08/04/2023] Open
Abstract
Electroencephalogram(EEG)-based authentication has received increasing attention from researchers as they believe it could serve as an alternative to more conventional personal authentication methods. Unfortunately, EEG signals are non-stationary and could be easily contaminated by noise and artifacts. Therefore, further processing of data analysis is needed to retrieve useful information. Various machine learning approaches have been proposed and implemented in the EEG-based domain, with deep learning being the most current trend. However, retaining the performance of a deep learning model requires substantial computational effort and a vast amount of data, especially when the models go deeper to generate consistent results. Deep learning models trained with small data sets from scratch may experience an overfitting issue. Transfer learning becomes an alternative solution. It is a technique to recognize and apply the knowledge and skills learned from the previous tasks to a new domain with limited training data. This study attempts to explore the applicability of transferring various pre-trained models' knowledge to the EEG-based authentication domain. A self-collected database that consists of 30 subjects was utilized in the analysis. The database enrolment is divided into two sessions, with each session producing two sets of EEG recording data. The frequency spectrums of the preprocessed EEG signals are extracted and fed into the pre-trained models as the input data. Three experimental tests are carried out and the best performance is reported with accuracy in the range of 99.1-99.9%. The acquired results demonstrate the efficiency of transfer learning in authenticating an individual in this domain.
Collapse
Affiliation(s)
- Hui Yen Yap
- Faculty of Information Science and Technology, Multimedia University (MMU), Melaka, Malaysia.
- Faculty of Information and Communication Technology, Universiti Teknikal Malaysia Melaka (UTeM), Melaka, Malaysia.
| | - Yun-Huoy Choo
- Faculty of Information and Communication Technology, Universiti Teknikal Malaysia Melaka (UTeM), Melaka, Malaysia
| | - Zeratul Izzah Mohd Yusoh
- Faculty of Information and Communication Technology, Universiti Teknikal Malaysia Melaka (UTeM), Melaka, Malaysia
| | - Wee How Khoh
- Faculty of Information Science and Technology, Multimedia University (MMU), Melaka, Malaysia
| |
Collapse
|
39
|
Chaddad A, Wu Y, Kateb R, Bouridane A. Electroencephalography Signal Processing: A Comprehensive Review and Analysis of Methods and Techniques. SENSORS (BASEL, SWITZERLAND) 2023; 23:6434. [PMID: 37514728 PMCID: PMC10385593 DOI: 10.3390/s23146434] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/21/2023] [Revised: 06/16/2023] [Accepted: 07/13/2023] [Indexed: 07/30/2023]
Abstract
The electroencephalography (EEG) signal is a noninvasive and complex signal that has numerous applications in biomedical fields, including sleep and the brain-computer interface. Given its complexity, researchers have proposed several advanced preprocessing and feature extraction methods to analyze EEG signals. In this study, we analyze a comprehensive review of numerous articles related to EEG signal processing. We searched the major scientific and engineering databases and summarized the results of our findings. Our survey encompassed the entire process of EEG signal processing, from acquisition and pretreatment (denoising) to feature extraction, classification, and application. We present a detailed discussion and comparison of various methods and techniques used for EEG signal processing. Additionally, we identify the current limitations of these techniques and analyze their future development trends. We conclude by offering some suggestions for future research in the field of EEG signal processing.
Collapse
Affiliation(s)
- Ahmad Chaddad
- School of Artificial Intelligence, Guilin University of Electronic Technology, Guilin 541004, China
- The Laboratory for Imagery, Vision and Artificial Intelligence, Ecole de Technologie Supérieure, Montreal, QC H3C 1K3, Canada
| | - Yihang Wu
- School of Artificial Intelligence, Guilin University of Electronic Technology, Guilin 541004, China
| | - Reem Kateb
- College of Computer Science and Engineering, Taibah University, Madinah 41477, Saudi Arabia
| | - Ahmed Bouridane
- Centre for Data Analytics and Cybersecurity, University of Sharjah, Sharjah 27272, United Arab Emirates
| |
Collapse
|
40
|
Pandey A, Singh SK, Udmale SS, Shukla K. An intelligent optimized deep learning model to achieve early prediction of epileptic seizures. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104798] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/12/2023]
|
41
|
Zhang H, Lv X, Liu Y, Zou X. Hedge transfer learning routing for dynamic searching and reconnoitering applications in 3D multimedia FANETs. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1-35. [PMID: 37362678 PMCID: PMC10250183 DOI: 10.1007/s11042-023-15932-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/04/2022] [Revised: 03/20/2023] [Accepted: 05/29/2023] [Indexed: 06/28/2023]
Abstract
With the fast development of unmanned aerial vehicles (UAVs) and the user increasing demand of UAV video transmission, UAV video service is widely used in dynamic searching and reconnoitering applications. Video transmissions not only consider the complexity and instability of 3D UAV network topology but also ensure reliable quality of service (QoS) in flying ad hoc networks (FANETs). We propose hedge transfer learning routing (HTLR) for dynamic searching and reconnoitering applications to address this problem. Compared with the previous transfer learning framework, HTRL has the following innovations. First, hedge principle is introduced into transfer learning. Online model is continuously trained on the basis of offline model, and their weight factors are adjusted in real-time by transfer learning, so as to adapt to the complex 3D FANETs. Secondly, distributed multi-hop link state scheme is used to estimate multi-hop link states in the whole network, thus enhancing the stability of transmission links. Among them, we propose the multiplication rule of multi-hop link states, which is a new idea to evaluate link states. Finally, we use packet delivery rate (PDR) and energy efficiency rate (EER) as two main evaluation metrics. In the same NS3 experimental scenario, the PDR of HTLR is at least 5.11% higher and the EER is at least 1.17 lower than compared protocols. Besides, we use Wilcoxon test to compare HTLR with the simplified version of HTLR without hedge transfer learning (N-HTLR). The results show that HTRL is superior to N-HTRL, effectively ensuring QoS.
Collapse
Affiliation(s)
- HongGuang Zhang
- School of Electronic Engineering, Beijing Key Laboratory of Work Safety Intelligent Monitoring, Beijing University of Posts and Telecommunications, Beijing, China
| | - XiuSha Lv
- School of Electronic Engineering, Beijing Key Laboratory of Work Safety Intelligent Monitoring, Beijing University of Posts and Telecommunications, Beijing, China
| | - YuanAn Liu
- School of Electronic Engineering, Beijing Key Laboratory of Work Safety Intelligent Monitoring, Beijing University of Posts and Telecommunications, Beijing, China
| | - XinYing Zou
- School of Electronic Engineering, Beijing Key Laboratory of Work Safety Intelligent Monitoring, Beijing University of Posts and Telecommunications, Beijing, China
| |
Collapse
|
42
|
Shuqfa Z, Belkacem AN, Lakas A. Decoding Multi-Class Motor Imagery and Motor Execution Tasks Using Riemannian Geometry Algorithms on Large EEG Datasets. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23115051. [PMID: 37299779 DOI: 10.3390/s23115051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/13/2023] [Revised: 05/11/2023] [Accepted: 05/15/2023] [Indexed: 06/12/2023]
Abstract
The use of Riemannian geometry decoding algorithms in classifying electroencephalography-based motor-imagery brain-computer interfaces (BCIs) trials is relatively new and promises to outperform the current state-of-the-art methods by overcoming the noise and nonstationarity of electroencephalography signals. However, the related literature shows high classification accuracy on only relatively small BCI datasets. The aim of this paper is to provide a study of the performance of a novel implementation of the Riemannian geometry decoding algorithm using large BCI datasets. In this study, we apply several Riemannian geometry decoding algorithms on a large offline dataset using four adaptation strategies: baseline, rebias, supervised, and unsupervised. Each of these adaptation strategies is applied in motor execution and motor imagery for both scenarios 64 electrodes and 29 electrodes. The dataset is composed of four-class bilateral and unilateral motor imagery and motor execution of 109 subjects. We run several classification experiments and the results show that the best classification accuracy is obtained for the scenario where the baseline minimum distance to Riemannian mean has been used. The mean accuracy values up to 81.5% for motor execution, and up to 76.4% for motor imagery. The accurate classification of EEG trials helps to realize successful BCI applications that allow effective control of devices.
Collapse
Affiliation(s)
- Zaid Shuqfa
- Connected Autonomous Intelligent Systems Laboratory, Department of Computer and Network Engineering, College of IT (CIT), United Arab Emirates University (UAEU), Al Ain 15551, United Arab Emirates
| | - Abdelkader Nasreddine Belkacem
- Connected Autonomous Intelligent Systems Laboratory, Department of Computer and Network Engineering, College of IT (CIT), United Arab Emirates University (UAEU), Al Ain 15551, United Arab Emirates
| | - Abderrahmane Lakas
- Connected Autonomous Intelligent Systems Laboratory, Department of Computer and Network Engineering, College of IT (CIT), United Arab Emirates University (UAEU), Al Ain 15551, United Arab Emirates
| |
Collapse
|
43
|
Alves CL, Toutain TGLDO, de Carvalho Aguiar P, Pineda AM, Roster K, Thielemann C, Porto JAM, Rodrigues FA. Diagnosis of autism spectrum disorder based on functional brain networks and machine learning. Sci Rep 2023; 13:8072. [PMID: 37202411 DOI: 10.1038/s41598-023-34650-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2022] [Accepted: 05/04/2023] [Indexed: 05/20/2023] Open
Abstract
Autism is a multifaceted neurodevelopmental condition whose accurate diagnosis may be challenging because the associated symptoms and severity vary considerably. The wrong diagnosis can affect families and the educational system, raising the risk of depression, eating disorders, and self-harm. Recently, many works have proposed new methods for the diagnosis of autism based on machine learning and brain data. However, these works focus on only one pairwise statistical metric, ignoring the brain network organization. In this paper, we propose a method for the automatic diagnosis of autism based on functional brain imaging data recorded from 500 subjects, where 242 present autism spectrum disorder considering the regions of interest throughout Bootstrap Analysis of Stable Cluster map. Our method can distinguish the control group from autism spectrum disorder patients with high accuracy. Indeed the best performance provides an AUC near 1.0, which is higher than that found in the literature. We verify that the left ventral posterior cingulate cortex region is less connected to an area in the cerebellum of patients with this neurodevelopment disorder, which agrees with previous studies. The functional brain networks of autism spectrum disorder patients show more segregation, less distribution of information across the network, and less connectivity compared to the control cases. Our workflow provides medical interpretability and can be used on other fMRI and EEG data, including small data sets.
Collapse
Affiliation(s)
- Caroline L Alves
- Institute of Mathematical and Computer Sciences (ICMC), University of São Paulo (USP), São Paulo, Brazil.
- BioMEMS Lab, Aschaffenburg University of Applied Sciences, Aschaffenburg, Germany.
| | | | - Patricia de Carvalho Aguiar
- Hospital Israelita Albert Einstein, São Paulo, Brazil
- Department of Neurology and Neurosurgery, Federal University of São Paulo, São Paulo, Brazil
| | - Aruane M Pineda
- Institute of Mathematical and Computer Sciences (ICMC), University of São Paulo (USP), São Paulo, Brazil
| | - Kirstin Roster
- Institute of Mathematical and Computer Sciences (ICMC), University of São Paulo (USP), São Paulo, Brazil
| | | | | | - Francisco A Rodrigues
- Institute of Mathematical and Computer Sciences (ICMC), University of São Paulo (USP), São Paulo, Brazil
| |
Collapse
|
44
|
Habashi AG, Azab AM, Eldawlatly S, Aly GM. Generative adversarial networks in EEG analysis: an overview. J Neuroeng Rehabil 2023; 20:40. [PMID: 37038142 PMCID: PMC10088201 DOI: 10.1186/s12984-023-01169-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Accepted: 03/30/2023] [Indexed: 04/12/2023] Open
Abstract
Electroencephalogram (EEG) signals have been utilized in a variety of medical as well as engineering applications. However, one of the challenges associated with recording EEG data is the difficulty of recording large amounts of data. Consequently, data augmentation is a potential solution to overcome this challenge in which the objective is to increase the amount of data. Inspired by the success of Generative Adversarial Networks (GANs) in image processing applications, generating artificial EEG data from the limited recorded data using GANs has seen recent success. This article provides an overview of various techniques and approaches of GANs for augmenting EEG signals. We focus on the utility of GANs in different applications including Brain-Computer Interface (BCI) paradigms such as motor imagery and P300-based systems, in addition to emotion recognition, epileptic seizures detection and prediction, and various other applications. We address in this article how GANs have been used in each study, the impact of using GANs on the model performance, the limitations of each algorithm, and future possibilities for developing new algorithms. We emphasize the utility of GANs in augmenting the limited EEG data typically available in the studied applications.
Collapse
Affiliation(s)
- Ahmed G Habashi
- Computer and Systems Engineering Department, Faculty of Engineering, Ain Shams University, 1 El-Sarayat St., Abbassia, Cairo, Egypt
| | - Ahmed M Azab
- Biomedical Engineering Department, Technical Research Center, Cairo, Egypt
| | - Seif Eldawlatly
- Computer and Systems Engineering Department, Faculty of Engineering, Ain Shams University, 1 El-Sarayat St., Abbassia, Cairo, Egypt.
- Computer Science and Engineering Department, The American University in Cairo, Cairo, Egypt.
| | - Gamal M Aly
- Computer and Systems Engineering Department, Faculty of Engineering, Ain Shams University, 1 El-Sarayat St., Abbassia, Cairo, Egypt
| |
Collapse
|
45
|
Wang L, Huang M, Yang R, Liang HN, Han J, Sun Y. Survey of Movement Reproduction in Immersive Virtual Rehabilitation. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2023; 29:2184-2202. [PMID: 35015645 DOI: 10.1109/tvcg.2022.3142198] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Virtual reality (VR) has emerged as a powerful tool for rehabilitation. Many effective VR applications have been developed to support motor rehabilitation of people affected by motor issues. Movement reproduction, which transfers users' movements from the physical world to the virtual environment, is commonly used in VR rehabilitation applications. Three major components are required for movement reproduction in VR: (1) movement input, (2) movement representation, and (3) movement modulation. Until now, movement reproduction in virtual rehabilitation has not yet been systematically studied. This article aims to provide a state-of-the-art review on this subject by focusing on existing literature on immersive motor rehabilitation using VR. In this review, we provided in-depth discussions on the rehabilitation goals and outcomes, technology issues behind virtual rehabilitation, and user experience regarding movement reproduction. Similarly, we present good practices and highlight challenges and opportunities that can form constructive suggestions for the design and development of fit-for-purpose VR rehabilitation applications and can help frame future research directions for this emerging area that combines VR and health.
Collapse
|
46
|
Li J, Wu SR, Zhang X, Luo TJ, Li R, Zhao Y, Liu B, Peng H. Cross-subject aesthetic preference recognition of Chinese dance posture using EEG. Cogn Neurodyn 2023; 17:311-329. [PMID: 37007204 PMCID: PMC10050299 DOI: 10.1007/s11571-022-09821-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2021] [Revised: 04/21/2022] [Accepted: 05/09/2022] [Indexed: 11/03/2022] Open
Abstract
Due to the differences in knowledge, experience, background, and social influence, people have subjective characteristics in the process of dance aesthetic cognition. To explore the neural mechanism of the human brain in the process of dance aesthetic preference, and to find a more objective determining criterion for dance aesthetic preference, this paper constructs a cross-subject aesthetic preference recognition model of Chinese dance posture. Specifically, Dai nationality dance (a classic Chinese folk dance) was used to design dance posture materials, and an experimental paradigm for aesthetic preference of Chinese dance posture was built. Then, 91 subjects were recruited for the experiment, and their EEG signals were collected. Finally, the transfer learning method and convolutional neural networks were used to identify the aesthetic preference of the EEG signals. Experimental results have shown the feasibility of the proposed model, and the objective aesthetic measurement in dance appreciation has been implemented. Based on the classification model, the accuracy of aesthetic preference recognition is 79.74%. Moreover, the recognition accuracies of different brain regions, different hemispheres, and different model parameters were also verified by the ablation study. Additionally, the experimental results reflected the following two facts: (1) in the visual aesthetic processing of Chinese dance posture, the occipital and frontal lobes are more activated and participate in dance aesthetic preference; (2) the right brain is more involved in the visual aesthetic processing of Chinese dance posture, which is consistent with the common knowledge that the right brain is responsible for processing artistic activities.
Collapse
Affiliation(s)
- Jing Li
- Academy of Arts, Shaoxing University, Shaoxing, 312000 China
| | - Shen-rui Wu
- Department of Computer Science and Engineering, Shaoxing University, Shaoxing, 312000 China
| | - Xiang Zhang
- Department of Computer Science and Engineering, Shaoxing University, Shaoxing, 312000 China
| | - Tian-jian Luo
- College of Computer and Cyber Security, Fujian Normal University, Fuzhou, 350117 China
| | - Rui Li
- National Engineering Laboratory for Educational Big Data, Central China Normal University, Wuhan, 430079 China
| | - Ying Zhao
- Department of Computer Science and Engineering, Shaoxing University, Shaoxing, 312000 China
| | - Bo Liu
- Department of Computer Science and Engineering, Shaoxing University, Shaoxing, 312000 China
| | - Hua Peng
- Department of Computer Science and Engineering, Shaoxing University, Shaoxing, 312000 China
- College of Information Science and Engineering, Jishou University, Jishou, 416000 China
| |
Collapse
|
47
|
Wang C, Wang Z, Ma L, Dong H, Sheng W. A novel contrastive adversarial network for minor-class data augmentation: Applications to pipeline fault diagnosis. Knowl Based Syst 2023. [DOI: 10.1016/j.knosys.2023.110516] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/30/2023]
|
48
|
Multi-Source geometric metric transfer learning for EEG classification. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104435] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022]
|
49
|
Mintz Hemed N, Melosh NA. An integrated perspective for the diagnosis and therapy of neurodevelopmental disorders - From an engineering point of view. Adv Drug Deliv Rev 2023; 194:114723. [PMID: 36746077 DOI: 10.1016/j.addr.2023.114723] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2022] [Revised: 12/14/2022] [Accepted: 01/29/2023] [Indexed: 02/05/2023]
Abstract
Neurodevelopmental disorders (NDDs) are complex conditions with largely unknown pathophysiology. While many NDD symptoms are familiar, the cause of these disorders remains unclear and may involve a combination of genetic, biological, psychosocial, and environmental risk factors. Current diagnosis relies heavily on behaviorally defined criteria, which may be biased by the clinical team's professional and cultural expectations, thus a push for new biological-based biomarkers for NDDs diagnosis is underway. Emerging new research technologies offer an unprecedented view into the electrical, chemical, and physiological activity in the brain and with further development in humans may provide clinically relevant diagnoses. These could also be extended to new treatment options, which can start to address the underlying physiological issues. When combined with current speech, language, occupational therapy, and pharmacological treatment these could greatly improve patient outcomes. The current review will discuss the latest technologies that are being used or may be used for NDDs diagnosis and treatment. The aim is to provide an inspiring and forward-looking view for future research in the field.
Collapse
Affiliation(s)
- Nofar Mintz Hemed
- Department of Materials Science and Engineering, Stanford University, Stanford, CA 94305, USA.
| | - Nicholas A Melosh
- Department of Materials Science and Engineering, Stanford University, Stanford, CA 94305, USA
| |
Collapse
|
50
|
Yang CY, Chen PC, Huang WC. Cross-Domain Transfer of EEG to EEG or ECG Learning for CNN Classification Models. SENSORS (BASEL, SWITZERLAND) 2023; 23:2458. [PMID: 36904661 PMCID: PMC10007254 DOI: 10.3390/s23052458] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Revised: 02/19/2023] [Accepted: 02/20/2023] [Indexed: 06/18/2023]
Abstract
Electroencephalography (EEG) is often used to evaluate several types of neurological brain disorders because of its noninvasive and high temporal resolution. In contrast to electrocardiography (ECG), EEG can be uncomfortable and inconvenient for patients. Moreover, deep-learning techniques require a large dataset and a long time for training from scratch. Therefore, in this study, EEG-EEG or EEG-ECG transfer learning strategies were applied to explore their effectiveness for the training of simple cross-domain convolutional neural networks (CNNs) used in seizure prediction and sleep staging systems, respectively. The seizure model detected interictal and preictal periods, whereas the sleep staging model classified signals into five stages. The patient-specific seizure prediction model with six frozen layers achieved 100% accuracy for seven out of nine patients and required only 40 s of training time for personalization. Moreover, the cross-signal transfer learning EEG-ECG model for sleep staging achieved an accuracy approximately 2.5% higher than that of the ECG model; additionally, the training time was reduced by >50%. In summary, transfer learning from an EEG model to produce personalized models for a more convenient signal can both reduce the training time and increase the accuracy; moreover, challenges such as data insufficiency, variability, and inefficiency can be effectively overcome.
Collapse
|