1
|
Wang H, Wang Z, Sun Y, Yuan Z, Xu T, Li J. A Cascade xDAWN EEGNet Structure for Unified Visual-Evoked Related Potential Detection. IEEE Trans Neural Syst Rehabil Eng 2024; 32:2270-2280. [PMID: 38885099 DOI: 10.1109/tnsre.2024.3415474] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/20/2024]
Abstract
Visual-based brain-computer interface (BCI) enables people to communicate with others by spelling words from the brain and helps professionals recognize targets in large numbers of images. P300 signals evoked by different types of stimuli, such as words or images, may vary significantly in terms of both amplitude and latency. A unified approach is required to detect variable P300 signals, which facilitates BCI applications, as well as deepens the understanding of the P300 generation mechanism. In this study, our proposed approach involves a cascade network structure that combines xDAWN and classical EEGNet techniques. This network is designed to classify target and non-target stimuli in both P300 speller and rapid serial visual presentation (RSVP) paradigms. The proposed approach is capable of recognizing more symbols with fewer repetitions (up to 5 rounds) compared to other models while possessing a better information transfer rate (ITR) as demonstrated on Dataset II (17.22 bits/min in the second repetition round) of BCI Competition III. Additionally, our approach has the highest unweighted average recall (UAR) performance for both 5 Hz ( 0.8134±0.0259 ) and 20 Hz ( 0.6527±0.0321 ) RSVP. The results show that the cascade network structure has better performance between both the P300 Speller and RSVP paradigms, manifesting that such a cascade structure is robust enough for dealing with P300-related signals (source code is available at https://github.com/embneural/Cascade-xDAWN-EEGNet-for-ERP-Detection).
Collapse
|
2
|
Wang Z, Xiang L, Zhang R. P300 intention recognition based on phase lag index (PLI)-rich-club brain functional network. THE REVIEW OF SCIENTIFIC INSTRUMENTS 2024; 95:045116. [PMID: 38624364 DOI: 10.1063/5.0202770] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/26/2023] [Accepted: 03/28/2024] [Indexed: 04/17/2024]
Abstract
Brain-computer interface (BCI) technology based on P300 signals has a broad application prospect in the assessment and diagnosis of clinical diseases and game control. The paper of selecting key electrodes to realize a wearable intention recognition system has become a hotspot for scholars at home and abroad. In this paper, based on the rich-club phenomenon that exists in the process of intention generation, a phase lag index (PLI)-rich-club-based intention recognition method for P300 is proposed. The rich-club structure is a network consisting of electrodes that are highly connected with other electrodes in the process of P300 generation. To construct the rich-club network, this paper uses PLI to construct the brain functional network, calculates rich-club coefficients of the network in the range of k degrees, initially identifies rich-club nodes based on the feature of node degree, and then performs a descending order of betweenness centrality and identifies the nodes with larger betweenness centrality as the specific rich-club nodes, extracts the non-linear features and frequency domain features of Rich-club nodes, and finally uses support vector machine for classification. The experimental results show that the range of rich-club coefficients is smaller with intent compared to that without intent. Validation was performed on the BCI Competition III dataset by reducing the number of channels to 17 and 16 for subject A and subject B, with recognition quasi-departure rates of 96.93% and 94.93%, respectively, and on the BCI Competition II dataset by reducing the number of channels to 17 for subjects, with a recognition accuracy of 95.50%.
Collapse
Affiliation(s)
- Zhongmin Wang
- School of Computer Science and Technology, Xi'an University of Posts and Telecommunications, Xi'an, Shaanxi 710121, China
- Shaanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi'an University of Posts and Telecommunications, Xi'an, Shaanxi 710121, China
- Xi'an Key Laboratory of Big Data and Intelligent Computing, Xi'an 710121, Shaanxi, China
| | - Leihua Xiang
- School of Computer Science and Technology, Xi'an University of Posts and Telecommunications, Xi'an, Shaanxi 710121, China
| | - Rong Zhang
- School of Computer Science and Technology, Xi'an University of Posts and Telecommunications, Xi'an, Shaanxi 710121, China
- Shaanxi Key Laboratory of Network Data Analysis and Intelligent Processing, Xi'an University of Posts and Telecommunications, Xi'an, Shaanxi 710121, China
- Xi'an Key Laboratory of Big Data and Intelligent Computing, Xi'an 710121, Shaanxi, China
| |
Collapse
|
3
|
Chen C, Ji Z, Sun Y, Bezerianos A, Thakor N, Wang H. Self-Attentive Channel-Connectivity Capsule Network for EEG-Based Driving Fatigue Detection. IEEE Trans Neural Syst Rehabil Eng 2023; 31:3152-3162. [PMID: 37494165 DOI: 10.1109/tnsre.2023.3299156] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/28/2023]
Abstract
Deep neural networks have recently been successfully extended to EEG-based driving fatigue detection. Nevertheless, most existing models fail to reveal the intrinsic inter-channel relations that are known to be beneficial for EEG-based classification. Additionally, these models require substantial data for training, which is often impractical due to the high cost of data collection. To simultaneously address these two issues, we propose a Self-Attentive Channel-Connectivity Capsule Network (SACC-CapsNet) for EEG-based driving fatigue detection in this paper. SACC-CapsNet starts with a temporal-channel attention module to investigate the critical temporal information and important channels for driving fatigue detection, refining the input EEG signals. Subsequently, the refined EEG data are transformed into a channel covariance matrix to capture the inter-channel relations, followed by selective kernel attention to extract the highly discriminative channel-connectivity features. Finally, a capsule neural network is employed to effectively learn the relationships between connectivity features, which is more suitable for limited data. To confirm the effectiveness of SACC-CapsNet, we collected 24-channel EEG data from 31 subjects (mean age=23.13±2.68 years, male/female=18/13) in a simulated fatigue driving environment. Extensive experiments were conducted with the acquired data, and the comparison results show that our proposed model outperforms state-of-the-art methods. Additionally, the channel covariance matrix learned from SACC-CapsNet reveals that the frontal pole is most informative for detecting driving fatigue, followed by the parietal and central regions. Intriguingly, the temporal-channel attention module can enhance the significance of these critical regions, and the reconstructed channel covariance matrix generated by the decoder network of SACC-CapsNet can effectively preserve valuable information about them.
Collapse
|
4
|
Fu B, Gu C, Fu M, Xia Y, Liu Y. A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals. Front Neurosci 2023; 17:1234162. [PMID: 37600016 PMCID: PMC10436100 DOI: 10.3389/fnins.2023.1234162] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2023] [Accepted: 07/20/2023] [Indexed: 08/22/2023] Open
Abstract
Emotion recognition is a challenging task, and the use of multimodal fusion methods for emotion recognition has become a trend. Fusion vectors can provide a more comprehensive representation of changes in the subject's emotional state, leading to more accurate emotion recognition results. Different fusion inputs or feature fusion methods have varying effects on the final fusion outcome. In this paper, we propose a novel Multimodal Feature Fusion Neural Network model (MFFNN) that effectively extracts complementary information from eye movement signals and performs feature fusion with EEG signals. We construct a dual-branch feature extraction module to extract features from both modalities while ensuring temporal alignment. A multi-scale feature fusion module is introduced, which utilizes cross-channel soft attention to adaptively select information from different spatial scales, enabling the acquisition of features at different spatial scales for effective fusion. We conduct experiments on the publicly available SEED-IV dataset, and our model achieves an accuracy of 87.32% in recognizing four emotions (happiness, sadness, fear, and neutrality). The results demonstrate that the proposed model can better explore complementary information from EEG and eye movement signals, thereby improving accuracy, and stability in emotion recognition.
Collapse
Affiliation(s)
- Baole Fu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Chunrui Gu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Ming Fu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Yuxiao Xia
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Yinhua Liu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
- Shandong Key Laboratory of Industrial Control Technology, Qingdao, China
| |
Collapse
|
5
|
TaghiBeyglou B, Shamsollahi MB. ETucker: a constrained tensor decomposition for single trial ERP extraction. Physiol Meas 2023; 44:075005. [PMID: 37414004 DOI: 10.1088/1361-6579/ace510] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2023] [Accepted: 07/06/2023] [Indexed: 07/08/2023]
Abstract
Objective.In this paper, we propose a new tensor decomposition to extract event-related potentials (ERP) by adding a physiologically meaningful constraint to the Tucker decomposition.Approach.We analyze the performance of the proposed model and compare it with Tucker decomposition by synthesizing a dataset. The simulated dataset is generated using a 12th-order autoregressive model in combination with independent component analysis (ICA) on real no-task electroencephalogram (EEG) recordings. The dataset is manipulated to contain the P300 ERP component and to cover different SNR conditions, ranging from 0 to -30 dB, to simulate the presence of the P300 component in extremely noisy recordings. Furthermore, in order to assess the practicality of the proposed methodology in real-world scenarios, we utilized the brain-computer interface (BCI) competition III-dataset II.Main results.Our primary results demonstrate the superior performance of our approach compared to conventional methods commonly employed for single-trial estimation. Additionally, our method outperformed both Tucker decomposition and non-negative Tucker decomposition in the synthesized dataset. Furthermore, the results obtained from real-world data exhibited meaningful performance and provided insightful interpretations for the extracted P300 component.Significance.The findings suggest that the proposed decomposition is eminently capable of extracting the target P300 component's waveform, including latency and amplitude as well as its spatial location, using single-trial EEG recordings.
Collapse
|
6
|
Mao J, Qiu S, Wei W, He H. Cross-modal guiding and reweighting network for multi-modal RSVP-based target detection. Neural Netw 2023; 161:65-82. [PMID: 36736001 DOI: 10.1016/j.neunet.2023.01.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Revised: 10/31/2022] [Accepted: 01/11/2023] [Indexed: 01/17/2023]
Abstract
Rapid Serial Visual Presentation (RSVP) based Brain-Computer Interface (BCI) facilities the high-throughput detection of rare target images by detecting evoked event-related potentials (ERPs). At present, the decoding accuracy of the RSVP-based BCI system limits its practical applications. This study introduces eye movements (gaze and pupil information), referred to as EYE modality, as another useful source of information to combine with EEG-based BCI and forms a novel target detection system to detect target images in RSVP tasks. We performed an RSVP experiment, recorded the EEG signals and eye movements simultaneously during a target detection task, and constructed a multi-modal dataset including 20 subjects. Also, we proposed a cross-modal guiding and fusion network to fully utilize EEG and EYE modalities and fuse them for better RSVP decoding performance. In this network, a two-branch backbone was built to extract features from these two modalities. A Cross-Modal Feature Guiding (CMFG) module was proposed to guide EYE modality features to complement the EEG modality for better feature extraction. A Multi-scale Multi-modal Reweighting (MMR) module was proposed to enhance the multi-modal features by exploring intra- and inter-modal interactions. And, a Dual Activation Fusion (DAF) was proposed to modulate the enhanced multi-modal features for effective fusion. Our proposed network achieved a balanced accuracy of 88.00% (±2.29) on the collected dataset. The ablation studies and visualizations revealed the effectiveness of the proposed modules. This work implies the effectiveness of introducing the EYE modality in RSVP tasks. And, our proposed network is a promising method for RSVP decoding and further improves the performance of RSVP-based target detection systems.
Collapse
Affiliation(s)
- Jiayu Mao
- Laboratory of Brain Atlas and Brain-Inspired Intelligence, State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China; School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Shuang Qiu
- Laboratory of Brain Atlas and Brain-Inspired Intelligence, State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China; School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Wei Wei
- Laboratory of Brain Atlas and Brain-Inspired Intelligence, State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Huiguang He
- Laboratory of Brain Atlas and Brain-Inspired Intelligence, State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China; School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China.
| |
Collapse
|
7
|
Du P, Li P, Cheng L, Li X, Su J. Single-trial P300 classification algorithm based on centralized multi-person data fusion CNN. Front Neurosci 2023; 17:1132290. [PMID: 36908799 PMCID: PMC9992797 DOI: 10.3389/fnins.2023.1132290] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Accepted: 02/01/2023] [Indexed: 02/24/2023] Open
Abstract
Introduction Currently, it is still a challenge to detect single-trial P300 from electroencephalography (EEG) signals. In this paper, to address the typical problems faced by existing single-trial P300 classification, such as complex, time-consuming and low accuracy processes, a single-trial P300 classification algorithm based on multiplayer data fusion convolutional neural network (CNN) is proposed to construct a centralized collaborative brain-computer interfaces (cBCI) for fast and highly accurate classification of P300 EEG signals. Methods In this paper, two multi-person data fusion methods (parallel data fusion and serial data fusion) are used in the data pre-processing stage to fuse multi-person EEG information stimulated by the same task instructions, and then the fused data is fed as input to the CNN for classification. In building the CNN network for single-trial P300 classification, the Conv layer was first used to extract the features of single-trial P300, and then the Maxpooling layer was used to connect the Flatten layer for secondary feature extraction and dimensionality reduction, thereby simplifying the computation. Finally batch normalisation is used to train small batches of data in order to better generalize the network and speed up single-trial P300 signal classification. Results In this paper, the above new algorithms were tested on the Kaggle dataset and the Brain-Computer Interface (BCI) Competition III dataset, and by analyzing the P300 waveform features and EEG topography and the four standard evaluation metrics, namely Accuracy, Precision, Recall and F1-score,it was demonstrated that the single-trial P300 classification algorithm after two multi-person data fusion CNNs significantly outperformed other classification algorithms. Discussion The results show that the single-trial P300 classification algorithm after two multi-person data fusion CNNs significantly outperformed the single-person model, and that the single-trial P300 classification algorithm with two multi-person data fusion CNNs involves smaller models, fewer training parameters, higher classification accuracy and improves the overall P300-cBCI classification rate and actual performance more effectively with a small amount of sample information compared to other algorithms.
Collapse
Affiliation(s)
- Pu Du
- School of Integrated Circuit Science and Engineering, Tianjin University of Technology, Tianjin, China
| | - Penghai Li
- School of Integrated Circuit Science and Engineering, Tianjin University of Technology, Tianjin, China
| | - Longlong Cheng
- School of Integrated Circuit Science and Engineering, Tianjin University of Technology, Tianjin, China.,China Electronics Cloud Brain Technology Co., Ltd., Tianjin, China
| | - Xueqing Li
- School of Integrated Circuit Science and Engineering, Tianjin University of Technology, Tianjin, China
| | - Jianxian Su
- School of Integrated Circuit Science and Engineering, Tianjin University of Technology, Tianjin, China
| |
Collapse
|
8
|
Gannouni S, Belwafi K, Alangari N, AboAlsamh H, Belghith A. Classification Strategies for P300-Based BCI-Spellers Adopting the Row Column Paradigm. SENSORS (BASEL, SWITZERLAND) 2022; 22:9159. [PMID: 36501860 PMCID: PMC9738959 DOI: 10.3390/s22239159] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/19/2022] [Revised: 11/14/2022] [Accepted: 11/22/2022] [Indexed: 06/17/2023]
Abstract
Acknowledging the importance of the ability to communicate with other people, the researcher community has developed a series of BCI-spellers, with the goal of regaining communication and interaction capabilities with the environment for people with disabilities. In order to bridge the gap in the digital divide between the disabled and the non-disabled people, we believe that the development of efficient signal processing algorithms and strategies will go a long way towards achieving novel assistive technologies using new human-computer interfaces. In this paper, we present various classification strategies that would be adopted by P300 spellers adopting the row/column paradigm. The presented strategies have obtained high accuracy rates compared with existent similar research works.
Collapse
Affiliation(s)
- Sofien Gannouni
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
| | - Kais Belwafi
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
- Electrical Engineering and Computer Science Department, Khalifa University, Abu Dhabi P.O. Box 127788, United Arab Emirates
| | - Nourah Alangari
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
| | - Hatim AboAlsamh
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
| | - Abdelfettah Belghith
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
| |
Collapse
|
9
|
Lu R, Zeng Y, Zhang R, Yan B, Tong L. SAST-GCN: Segmentation Adaptive Spatial Temporal-Graph Convolutional Network for P3-Based Video Target Detection. Front Neurosci 2022; 16:913027. [PMID: 35720707 PMCID: PMC9201684 DOI: 10.3389/fnins.2022.913027] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2022] [Accepted: 05/12/2022] [Indexed: 11/27/2022] Open
Abstract
Detecting video-induced P3 is crucial to building the video target detection system based on the brain-computer interface. However, studies have shown that the brain response patterns corresponding to video-induced P3 are dynamic and determined by the interaction of multiple brain regions. This paper proposes a segmentation adaptive spatial-temporal graph convolutional network (SAST-GCN) for P3-based video target detection. To make full use of the dynamic characteristics of the P3 signal data, the data is segmented according to the processing stages of the video-induced P3, and the brain network connections are constructed correspondingly. Then, the spatial-temporal feature of EEG data is extracted by adaptive spatial-temporal graph convolution to discriminate the target and non-target in the video. Especially, a style-based recalibration module is added to select feature maps with higher contributions and increase the feature extraction ability of the network. The experimental results demonstrate the superiority of our proposed model over the baseline methods. Also, the ablation experiments indicate that the segmentation of data to construct the brain connection can effectively improve the recognition performance by reflecting the dynamic connection relationship between EEG channels more accurately.
Collapse
Affiliation(s)
- Runnan Lu
- Henan Key Laboratory of Imaging and Intelligent Processing, People’s Liberation Army of China (PLA) Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Ying Zeng
- Henan Key Laboratory of Imaging and Intelligent Processing, People’s Liberation Army of China (PLA) Strategic Support Force Information Engineering University, Zhengzhou, China
- Key Laboratory for Neuroinformation of Ministry of Education, School of Life Sciences and Technology, University of Electronic Science and Technology of China, Chengdu, China
| | - Rongkai Zhang
- Henan Key Laboratory of Imaging and Intelligent Processing, People’s Liberation Army of China (PLA) Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Bin Yan
- Henan Key Laboratory of Imaging and Intelligent Processing, People’s Liberation Army of China (PLA) Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Li Tong
- Henan Key Laboratory of Imaging and Intelligent Processing, People’s Liberation Army of China (PLA) Strategic Support Force Information Engineering University, Zhengzhou, China
| |
Collapse
|
10
|
Zhang H, Wang Z, Yu Y, Yin H, Chen C, Wang H. An improved EEGNet for single-trial EEG classification in rapid serial visual presentation task. BRAIN SCIENCE ADVANCES 2022. [DOI: 10.26599/bsa.2022.9050007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022] Open
Abstract
As a new type of brain–computer interface (BCI), the rapid serial visual presentation (RSVP) paradigm has attracted significant attention. The mechanism of RSVP is detecting the P300 component corresponding to the target image to realize fast and correct recognition. This paper proposed an improved EEGNet model to achieve good performance in offline and online data. Specifically, the data were filtered by xDAWN to enhance the signal-to-noise ratio of the electroencephalogram (EEG) signals. The focal loss function was used instead of the cross-entropy loss function to solve the classification problems of unbalanced samples. Additionally, the subject-specific data were fed to the improved EEGNet model to obtain a subject-specific model. We applied the proposed model at the BCI Controlled Robot Contest in World Robot Contest 2021 and won the second place. The average recall rate of the four participants reached 51.56% in triple classification. In the offline data benchmark dataset (64 subjects-RSVP tasks), the average recall rates of groups A and B reached 76.07% and 78.11%, respectively. We provided an alternative method to identify targets based on the RSVP paradigm.
Collapse
Affiliation(s)
- Hongfei Zhang
- Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, Guangdong, China
| | - Zehui Wang
- Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, Guangdong, China
| | - Yinhu Yu
- Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, Guangdong, China
| | - Haojun Yin
- Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, Guangdong, China
| | - Chuangquan Chen
- Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, Guangdong, China
| | - Hongtao Wang
- Faculty of Intelligent Manufacturing, Wuyi University, Jiangmen 529020, Guangdong, China
| |
Collapse
|
11
|
Singanamalla SKR, Lin CT. Spike-Representation of EEG Signals for Performance Enhancement of Brain-Computer Interfaces. Front Neurosci 2022; 16:792318. [PMID: 35444515 PMCID: PMC9014221 DOI: 10.3389/fnins.2022.792318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2021] [Accepted: 01/12/2022] [Indexed: 11/13/2022] Open
Abstract
Brain-computer interfaces (BCI) relying on electroencephalography (EEG) based neuroimaging mode has shown prospects for real-world usage due to its portability and optional selectivity of fewer channels for compactness. However, noise and artifacts often limit the capacity of BCI systems especially for event-related potentials such as P300 and error-related negativity (ERN), whose biomarkers are present in short time segments at the time-series level. Contrary to EEG, invasive recording is less prone to noise but requires a tedious surgical procedure. But EEG signal is the result of aggregation of neuronal spiking information underneath the scalp surface and transforming the relevant BCI task's EEG signal to spike representation could potentially help improve the BCI performance. In this study, we designed an approach using a spiking neural network (SNN) which is trained using surrogate-gradient descent to generate task-related multi-channel EEG template signals of all classes. The trained model is in turn leveraged to obtain the latent spike representation for each EEG sample. Comparing the classification performance of EEG signal and its spike-representation, the proposed approach enhanced the performance of ERN dataset from 79.22 to 82.27% with naive bayes and for P300 dataset, the accuracy was improved from 67.73 to 69.87% using xGboost. In addition, principal component analysis and correlation metrics were evaluated on both EEG signals and their spike-representation to identify the reason for such improvement.
Collapse
Affiliation(s)
- Sai Kalyan Ranga Singanamalla
- Computational Intelligence and Brain Computer Interface Lab, School of Computer Science, University of Technology Sydney, Sydney, NSW, Australia
| | - Chin-Teng Lin
- Computational Intelligence and Brain Computer Interface Lab, School of Computer Science, University of Technology Sydney, Sydney, NSW, Australia
- Australian Artificial Intelligence Institute, University of Technology Sydney, Sydney, NSW, Australia
- *Correspondence: Chin-Teng Lin
| |
Collapse
|