1
|
Lei D, Dong C, Guo H, Ma P, Liu H, Bao N, Kang H, Chen X, Wu Y. A fused multi-subfrequency bands and CBAM SSVEP-BCI classification method based on convolutional neural network. Sci Rep 2024; 14:8616. [PMID: 38616204 PMCID: PMC11016546 DOI: 10.1038/s41598-024-59348-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Accepted: 04/09/2024] [Indexed: 04/16/2024] Open
Abstract
For the brain-computer interface (BCI) system based on steady-state visual evoked potential (SSVEP), it is difficult to obtain satisfactory classification performance for short-time window SSVEP signals by traditional methods. In this paper, a fused multi-subfrequency bands and convolutional block attention module (CBAM) classification method based on convolutional neural network (CBAM-CNN) is proposed for discerning SSVEP-BCI tasks. This method extracts multi-subfrequency bands SSVEP signals as the initial input of the network model, and then carries out feature fusion on all feature inputs. In addition, CBAM is embedded in both parts of the initial input and feature fusion for adaptive feature refinement. To verify the effectiveness of the proposed method, this study uses the datasets of Inner Mongolia University of Technology (IMUT) and Tsinghua University (THU) to evaluate the performance of the proposed method. The experimental results show that the highest accuracy of CBAM-CNN reaches 0.9813 percentage point (pp). Within 0.1-2 s time window, the accuracy of CBAM-CNN is 0.0201-0.5388 (pp) higher than that of CNN, CCA-CWT-SVM, CCA-SVM, CCA-GNB, FBCCA, and CCA. Especially in the short-time window range of 0.1-1 s, the performance advantage of CBAM-CNN is more significant. The maximum information transmission rate (ITR) of CBAM-CNN is 503.87 bit/min, which is 227.53 bit/min-503.41 bit/min higher than the above six EEG decoding methods. The study further results show that CBAM-CNN has potential application value in SSVEP decoding.
Collapse
Affiliation(s)
- Dongyang Lei
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China
| | - Chaoyi Dong
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China.
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China.
- Engineering Research Center of Large Energy Storage Technology, Ministry of Education, Hohhot, 010080, China.
- Inner Mongolia Academy of Science and Technology, Hohhot, 010010, China.
| | - Hongfei Guo
- Inner Mongolia Academy of Science and Technology, Hohhot, 010010, China.
| | - Pengfei Ma
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China
| | - Huanzi Liu
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China
| | - Naqin Bao
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China
| | - Hongzhuo Kang
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China
| | - Xiaoyan Chen
- College of Electric Power, Inner Mongolia University of Technology, Hohhot, 010080, China
- Intelligent Energy Technology and Equipment Engineering Research Centre of Colleges and Universities in Inner Mongolia Autonomous Region, Hohhot, 010051, China
- Engineering Research Center of Large Energy Storage Technology, Ministry of Education, Hohhot, 010080, China
- Inner Mongolia Academy of Science and Technology, Hohhot, 010010, China
| | - Yi Wu
- Inner Mongolia Academy of Science and Technology, Hohhot, 010010, China
| |
Collapse
|
2
|
Mai X, Ai J, Wei Y, Zhu X, Meng J. Phase-Locked Time-Shift Data Augmentation Method for SSVEP Brain-Computer Interfaces. IEEE Trans Neural Syst Rehabil Eng 2023; 31:4096-4105. [PMID: 37815966 DOI: 10.1109/tnsre.2023.3323351] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/12/2023]
Abstract
Steady-state visual evoked potential (SSVEP) based brain-computer interfaces (BCIs) have achieved an information transfer rate (ITR) of over 300 bits/min, but abundant training data is required. The performance of SSVEP algorithms deteriorates greatly under limited data, and the existing time-shift data augmentation method fails to improve it because the phase-locked requirement between training samples is violated. To address this issue, this study proposes a novel augmentation method, namely phase-locked time-shift (PLTS), for SSVEP-BCI. The similarity between epochs at different time moments was evaluated, and a unique time-shift step was calculated for each class to augment additional data epochs in each trial. The results showed that the PLTS significantly improved the classification performance of SSVEP algorithms on the BETA SSVEP datasets. Moreover, under the condition of one calibration block, by slightly prolonging the calibration duration (from 48 s to 51.5 s), the ITR increased from 40.88±4.54 bits/min to 122.61±7.05 bits/min with the PLTS. This study provides a new perspective on augmenting data epochs for training-based SSVEP-BCI, promotes the classification accuracy and ITR under limited training data, and thus facilitates the real-life applications of SSVEP-based brain spellers.
Collapse
|
3
|
Swami P, Gramann K, Vonstad EK, Vereijken B, Holt A, Holt T, Sandstrak G, Nilsen JH, Su X. CLET: Computation of Latencies in Event-related potential Triggers using photodiode on virtual reality apparatuses. Front Hum Neurosci 2023; 17:1223774. [PMID: 37795210 PMCID: PMC10546026 DOI: 10.3389/fnhum.2023.1223774] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2023] [Accepted: 08/31/2023] [Indexed: 10/06/2023] Open
Abstract
To investigate event-related activity in human brain dynamics as measured with EEG, triggers must be incorporated to indicate the onset of events in the experimental protocol. Such triggers allow for the extraction of ERP, i.e., systematic electrophysiological responses to internal or external stimuli that must be extracted from the ongoing oscillatory activity by averaging several trials containing similar events. Due to the technical setup with separate hardware sending and recording triggers, the recorded data commonly involves latency differences between the transmitted and received triggers. The computation of these latencies is critical for shifting the epochs with respect to the triggers sent. Otherwise, timing differences can lead to a misinterpretation of the resulting ERPs. This study presents a methodical approach for the CLET using a photodiode on a non-immersive VR (i.e., LED screen) and an immersive VR (i.e., HMD). Two sets of algorithms are proposed to analyze the photodiode data. The experiment designed for this study involved the synchronization of EEG, EMG, PPG, photodiode sensors, and ten 3D MoCap cameras with a VR presentation platform (Unity). The average latency computed for LED screen data for a set of white and black stimuli was 121.98 ± 8.71 ms and 121.66 ± 8.80 ms, respectively. In contrast, the average latency computed for HMD data for the white and black stimuli sets was 82.80 ± 7.63 ms and 69.82 ± 5.52 ms. The codes for CLET and analysis, along with datasets, tables, and a tutorial video for using the codes, have been made publicly available.
Collapse
Affiliation(s)
- Piyush Swami
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
- Section for Visual Computing, Department of Applied Mathematics and Computer Science, Technical University of Denmark, Kongens Lyngby, Denmark
- Biomedical Engineering Techies, Broendby, Denmark
| | - Klaus Gramann
- Biological Psychology and Neuroergonomics, Technical University of Berlin, Berlin, Germany
| | - Elise Klæbo Vonstad
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| | - Beatrix Vereijken
- Department of Neuromedicine and Movement Science, Norwegian University of Science and Technology, Trondheim, Norway
| | - Alexander Holt
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| | - Tomas Holt
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| | - Grethe Sandstrak
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| | - Jan Harald Nilsen
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| | - Xiaomeng Su
- Motion Capture and Visualization Laboratory, Applied Information Technology Group, Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| |
Collapse
|
4
|
Huang J, Zhang ZQ, Xiong B, Wang Q, Wan B, Li F, Yang P. Cross-Subject Transfer Method Based on Domain Generalization for Facilitating Calibration of SSVEP-Based BCIs. IEEE Trans Neural Syst Rehabil Eng 2023; 31:3307-3319. [PMID: 37578926 DOI: 10.1109/tnsre.2023.3305202] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/16/2023]
Abstract
In steady-state visual evoked potential (SSVEP)-based brain-computer interfaces (BCIs), various spatial filtering methods based on individual calibration data have been proposed to alleviate the interference of spontaneous activities in SSVEP signals for enhancing the SSVEP detection performance. However, the time-consuming calibration session would increase the visual fatigue of subjects and reduce the usability of the BCI system. The key idea of this study is to propose a cross-subject transfer method based on domain generalization, which transfers the domain-invariant spatial filters and templates learned from source subjects to the target subject with no access to the EEG data from the target subject. The transferred spatial filters and templates are obtained by maximizing the intra- and inter-subject correlations using the SSVEP data corresponding to the target and its neighboring stimuli. For SSVEP detection of the target subject, four types of correlation coefficients are calculated to construct the feature vector. Experimental results estimated with three SSVEP datasets show that the proposed cross-subject transfer method improves the SSVEP detection performance compared to state-of-art methods. The satisfactory results demonstrate that the proposed method provides an effective transfer learning strategy requiring no tedious data collection process for new users, holding the potential of promoting practical applications of SSVEP-based BCI.
Collapse
|
5
|
Pan Y, Chen J, Zhang Y, Zhang Y. An efficient CNN-LSTM Network with spectral normalization and label smoothing technologies for SSVEP frequency recognition. J Neural Eng 2022; 19. [PMID: 36041426 DOI: 10.1088/1741-2552/ac8dc5] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2022] [Accepted: 08/30/2022] [Indexed: 11/12/2022]
Abstract
OBJECTIVE Steady-state visual evoked potentials(SSVEPs) based braincomputer interface(BCI) has received great interests owing to the high information transfer rate(ITR) and available large number of targets. However, the performance of frequency recognition methods heavily depends on the amount of the calibration data for intra-subject classification. Some research adopted the deep learning(DL) algorithm to conduct the inter-subject classification, which could reduce the calculation procedure, but the performance still has large room to improve compared with the intra-subject classification. APPROACH To address these issues, we proposed an efficient SSVEP DL NETwork (termed SSVEPNET) based on 1D convolution and long short-term memory (LSTM) module. To enhance the performance of SSVEPNT, we adopted the spectral normalization and label smoothing technologies during implementing the network architecture. We evaluated the SSVEPNET and compared it with other methods for the intra- and inter-subject classification under different conditions, i.e., two datasets, two time-window lengths (1 s and 0.5 s), three sizes of training data. MAIN RESULTS Under all the experimental settings, the proposed SSVEPNET achieved the highest average accuracy for the intra- and inter-subject classification on the two SSVEP datasets, when compared with other traditional and DL baseline methods. Signif icance. The extensive experimental results demonstrate that the proposed DL model holds promise to enhance frequency recognition performance in SSVEP-based BCIs. Besides, the mixed network structures with CNN and LSTM, and the spectral normalization and label smoothing could be useful optimization strategies to design efficient models for EEG data.
Collapse
Affiliation(s)
- YuDong Pan
- Laboratory for Brain Science and Medical Artificial Intelligence, Southwest University of Science and Technology, Mianyang,CN,621010, Mianyang, 621010, CHINA
| | - Jianbo Chen
- Laboratory for Brain Science and Medical Artificial Intelligence, Southwest University of Science and Technology, Mianyang 621010, China, Mianyang, 621010, CHINA
| | - Yangsong Zhang
- School of Computer Science and Technology, Southwest University of Science and Technology, Mianyang,CN,621010, Mianyang, 621010, CHINA
| | - Yu Zhang
- Department of Bioengineering, Lehigh University, Bethlehem, PA 18015, USA, Bethlehem, 18015-3027, UNITED STATES
| |
Collapse
|