1
|
Zhou Z, Lin M, Zhou X, Zhang C. Implementation of memristive emotion associative learning circuit. Cogn Neurodyn 2025; 19:13. [PMID: 39801920 PMCID: PMC11717764 DOI: 10.1007/s11571-024-10211-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2024] [Revised: 07/10/2024] [Accepted: 09/23/2024] [Indexed: 01/16/2025] Open
Abstract
Psychological studies have demonstrated that the music can affect memory by triggering different emotions. Building on the relationships among music, emotion, and memory, a memristor-based emotion associative learning circuit is designed by utilizing the nonlinear and non-volatile characteristics of memristors, which includes a music judgment module, three emotion generation modules, three emotional homeostasis modules, and a memory module to implement functions such as learning, second learning, forgetting, emotion generation, and emotional homeostasis. The experimental results indicate that the proposed circuit can simulate the learning and forgetting processes of human under different music circumstances, demonstrate the feasibility of memristors in biomimetic circuits, verify the impact of music on memory, and provide a foundation for in-depth research and application development of the interaction mechanism between emotion and memory.
Collapse
Affiliation(s)
- Zhangzhi Zhou
- School of Electronics and Information, Hangzhou Dianzi University, Hangzhou, 310018 China
| | - Mi Lin
- School of Electronics and Information, Hangzhou Dianzi University, Hangzhou, 310018 China
| | - Xuanxuan Zhou
- School of Electronics and Information, Hangzhou Dianzi University, Hangzhou, 310018 China
| | - Chong Zhang
- School of Electronics and Information, Hangzhou Dianzi University, Hangzhou, 310018 China
| |
Collapse
|
2
|
Yan R, Lu N, Yan Y, Niu X, Wu J. A Fine-grained Hemispheric Asymmetry Network for accurate and interpretable EEG-based emotion classification. Neural Netw 2025; 184:107127. [PMID: 39809039 DOI: 10.1016/j.neunet.2025.107127] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2024] [Revised: 12/15/2024] [Accepted: 01/03/2025] [Indexed: 01/16/2025]
Abstract
In this work, we propose a Fine-grained Hemispheric Asymmetry Network (FG-HANet), an end-to-end deep learning model that leverages hemispheric asymmetry features within 2-Hz narrow frequency bands for accurate and interpretable emotion classification over raw EEG data. In particular, the FG-HANet extracts features not only from original inputs but also from their mirrored versions, and applies Finite Impulse Response (FIR) filters at a granularity as fine as 2-Hz to acquire fine-grained spectral information. Furthermore, to guarantee sufficient attention to hemispheric asymmetry features, we tailor a three-stage training pipeline for the FG-HANet to further boost its performance. We conduct extensive evaluations on two public datasets, SEED and SEED-IV, and experimental results well demonstrate the superior performance of the proposed FG-HANet, i.e. 97.11% and 85.70% accuracy, respectively, building a new state-of-the-art. Our results also reveal the hemispheric dominance under different emotional states and the hemisphere asymmetry within 2-Hz frequency bands in individuals. These not only align with previous findings in neuroscience but also provide new insights into underlying emotion generation mechanisms.
Collapse
Affiliation(s)
- Ruofan Yan
- Systems Engineering Institute, School of Automation Science and Engineering, Xi'an Jiaotong University, People's Republic of China; Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region
| | - Na Lu
- Systems Engineering Institute, School of Automation Science and Engineering, Xi'an Jiaotong University, People's Republic of China.
| | - Yuxuan Yan
- Systems Engineering Institute, School of Automation Science and Engineering, Xi'an Jiaotong University, People's Republic of China
| | - Xu Niu
- Systems Engineering Institute, School of Automation Science and Engineering, Xi'an Jiaotong University, People's Republic of China
| | - Jibin Wu
- Department of Data Science and Artificial Intelligence, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region; Department of Computing, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region.
| |
Collapse
|
3
|
Chen Q, Mao X, Song Y, Wang K. An EEG-based emotion recognition method by fusing multi-frequency-spatial features under multi-frequency bands. J Neurosci Methods 2025; 415:110360. [PMID: 39778774 DOI: 10.1016/j.jneumeth.2025.110360] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2024] [Revised: 12/23/2024] [Accepted: 01/03/2025] [Indexed: 01/11/2025]
Abstract
BACKGROUND Recognition of emotion changes is of great significance to a person's physical and mental health. At present, EEG-based emotion recognition methods are mainly focused on time or frequency domains, but rarely on spatial information. Therefore, the goal of this study is to improve the performance of emotion recognition by integrating frequency and spatial domain information under multi-frequency bands. NEW METHODS Firstly, EEG signals of four frequency bands are extracted, and then three frequency-spatial features of differential entropy (DE) symmetric difference (SD) and symmetric quotient (SQ) are separately calculated. Secondly, according to the distribution of EEG electrodes, a series of brain maps are constructed by three frequency-spatial features for each frequency band. Thirdly, a Multi-Parallel-Input Convolutional Neural Network (MPICNN) uses the constructed brain maps to train and obtain the emotion recognition model. Finally, the subject-dependent experiments are conducted on DEAP and SEED-IV datasets. RESULTS The experimental results of DEAP dataset show that the average accuracy of four-class emotion recognition, namely, high-valence high-arousal, high-valence low-arousal, low-valence high-arousal and low-valence low-arousal, reaches 98.71 %. The results of SEED-IV dataset show the average accuracy of four-class emotion recognition, namely, happy, sad, neutral and fear reaches 92.55 %. COMPARISON WITH EXISTING METHODS This method has a best classification performance compared with the state-of-the-art methods on both four-class emotion recognition datasets. CONCLUSIONS This EEG-based emotion recognition method fused multi-frequency-spatial features under multi-frequency bands, and effectively improved the recognition performance compared with the existing methods.
Collapse
Affiliation(s)
- Qiuyu Chen
- College of Automation and Electronic Engineering, Qingdao University of Science and Technology, Qingdao, China
| | - Xiaoqian Mao
- College of Automation and Electronic Engineering, Qingdao University of Science and Technology, Qingdao, China.
| | - Yuebin Song
- College of Automation and Electronic Engineering, Qingdao University of Science and Technology, Qingdao, China
| | - Kefa Wang
- College of Automation and Electronic Engineering, Qingdao University of Science and Technology, Qingdao, China
| |
Collapse
|
4
|
Yan F, Guo Z, Iliyasu AM, Hirota K. Multi-branch convolutional neural network with cross-attention mechanism for emotion recognition. Sci Rep 2025; 15:3976. [PMID: 39893256 PMCID: PMC11787301 DOI: 10.1038/s41598-025-88248-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Accepted: 01/28/2025] [Indexed: 02/04/2025] Open
Abstract
Research on emotion recognition is an interesting area because of its wide-ranging applications in education, marketing, and medical fields. This study proposes a multi-branch convolutional neural network model based on cross-attention mechanism (MCNN-CA) for accurate recognition of different emotions. The proposed model provides automated extraction of relevant features from multimodal data and fusion of feature maps from diverse sources as modules for the subsequent emotion recognition. In the feature extraction stage, various convolutional neural networks were designed to extract critical information from multiple dimensional features. The feature fusion module was used to enhance the inter-correlation between features based on channel-efficient attention mechanism. This innovation proves effective in fusing distinctive features within a single mode and across different modes. The model was assessed based on EEG emotion recognition experiments on the SEED and SEED-IV datasets. Furthermore, the efficiency of the proposed model was evaluated via multimodal emotion experiments using EEG and text data from the ZuCo dataset. Comparative analysis alongside contemporary studies shows that our model excels in terms of accuracy, precision, recall, and F1-score.
Collapse
Affiliation(s)
- Fei Yan
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, China
| | - Zekai Guo
- School of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, China
| | - Abdullah M Iliyasu
- College of Engineering, Prince Sattam Bin Abdulaziz University, Al-Kharj, 11942, Saudi Arabia.
- School of Computing, Tokyo Institute of Technology, Yokohama, 226-8502, Japan.
| | - Kaoru Hirota
- School of Computing, Tokyo Institute of Technology, Yokohama, 226-8502, Japan
- School of Automation, Beijing Institute of Technology, Beijing, 100081, China
| |
Collapse
|
5
|
Xiong H, Yan Y, Chen Y, Liu J. Graph convolution network-based eeg signal analysis: a review. Med Biol Eng Comput 2025:10.1007/s11517-025-03295-0. [PMID: 39883372 DOI: 10.1007/s11517-025-03295-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2024] [Accepted: 01/07/2025] [Indexed: 01/31/2025]
Abstract
With the advancement of artificial intelligence technology, more and more effective methods are being used to identify and classify Electroencephalography (EEG) signals to address challenges in healthcare and brain-computer interface fields. The applications and major achievements of Graph Convolution Network (GCN) techniques in EEG signal analysis are reviewed in this paper. Through an exhaustive search of the published literature, a module-by-module discussion is carried out for the first time to address the current research status of GCN. An exhaustive classification of methods and a systematic analysis of key modules, such as brain map construction, node feature extraction, and GCN architecture design, are presented. In addition, we pay special attention to several key research issues related to GCN. This review enhances the understanding of the future potential of GCN in the field of EEG signal analysis. At the same time, several valuable development directions are sorted out for researchers in related fields, such as analysing the applicability of different GCN layers, building task-oriented GCN models, and improving adaptation to limited data.
Collapse
Affiliation(s)
- Hui Xiong
- School of Control Science and Engineering, Tiangong University, Tianjin, 300387, China.
- Key Laboratory of Intelligent Control of Electrical Equipment, Tiangong University, Tianjin, 300387, China.
| | - Yan Yan
- Key Laboratory of Intelligent Control of Electrical Equipment, Tiangong University, Tianjin, 300387, China
- School of Artificial Intelligence, Tiangong University, Tianjin, 300387, China
| | - Yimei Chen
- School of Control Science and Engineering, Tiangong University, Tianjin, 300387, China
| | - Jinzhen Liu
- School of Control Science and Engineering, Tiangong University, Tianjin, 300387, China
- Key Laboratory of Intelligent Control of Electrical Equipment, Tiangong University, Tianjin, 300387, China
| |
Collapse
|
6
|
Liu J, He L, Chen H, Jiang D. Directional Spatial and Spectral Attention Network (DSSA Net) for EEG-based emotion recognition. Front Neurorobot 2025; 18:1481746. [PMID: 39840233 PMCID: PMC11747697 DOI: 10.3389/fnbot.2024.1481746] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2024] [Accepted: 12/09/2024] [Indexed: 01/23/2025] Open
Abstract
Significant strides have been made in emotion recognition from Electroencephalography (EEG) signals. However, effectively modeling the diverse spatial, spectral, and temporal features of multi-channel brain signals remains a challenge. This paper proposes a novel framework, the Directional Spatial and Spectral Attention Network (DSSA Net), which enhances emotion recognition accuracy by capturing critical spatial-spectral-temporal features from EEG signals. The framework consists of three modules: Positional Attention (PA), Spectral Attention (SA), and Temporal Attention (TA). The PA module includes Vertical Attention (VA) and Horizontal Attention (HA) branches, designed to detect active brain regions from different orientations. Experimental results on three benchmark EEG datasets demonstrate that DSSA Net outperforms most competitive methods. On the SEED and SEED-IV datasets, it achieves accuracies of 96.61% and 85.07% for subject-dependent emotion recognition, respectively, and 87.03% and 75.86% for subject-independent recognition. On the DEAP dataset, it attains accuracies of 94.97% for valence and 94.73% for arousal. These results showcase the framework's ability to leverage both spatial and spectral differences across brain hemispheres and regions, enhancing classification accuracy for emotion recognition.
Collapse
Affiliation(s)
- Jiyao Liu
- School of Computer Science, Northwestern Polytechnical University, Xi'an, China
| | - Lang He
- School of Computer Science and Technology, Xi'an University of Posts and Telecommunications, Xi'an, Shaanxi, China
| | - Haifeng Chen
- School of Electronic Information and Artificial Intelligence, Shaanxi University of Science and Technology, Xi'an, China
| | - Dongmei Jiang
- School of Computer Science, Northwestern Polytechnical University, Xi'an, China
| |
Collapse
|
7
|
Cheng Z, Bu X, Wang Q, Yang T, Tu J. EEG-based emotion recognition using multi-scale dynamic CNN and gated transformer. Sci Rep 2024; 14:31319. [PMID: 39733023 DOI: 10.1038/s41598-024-82705-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2024] [Accepted: 12/09/2024] [Indexed: 12/30/2024] Open
Abstract
Emotions play a crucial role in human thoughts, cognitive processes, and decision-making. EEG has become a widely utilized tool in emotion recognition due to its high temporal resolution, real-time monitoring capabilities, portability, and cost-effectiveness. In this paper, we propose a novel end-to-end emotion recognition method from EEG signals, called MSDCGTNet, which is based on the Multi-Scale Dynamic 1D CNN and the Gated Transformer. First, the Multi-Scale Dynamic CNN is used to extract complex spatial and spectral features from raw EEG signals, which not only avoids information loss but also reduces computational costs associated with the time-frequency conversion of signals. Then, the Gated Transformer Encoder is utilized to capture global dependencies of EEG signals. This encoder focuses on specific regions of the input sequence while reducing computational resources through parallel processing with the improved multi-head self-attention mechanisms. Third, the Temporal Convolution Network is used to extract temporal features from the EEG signals. Finally, the extracted abstract features are fed into a classification module for emotion recognition. The proposed method was evaluated on three publicly available datasets: DEAP, SEED, and SEED_IV. Experimental results demonstrate the high accuracy and efficiency of the proposed method for emotion recognition. This approach proves to be robust and suitable for various practical applications. By addressing challenges posed by existing methods, the proposed method provides a valuable and effective solution for the field of Brain-Computer Interface (BCI).
Collapse
Affiliation(s)
- Zhuoling Cheng
- School of Electronic Information and Electrical Engineering, Yangtze University, Jingzhou, 434100, Hubei, China
| | - Xuekui Bu
- School of Electronic Information and Electrical Engineering, Yangtze University, Jingzhou, 434100, Hubei, China
| | - Qingnan Wang
- School of Physics, Electronics and Intelligent Manufacturing, Huaihua University, Hunan, 418000, China
| | - Tao Yang
- Department of Neurology, Jingzhou First People's Hospital, Jingzhou, 434000, Hubei, China
| | - Jihui Tu
- School of Electronic Information and Electrical Engineering, Yangtze University, Jingzhou, 434100, Hubei, China.
| |
Collapse
|
8
|
Lu W, Xia L, Tan TP, Ma H. CIT-EmotionNet: convolution interactive transformer network for EEG emotion recognition. PeerJ Comput Sci 2024; 10:e2610. [PMID: 39896395 PMCID: PMC11784834 DOI: 10.7717/peerj-cs.2610] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2024] [Accepted: 11/25/2024] [Indexed: 02/04/2025]
Abstract
Emotion recognition is a significant research problem in affective computing as it has a lot of potential areas of application. One of the approaches in emotion recognition uses electroencephalogram (EEG) signals to identify the emotion of a person. However, effectively using the global and local features of EEG signals to improve the performance of emotion recognition is still a challenge. In this study, we propose a novel Convolution Interactive Transformer Network for EEG Emotion Recognition, known as CIT-EmotionNet, which efficiently integrates the global and local features of EEG signals. We convert the raw EEG signals into spatial-spectral representations, which serve as the inputs into the model. The model integrates convolutional neural network (CNN) and Transformer within a single framework in a parallel manner. We propose a Convolution Interactive Transformer module, which facilitates the interaction and fusion of local and global features extracted by CNN and Transformer respectively, thereby improving the average accuracy of emotion recognition. The proposed CIT-EmotionNet outperforms state-of-the-art methods, achieving an average recognition accuracy of 98.57% and 92.09% on two publicly available datasets, SEED and SEED-IV, respectively.
Collapse
Affiliation(s)
- Wei Lu
- Henan High-Speed Railway Operation and Maintenance Engineering Research Center, Zhengzhou Railway Vocational and Technical College, Zhengzhou, Henan, China
- School of Computer Sciences, Universiti Sains Malaysia, USM, Pulau Pinang, Malaysia
- Zhengzhou University Industrial Technology Research Institute, Zhengzhou, Henan, China
| | - Lingnan Xia
- Henan High-Speed Railway Operation and Maintenance Engineering Research Center, Zhengzhou Railway Vocational and Technical College, Zhengzhou, Henan, China
| | - Tien Ping Tan
- School of Computer Sciences, Universiti Sains Malaysia, USM, Pulau Pinang, Malaysia
| | - Hua Ma
- Henan High-Speed Railway Operation and Maintenance Engineering Research Center, Zhengzhou Railway Vocational and Technical College, Zhengzhou, Henan, China
- Zhengzhou University Industrial Technology Research Institute, Zhengzhou, Henan, China
| |
Collapse
|
9
|
Ma Y, Huang Z, Yang Y, Zhang S, Dong Q, Wang R, Hu L. Emotion Recognition Model of EEG Signals Based on Double Attention Mechanism. Brain Sci 2024; 14:1289. [PMID: 39766488 PMCID: PMC11674476 DOI: 10.3390/brainsci14121289] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2024] [Revised: 12/14/2024] [Accepted: 12/19/2024] [Indexed: 01/11/2025] Open
Abstract
BACKGROUND Emotions play a crucial role in people's lives, profoundly affecting their cognition, decision-making, and interpersonal communication. Emotion recognition based on brain signals has become a significant challenge in the fields of affective computing and human-computer interaction. METHODS Addressing the issue of inaccurate feature extraction and low accuracy of existing deep learning models in emotion recognition, this paper proposes a multi-channel automatic classification model for emotion EEG signals named DACB, which is based on dual attention mechanisms, convolutional neural networks, and bidirectional long short-term memory networks. DACB extracts features in both temporal and spatial dimensions, incorporating not only convolutional neural networks but also SE attention mechanism modules for learning the importance of different channel features, thereby enhancing the network's performance. DACB also introduces dot product attention mechanisms to learn the importance of spatial and temporal features, effectively improving the model's accuracy. RESULTS The accuracy of this method in single-shot validation tests on the SEED-IV and DREAMER (Valence-Arousal-Dominance three-classification) datasets is 99.96% and 87.52%, 90.06%, and 89.05%, respectively. In 10-fold cross-validation tests, the accuracy is 99.73% and 84.26%, 85.40%, and 85.02%, outperforming other models. CONCLUSIONS This demonstrates that the DACB model achieves high accuracy in emotion classification tasks, demonstrating outstanding performance and generalization ability and providing new directions for future research in EEG signal recognition.
Collapse
Affiliation(s)
- Yahong Ma
- Xi’an Key Laboratory of High Pricision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China; (Y.Y.); (S.Z.); (R.W.)
| | - Zhentao Huang
- Affiliation College of Computer Science and Technology, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
| | - Yuyao Yang
- Xi’an Key Laboratory of High Pricision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China; (Y.Y.); (S.Z.); (R.W.)
| | - Shanwen Zhang
- Xi’an Key Laboratory of High Pricision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China; (Y.Y.); (S.Z.); (R.W.)
| | - Qi Dong
- School of Mathematics and Statistics, Zhengzhou University, Zhengzhou 710003, China;
| | - Rongrong Wang
- Xi’an Key Laboratory of High Pricision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China; (Y.Y.); (S.Z.); (R.W.)
| | - Liangliang Hu
- West China Institute of Children’s Brain and Cognition, Chongqing University of Education, Chongqing 400065, China;
| |
Collapse
|
10
|
Hao S, Ruiying X, Lifei X, Jian W, Jiaxin J, Siping F, Xiaoqin W, Xin Q, Lu L, Yufeng Z. The effect of workload on mind-wandering of drilling operators measured by electroencephalography (EEG). Exp Brain Res 2024; 243:29. [PMID: 39704838 DOI: 10.1007/s00221-024-06974-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2024] [Accepted: 12/04/2024] [Indexed: 12/21/2024]
Abstract
Mind wandering can cause workers to overlook safety hazards and delay making accurate operational decisions, ultimately raising the potential for accidents. However, there is relatively little research on the physiological characteristics of drilling workers during mind wandering. The aim of this investigation was to tackle the constraints of previous studies and to establish a more comprehensive theoretical framework and practical guidance for safety management. To this end, the phenomenon of workload on mind wandering among drillers during the drilling process was investigated in depth. It focused on drilling site workers, using SART paradigm tasks and EEG devices to track cognitive states under various loads, exploring how they affect mind wandering and EEG mechanisms. Fifty workers participated, observing drilling images to judge accidents. Results showed workload influenced cognitive processes such as mind wandering occurrence, reaction time, accuracy, and brain connectivity. High workload increased reaction time, decreased accuracy, raised mind wandering frequency, altered theta, beta, and gamma waves, and reduced cerebral synchronisation and engagement. Workload affected employees' mind wandering, sensations, focus, and work status, with a positive correlation between workload and mind wandering, potentially harming work performance and safety. Analyzing EEG data helps identify mind wandering and develop intervention measures. In depth research on these features not only helps identify employee mind wandering, but also promotes the development of more precise and personalized intervention measures.
Collapse
Affiliation(s)
- Su Hao
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China.
- Key Laboratory of Energy Security and Low Carbon Development, Chengdu, 610500, China.
| | - Xie Ruiying
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China.
| | - Xu Lifei
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China
| | - Wang Jian
- Chuanqing Drilling Engineering Company Chuanxi Drilling Company Limited of China National Petroleum Corporation, Chengdu, 610051, China
| | - Jiang Jiaxin
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China
| | - Fan Siping
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China
| | - Wang Xiaoqin
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China
| | - Qing Xin
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China
| | - Liu Lu
- School of Economics and Management, Southwest Petroleum University, Chengdu, 610500, China
| | - Zhang Yufeng
- Department of Rehabilitation Medicine, Xindu District People's Hospital, Chengdu, 610500, China
| |
Collapse
|
11
|
Dong Y, Jing C, Mahmud M, Ng MKP, Wang S. Enhancing cross-subject emotion recognition precision through unimodal EEG: a novel emotion preceptor model. Brain Inform 2024; 11:31. [PMID: 39692977 DOI: 10.1186/s40708-024-00245-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2024] [Accepted: 11/26/2024] [Indexed: 12/19/2024] Open
Abstract
Affective computing is a key research area in computer science, neuroscience, and psychology, aimed at enabling computers to recognize, understand, and respond to human emotional states. As the demand for affective computing technology grows, emotion recognition methods based on physiological signals have become research hotspots. Among these, electroencephalogram (EEG) signals, which reflect brain activity, are highly promising. However, due to individual physiological and anatomical differences, EEG signals introduce noise, reducing emotion recognition performance. Additionally, the synchronous collection of multimodal data in practical applications requires high equipment and environmental standards, limiting the practical use of EEG signals. To address these issues, this study proposes the Emotion Preceptor, a cross-subject emotion recognition model based on unimodal EEG signals. This model introduces a Static Spatial Adapter to integrate spatial information in EEG signals, reducing individual differences and extracting robust encoding information. The Temporal Causal Network then leverages temporal information to extract beneficial features for emotion recognition, achieving precise recognition based on unimodal EEG signals. Extensive experiments on the SEED and SEED-V datasets demonstrate the superior performance of the Emotion Preceptor and validate the effectiveness of the new data processing method that combines DE features in a temporal sequence. Additionally, we analyzed the model's data flow and encoding methods from a biological interpretability perspective and validated it with neuroscience research related to emotion generation and regulation, promoting further development in emotion recognition research based on EEG signals.
Collapse
Affiliation(s)
- Yihang Dong
- Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- University of Chinese Academy of Sciences, Chinese Academy of Sciences, Beijing, China
| | - Changhong Jing
- Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Mufti Mahmud
- Information and Computer Science Department, King Fahd University of Petroleum and Minerals, Dhahran, Saudi Arabia
| | - Michael Kwok-Po Ng
- Department of Mathematics, Hong Kong Baptist University, Hong Kong, China
| | - Shuqiang Wang
- Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.
- University of Chinese Academy of Sciences, Chinese Academy of Sciences, Beijing, China.
| |
Collapse
|
12
|
Lu W, Zhang X, Xia L, Ma H, Tan TP. Domain adaptation spatial feature perception neural network for cross-subject EEG emotion recognition. Front Hum Neurosci 2024; 18:1471634. [PMID: 39741785 PMCID: PMC11685119 DOI: 10.3389/fnhum.2024.1471634] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2024] [Accepted: 11/04/2024] [Indexed: 01/03/2025] Open
Abstract
Emotion recognition is a critical research topic within affective computing, with potential applications across various domains. Currently, EEG-based emotion recognition, utilizing deep learning frameworks, has been effectively applied and achieved commendable performance. However, existing deep learning-based models face challenges in capturing both the spatial activity features and spatial topology features of EEG signals simultaneously. To address this challenge, a domain-adaptation spatial-feature perception-network has been proposed for cross-subject EEG emotion recognition tasks, named DSP-EmotionNet. Firstly, a spatial activity topological feature extractor module has been designed to capture spatial activity features and spatial topology features of EEG signals, named SATFEM. Then, using SATFEM as the feature extractor, DSP-EmotionNet has been designed, significantly improving the accuracy of the model in cross-subject EEG emotion recognition tasks. The proposed model surpasses state-of-the-art methods in cross-subject EEG emotion recognition tasks, achieving an average recognition accuracy of 82.5% on the SEED dataset and 65.9% on the SEED-IV dataset.
Collapse
Affiliation(s)
- Wei Lu
- Henan High-speed Railway Operation and Maintenance Engineering Research Center, Zhengzhou Railway Vocational and Technical College, Zhengzhou, China
- School of Computer Sciences, Universiti Sains Malaysia, Penang, Malaysia
| | - Xiaobo Zhang
- School of Computer Sciences, Universiti Sains Malaysia, Penang, Malaysia
- Jiangxi Vocational College of Finance and Economics, Jiujiang, China
| | - Lingnan Xia
- Henan High-speed Railway Operation and Maintenance Engineering Research Center, Zhengzhou Railway Vocational and Technical College, Zhengzhou, China
| | - Hua Ma
- Henan High-speed Railway Operation and Maintenance Engineering Research Center, Zhengzhou Railway Vocational and Technical College, Zhengzhou, China
| | - Tien-Ping Tan
- School of Computer Sciences, Universiti Sains Malaysia, Penang, Malaysia
| |
Collapse
|
13
|
Hu F, He K, Qian M, Liu X, Qiao Z, Zhang L, Xiong J. STAFNet: an adaptive multi-feature learning network via spatiotemporal fusion for EEG-based emotion recognition. Front Neurosci 2024; 18:1519970. [PMID: 39720230 PMCID: PMC11666491 DOI: 10.3389/fnins.2024.1519970] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2024] [Accepted: 11/18/2024] [Indexed: 12/26/2024] Open
Abstract
Introduction Emotion recognition using electroencephalography (EEG) is a key aspect of brain-computer interface research. Achieving precision requires effectively extracting and integrating both spatial and temporal features. However, many studies focus on a single dimension, neglecting the interplay and complementarity of multi-feature information, and the importance of fully integrating spatial and temporal dynamics to enhance performance. Methods We propose the Spatiotemporal Adaptive Fusion Network (STAFNet), a novel framework combining adaptive graph convolution and temporal transformers to enhance the accuracy and robustness of EEG-based emotion recognition. The model includes an adaptive graph convolutional module to capture brain connectivity patterns through spatial dynamic evolution and a multi-structured transformer fusion module to integrate latent correlations between spatial and temporal features for emotion classification. Results Extensive experiments were conducted on the SEED and SEED-IV datasets to evaluate the performance of STAFNet. The model achieved accuracies of 97.89% and 93.64%, respectively, outperforming state-of-the-art methods. Interpretability analyses, including confusion matrices and t-SNE visualizations, were employed to examine the influence of different emotions on the model's recognition performance. Furthermore, an investigation of varying GCN layer depths demonstrated that STAFNet effectively mitigates the over-smoothing issue in deeper GCN architectures. Discussion In summary, the findings validate the effectiveness of STAFNet in EEG-based emotion recognition. The results emphasize the critical role of spatiotemporal feature extraction and introduce an innovative framework for feature fusion, advancing the state of the art in emotion recognition.
Collapse
Affiliation(s)
- Fo Hu
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, China
| | - Kailun He
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, China
| | - Mengyuan Qian
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, China
| | - Xiaofeng Liu
- College of Information Engineering, Zhejiang University of Technology, Hangzhou, China
| | - Zukang Qiao
- Department of Tuina, The First Affiliated Hospital of Zhejiang Chinese Medical University (Zhejiang Provincial Hospital of Chinese Medicine), Hangzhou, China
| | - Lekai Zhang
- The School of Design and Architecture, Zhejiang University of Technology, Hangzhou, China
| | - Junlong Xiong
- Department of Tuina, The First Affiliated Hospital of Zhejiang Chinese Medical University (Zhejiang Provincial Hospital of Chinese Medicine), Hangzhou, China
| |
Collapse
|
14
|
Zhuang W, Zhang Y, Wang Y, He K. 3D-BCLAM: A Lightweight Neurodynamic Model for Assessing Student Learning Effectiveness. SENSORS (BASEL, SWITZERLAND) 2024; 24:7856. [PMID: 39686393 DOI: 10.3390/s24237856] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Revised: 12/04/2024] [Accepted: 12/06/2024] [Indexed: 12/18/2024]
Abstract
Evaluating students' learning effectiveness is of great importance for gaining a deeper understanding of the learning process, accurately diagnosing learning barriers, and developing effective teaching strategies. Emotion, as a key factor influencing learning outcomes, provides a novel perspective for identifying cognitive states and emotional experiences. However, traditional evaluation methods suffer from one sidedness in feature extraction and high complexity in model construction, often making it difficult to fully explore the deep value of emotional data. To address this challenge, we have innovatively proposed a lightweight neurodynamic model: 3D-BCLAM. This model cleverly integrates Bidirectional Convolutional Long Short-Term Memory (BCL) and dynamic attention mechanism, in order to efficiently capture emotional dynamic changes in time series with extremely low computational cost. 3D-BCLAM can achieve a comprehensive evaluation of students' learning outcomes, covering not only the cognitive level but also delving into the emotional dimension for detailed analysis. Under testing on public datasets, 3D-BCLAM has demonstrated outstanding performance, significantly outperforming traditional machine learning and deep learning models based on Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN). This achievement not only validates the effectiveness of the 3D-BCLAM model, but also provides strong support for promoting the innovation of student learning effectiveness assessment.
Collapse
Affiliation(s)
- Wei Zhuang
- School of Computer Science, Nanjing University of Information Science and Technology, Nanjing 210044, China
| | - Yunhong Zhang
- School of Computer Science, Nanjing University of Information Science and Technology, Nanjing 210044, China
| | - Yuan Wang
- School of Teacher and Education, Nanjing University of Information Science and Technology, Nanjing 210044, China
| | - Kaiyang He
- School of Mathematics and Physics, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China
| |
Collapse
|
15
|
Qiu L, Zhong L, Li J, Feng W, Zhou C, Pan J. SFT-SGAT: A semi-supervised fine-tuning self-supervised graph attention network for emotion recognition and consciousness detection. Neural Netw 2024; 180:106643. [PMID: 39186838 DOI: 10.1016/j.neunet.2024.106643] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Revised: 04/11/2024] [Accepted: 08/14/2024] [Indexed: 08/28/2024]
Abstract
Emotional recognition is highly important in the field of brain-computer interfaces (BCIs). However, due to the individual variability in electroencephalogram (EEG) signals and the challenges in obtaining accurate emotional labels, traditional methods have shown poor performance in cross-subject emotion recognition. In this study, we propose a cross-subject EEG emotion recognition method based on a semi-supervised fine-tuning self-supervised graph attention network (SFT-SGAT). First, we model multi-channel EEG signals by constructing a graph structure that dynamically captures the spatiotemporal topological features of EEG signals. Second, we employ a self-supervised graph attention neural network to facilitate model training, mitigating the impact of signal noise on the model. Finally, a semi-supervised approach is used to fine-tune the model, enhancing its generalization ability in cross-subject classification. By combining supervised and unsupervised learning techniques, the SFT-SGAT maximizes the utility of limited labeled data in EEG emotion recognition tasks, thereby enhancing the model's performance. Experiments based on leave-one-subject-out cross-validation demonstrate that SFT-SGAT achieves state-of-the-art cross-subject emotion recognition performance on the SEED and SEED-IV datasets, with accuracies of 92.04% and 82.76%, respectively. Furthermore, experiments conducted on a self-collected dataset comprising ten healthy subjects and eight patients with disorders of consciousness (DOCs) revealed that the SFT-SGAT attains high classification performance in healthy subjects (maximum accuracy of 95.84%) and was successfully applied to DOC patients, with four patients achieving emotion recognition accuracies exceeding 60%. The experiments demonstrate the effectiveness of the proposed SFT-SGAT model in cross-subject EEG emotion recognition and its potential for assessing levels of consciousness in patients with DOC.
Collapse
Affiliation(s)
- Lina Qiu
- School of Artificial Intelligence, South China Normal University, Guangzhou, 510630, China; Research Station in Mathematics, South China Normal University, Guangzhou, 510630, China.
| | - Liangquan Zhong
- School of Artificial Intelligence, South China Normal University, Guangzhou, 510630, China.
| | - Jianping Li
- School of Artificial Intelligence, South China Normal University, Guangzhou, 510630, China.
| | - Weisen Feng
- School of Artificial Intelligence, South China Normal University, Guangzhou, 510630, China.
| | - Chengju Zhou
- School of Artificial Intelligence, South China Normal University, Guangzhou, 510630, China.
| | - Jiahui Pan
- School of Artificial Intelligence, South China Normal University, Guangzhou, 510630, China.
| |
Collapse
|
16
|
Hou M, Zhang X, Chen G, Huang L, Sun Y. Emotion Recognition Based on a EEG-fNIRS Hybrid Brain Network in the Source Space. Brain Sci 2024; 14:1166. [PMID: 39766365 PMCID: PMC11674611 DOI: 10.3390/brainsci14121166] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Revised: 11/07/2024] [Accepted: 11/19/2024] [Indexed: 01/11/2025] Open
Abstract
Background/Objectives: Studies have shown that emotion recognition based on electroencephalogram (EEG) and functional near-infrared spectroscopy (fNIRS) multimodal physiological signals exhibits superior performance compared to that of unimodal approaches. Nonetheless, there remains a paucity of in-depth investigations analyzing the inherent relationship between EEG and fNIRS and constructing brain networks to improve the performance of emotion recognition. Methods: In this study, we introduce an innovative method to construct hybrid brain networks in the source space based on simultaneous EEG-fNIRS signals for emotion recognition. Specifically, we perform source localization on EEG signals to derive the EEG source signals. Subsequently, causal brain networks are established in the source space by analyzing the Granger causality between the EEG source signals, while coupled brain networks in the source space are formed by assessing the coupling strength between the EEG source signals and the fNIRS signals. The resultant causal brain networks and coupled brain networks are integrated to create hybrid brain networks in the source space, which serve as features for emotion recognition. Results: The effectiveness of our proposed method is validated on multiple emotion datasets. The experimental results indicate that the recognition performance of our approach significantly surpasses that of the baseline method. Conclusions: This work offers a novel perspective on the fusion of EEG and fNIRS signals in an emotion-evoked experimental paradigm and provides a feasible solution for enhancing emotion recognition performance.
Collapse
Affiliation(s)
- Mingxing Hou
- College of Integrated Circuits, Taiyuan University of Technology, Taiyuan 030600, China;
- College of Computer Science and Technology, Taiyuan Normal University, Taiyuan 030619, China
| | - Xueying Zhang
- College of Electronic Information Engineering, Taiyuan University of Technology, Taiyuan 030600, China; (G.C.); (L.H.); (Y.S.)
| | - Guijun Chen
- College of Electronic Information Engineering, Taiyuan University of Technology, Taiyuan 030600, China; (G.C.); (L.H.); (Y.S.)
| | - Lixia Huang
- College of Electronic Information Engineering, Taiyuan University of Technology, Taiyuan 030600, China; (G.C.); (L.H.); (Y.S.)
| | - Ying Sun
- College of Electronic Information Engineering, Taiyuan University of Technology, Taiyuan 030600, China; (G.C.); (L.H.); (Y.S.)
| |
Collapse
|
17
|
Zhao X, Xu S, Geng K, Zhou T, Xu T, Wang Z, Feng S, Hu H. MP: A steady-state visual evoked potential dataset based on multiple paradigms. iScience 2024; 27:111030. [PMID: 39759080 PMCID: PMC11700636 DOI: 10.1016/j.isci.2024.111030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2024] [Revised: 08/27/2024] [Accepted: 09/10/2024] [Indexed: 01/07/2025] Open
Abstract
In the field of steady-state visual evoked potential (SSVEP), stimulus paradigms are regularly arranged or mimic the style of a keyboard with the same size. However, stimulation paradigms have important effects on the performance of SSVEP systems, which correlate with the electroencephalogram (EEG) signal amplitude and recognition accuracy. This paper provides MP dataset that was acquired using a 12-target BCI speller. MP dataset contains 9-channel EEG signals from the occipital region of 24 subjects under 5 stimulation paradigms with different stimulus sizes and arrangements. Stimuli were encoded using joint frequency and phase modulation (JFPM) method. Subjects completed an offline prompted spelling task using a speller under 5 paradigms. Each experiment contains 8 blocks, and each block contains 12 trials. Designers can use this dataset to test the performance of algorithms considering "stimulus size" and "stimulus arrangement". EEG data showed SSVEP features through amplitude-frequency analysis. FBCCA and TRCA confirmed its suitability.
Collapse
Affiliation(s)
- Xi Zhao
- School of Microelectronics, Shanghai University, Shanghai 200444, China
| | - Shencheng Xu
- School of Microelectronics, Shanghai University, Shanghai 200444, China
| | - Kexing Geng
- School of Microelectronics, Shanghai University, Shanghai 200444, China
| | - Ting Zhou
- School of Microelectronics, Shanghai University, Shanghai 200444, China
| | - Tianheng Xu
- Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China
| | - Zhenyu Wang
- Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China
| | - Shilun Feng
- State Key Laboratory of Transducer Technology, Shanghai Institute of Microsystem and Information Technology, Chinese Academy of Sciences, Shanghai 200050, China
| | - Honglin Hu
- Shanghai Advanced Research Institute, Chinese Academy of Sciences, Shanghai 201210, China
- School of Information Science and Technology, Shanghai Tech University, Shanghai 201210, China
| |
Collapse
|
18
|
Mi P, Yan L, Cheng Y, Liu Y, Wang J, Shoukat MU, Yan F, Qin G, Han P, Zhai Y. Driver Cognitive Architecture Based on EEG Signals: A Review. IEEE SENSORS JOURNAL 2024; 24:36261-36286. [DOI: 10.1109/jsen.2024.3471699] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/12/2025]
Affiliation(s)
- Peiwen Mi
- School of Automotive Engineering, Wuhan University of Technology, Wuhan, China
| | - Lirong Yan
- School of Automotive Engineering, Wuhan University of Technology, Wuhan, China
| | - Yu Cheng
- School of Automotive Engineering, Wuhan University of Technology, Wuhan, China
| | - Yan Liu
- School of Automotive Engineering, Wuhan University of Technology, Wuhan, China
| | - Jun Wang
- School of Automotive Engineering, Wuhan University of Technology, Wuhan, China
| | | | - Fuwu Yan
- School of Automotive Engineering, Wuhan University of Technology, Wuhan, China
| | - Guofeng Qin
- Teachers College for Vocational and Education, Guangxi Normal University, Guilin, China
| | - Peng Han
- Xiangyang DAAN Automobile Test Center Corporation Ltd., Xiangyang, China
| | - Yikang Zhai
- Xiangyang DAAN Automobile Test Center Corporation Ltd., Xiangyang, China
| |
Collapse
|
19
|
Thiruselvam SV, Reddy MR. Frontal EEG correlation based human emotion identification and classification. Phys Eng Sci Med 2024:10.1007/s13246-024-01495-w. [PMID: 39543049 DOI: 10.1007/s13246-024-01495-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2024] [Accepted: 10/28/2024] [Indexed: 11/17/2024]
Abstract
Humans express their feelings and intentions of their actions or communication through emotions. Recent advancements in technology involve machines in human communication in day-to-day life. Thus, understanding of human emotions by machines will be very helpful in assisting the user in a far better way. Various physiological and non-physiological signals can be used to make the machines to recognize the emotion of a person. The identification of emotional content in the signals is crucial to understand emotion and the machines act with emotional intelligence at appropriate times, thus providing a better human machine interaction with emotion identification system and mental health monitoring for psychiatric patients. This work includes the creation of an emotion EEG dataset, the development of an algorithm for identifying the emotion elicitation segments in the EEG signal, and the classification of emotions from EEG signals. The EEG signals are divided into 3s segments, and the segments with emotional content are selected based on the decrease in correlation between the frontal electrodes. The selected segments are validated with the facial expressions of the subjects in the appropriate time segments of the face video. EEGNet is used to classify the emotion from the EEG signal. The classification accuracy with the selected emotional EEG segments is higher compared to the accuracy using all the EEG segments. In subject-specific classification, an average accuracy of 80.87% is obtained from the network trained with selected EEG segments, and 70.5% is obtained from training with all EEG segments. In subject-independent classification, the accuracy of classification is 67% and 63.8% with and without segment selection, respectively. The proposed method of selection of EEG segments is validated using the DEAP dataset, and classification accuracies and F1-scores of subject dependent and subject-independent methods are presented.
Collapse
Affiliation(s)
- S V Thiruselvam
- Department of Applied Mechanics and Biomedical Engineering, Indian Institute of Technology Madras, Chennai, India.
| | - M Ramasubba Reddy
- Department of Applied Mechanics and Biomedical Engineering, Indian Institute of Technology Madras, Chennai, India
| |
Collapse
|
20
|
Alyan E, Arnau S, Reiser JE, Wascher E. Synchronization-based fusion of EEG and eye blink signals for enhanced decoding accuracy. Sci Rep 2024; 14:26918. [PMID: 39506076 PMCID: PMC11541762 DOI: 10.1038/s41598-024-78542-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 10/31/2024] [Indexed: 11/08/2024] Open
Abstract
Decoding locomotor tasks is crucial in cognitive neuroscience for understanding brain responses to physical tasks. Traditional methods like EEG offer brain activity insights but may require additional modalities for enhanced interpretative precision and depth. The integration of EEG with ocular metrics, particularly eye blinks, presents a promising avenue for understanding cognitive processes by combining neural and ocular behaviors. However, synchronizing EEG and eye blink activities poses a significant challenge due to their frequently inconsistent alignment. Our study with 35 participants performing various locomotor tasks such as standing, walking, and transversing obstacles introduced a novel methodology, pcEEG+, which fuses EEG principal components (pcEEG) with aligned eye blink data (syncBlink). The results demonstrated that pcEEG+ significantly improved decoding accuracy in locomotor tasks, reaching 78% in some conditions, and surpassed standalone pcEEG and syncBlink methods by 7.6% and 22.7%, respectively. The temporal generalization matrix confirmed the consistency of pcEEG+ across tasks and times. The results were replicated using two driving simulator datasets, thereby confirming the validity of our method. This study demonstrates the efficacy of the pcEEG+ method in decoding locomotor tasks, underscoring the importance of temporal synchronization for accuracy and offering a deeper insight into brain activity during complex movements.
Collapse
Affiliation(s)
- Emad Alyan
- Department of Ergonomics, Leibniz Research Centre for Working Environment and Human Factors, 44139, Dortmund, Germany.
| | - Stefan Arnau
- Department of Ergonomics, Leibniz Research Centre for Working Environment and Human Factors, 44139, Dortmund, Germany
| | - Julian Elias Reiser
- Department of Ergonomics, Leibniz Research Centre for Working Environment and Human Factors, 44139, Dortmund, Germany
| | - Edmund Wascher
- Department of Ergonomics, Leibniz Research Centre for Working Environment and Human Factors, 44139, Dortmund, Germany
| |
Collapse
|
21
|
Dan Y, Zhou D, Wang Z. Discriminative possibilistic clustering promoting cross-domain emotion recognition. Front Neurosci 2024; 18:1458815. [PMID: 39554850 PMCID: PMC11565435 DOI: 10.3389/fnins.2024.1458815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/03/2024] [Accepted: 10/03/2024] [Indexed: 11/19/2024] Open
Abstract
The affective Brain-Computer Interface (aBCI) systems strive to enhance prediction accuracy for individual subjects by leveraging data from multiple subjects. However, significant differences in EEG (Electroencephalogram) feature patterns among subjects often hinder these systems from achieving the desired outcomes. Although studies have attempted to address this challenge using subject-specific classifier strategies, the scarcity of labeled data remains a major hurdle. In light of this, Domain Adaptation (DA) technology has gradually emerged as a prominent approach in the field of EEG-based emotion recognition, attracting widespread research interest. The crux of DA learning lies in resolving the issue of distribution mismatch between training and testing datasets, which has become a focal point of academic attention. Currently, mainstream DA methods primarily focus on mitigating domain distribution discrepancies by minimizing the Maximum Mean Discrepancy (MMD) or its variants. Nevertheless, the presence of noisy samples in datasets can lead to pronounced shifts in domain means, thereby impairing the adaptive performance of DA methods based on MMD and its variants in practical applications to some extent. Research has revealed that the traditional MMD metric can be transformed into a 1-center clustering problem, and the possibility clustering model is adept at mitigating noise interference during the data clustering process. Consequently, the conventional MMD metric can be further relaxed into a possibilistic clustering model. Therefore, we construct a distributed distance measure with Discriminative Possibilistic Clustering criterion (DPC), which aims to achieve two objectives: (1) ensuring the discriminative effectiveness of domain distribution alignment by finding a shared subspace that minimizes the overall distribution distance between domains while maximizing the semantic distribution distance according to the principle of "sames attract and opposites repel"; and (2) enhancing the robustness of distribution distance measure by introducing a fuzzy entropy regularization term. Theoretical analysis confirms that the proposed DPC is an upper bound of the existing MMD metric under certain conditions. Therefore, the MMD objective can be effectively optimized by minimizing the DPC. Finally, we propose a domain adaptation in Emotion recognition based on DPC (EDPC) that introduces a graph Laplacian matrix to preserve the geometric structural consistency between data within the source and target domains, thereby enhancing label propagation performance. Simultaneously, by maximizing the use of source domain discriminative information to minimize domain discrimination errors, the generalization performance of the DA model is further improved. Comparative experiments on several representative domain adaptation learning methods using multiple EEG datasets (i.e., SEED and SEED-IV) show that, in most cases, the proposed method exhibits better or comparable consistent generalization performance.
Collapse
Affiliation(s)
- Yufang Dan
- Ningbo Polytechnic, Institute of Artificial Intelligence Application, Zhejiang, China
| | - Di Zhou
- Industrial Technological Institute of Intelligent Manufacturing, Sichuan University of Arts and Science, Sichuang, China
| | | |
Collapse
|
22
|
Zhang D, Li H, Xie J. Unsupervised and semi-supervised domain adaptation networks considering both global knowledge and prototype-based local class information for Motor Imagery Classification. Neural Netw 2024; 179:106497. [PMID: 38986186 DOI: 10.1016/j.neunet.2024.106497] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2023] [Revised: 05/21/2024] [Accepted: 06/25/2024] [Indexed: 07/12/2024]
Abstract
The non-stationarity of EEG signals results in variability across sessions, impeding model building and data sharing. In this paper, we propose a domain adaptation method called GPL, which simultaneously considers global knowledge and prototype-based local class information to enhance the classification accuracy of motor imagery signals. Depending on the amount of labeled data available in the target domain, the method is implemented in both unsupervised and semi-supervised versions. Specifically, at the global level, we employ the maximum mean difference (MMD) loss to globally constrain the feature space, achieving comprehensive alignment. In the context of class-level operations, we propose two memory banks designed to accommodate class prototypes in each domain and constrain feature embeddings by applying two prototype-based contrastive losses. The source contrastive loss is used to organize source features spatially based on categories, thereby reconciling inter-class and intra-class relationships, while the interactive contrastive loss is employed to facilitate cross-domain information interaction. Simultaneously, in unsupervised scenarios, to mitigate the adverse effects of excessive pseudo-labels, we introduce an entropy-aware strategy that dynamically evaluates the confidence level of target data and personalized constraints on the participation of interactive contrastive loss. To validate our approach, extensive experiments were conducted on a highly regarded public EEG dataset, namely Dataset IIa of the BCI Competition IV, as well as a large-scale EEG dataset called GigaDB. The experiments yielded average classification accuracies of 86.03% and 84.22% respectively. These results demonstrate that our method is an effective EEG decoding model, conducive to advancing the development of motor imagery brain-computer interfaces. The architecture proposed in this study and the code for data partitioning can be found at https://github.com/zhangdx21/GPL.
Collapse
Affiliation(s)
- Dongxue Zhang
- Jilin University, College of Computer Science and Technology, Changchun, Jilin Province, China; Key Laboratory of Symbol Computation and Knowledge Engineering, Jilin University, Changchun 130012, China
| | - Huiying Li
- Jilin University, College of Computer Science and Technology, Changchun, Jilin Province, China; Key Laboratory of Symbol Computation and Knowledge Engineering, Jilin University, Changchun 130012, China.
| | - Jingmeng Xie
- Xi'an Jiaotong University, College of Electronic information, Xi'an, Shanxi Province, China
| |
Collapse
|
23
|
Cheng C, Liu W, Feng L, Jia Z. Emotion recognition using hierarchical spatial-temporal learning transformer from regional to global brain. Neural Netw 2024; 179:106624. [PMID: 39163821 DOI: 10.1016/j.neunet.2024.106624] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2024] [Revised: 06/10/2024] [Accepted: 08/09/2024] [Indexed: 08/22/2024]
Abstract
Emotion recognition is an essential but challenging task in human-computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial-temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial-temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial-temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the long-range spatial-temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.
Collapse
Affiliation(s)
- Cheng Cheng
- Department of Computer Science and Technology, Dalian University of Technology, Dalian, China
| | - Wenzhe Liu
- School of Information Engineering, Huzhou University, Huzhou, China
| | - Lin Feng
- Department of Computer Science and Technology, Dalian University of Technology, Dalian, China; School of Information and Communication Engineering, Dalian Minzu University, Dlian, China.
| | - Ziyu Jia
- Brainnetome Center, Institute of Automation, Chinese Academy of Sciences, University of Chinese Academy of Sciences, Beijing, China.
| |
Collapse
|
24
|
Chen S, Wang Y, Lin X, Sun X, Li W, Ma W. Cross-subject emotion recognition in brain-computer interface based on frequency band attention graph convolutional adversarial neural networks. J Neurosci Methods 2024; 411:110276. [PMID: 39237038 DOI: 10.1016/j.jneumeth.2024.110276] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Revised: 08/19/2024] [Accepted: 09/01/2024] [Indexed: 09/07/2024]
Abstract
BACKGROUND Emotion is an important area in neuroscience. Cross-subject emotion recognition based on electroencephalogram (EEG) data is challenging due to physiological differences between subjects. Domain gap, which refers to the different distributions of EEG data at different subjects, has attracted great attention for cross-subject emotion recognition. COMPARISON WITH EXISTING METHODS This study focuses on narrowing the domain gap between subjects through the emotional frequency bands and the relationship information between EEG channels. Emotional frequency band features represent the energy distribution of EEG data in different frequency ranges, while relationship information between EEG channels provides spatial distribution information about EEG data. NEW METHOD To achieve this, this paper proposes a model called the Frequency Band Attention Graph convolutional Adversarial neural Network (FBAGAN). This model includes three components: a feature extractor, a classifier, and a discriminator. The feature extractor consists of a layer with a frequency band attention mechanism and a graph convolutional neural network. The mechanism effectively extracts frequency band information by assigning weights and Graph Convolutional Networks can extract relationship information between EEG channels by modeling the graph structure. The discriminator then helps minimize the gap in the frequency information and relationship information between the source and target domains, improving the model's ability to generalize. RESULTS The FBAGAN model is extensively tested on the SEED, SEED-IV, and DEAP datasets. The accuracy and standard deviation scores are 88.17% and 4.88, respectively, on the SEED dataset, and 77.35% and 3.72 on the SEED-IV dataset. On the DEAP dataset, the model achieves 69.64% for Arousal and 65.18% for Valence. These results outperform most existing models. CONCLUSIONS The experiments indicate that FBAGAN effectively addresses the challenges of transferring EEG channel domain and frequency band domain, leading to improved performance.
Collapse
Affiliation(s)
- Shinan Chen
- School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, PR China
| | - Yuchen Wang
- School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, PR China.
| | - Xuefen Lin
- School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, PR China
| | - Xiaoyong Sun
- School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, PR China
| | - Weihua Li
- School of Engineering, Computer & Mathematical Sciences, Auckland University of Technology, Auckland 1010, New Zealand
| | - Weifeng Ma
- School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, PR China
| |
Collapse
|
25
|
Wu X, Ju X, Dai S, Li X, Li M. Multi-source domain adaptation for EEG emotion recognition based on inter-domain sample hybridization. Front Hum Neurosci 2024; 18:1464431. [PMID: 39545146 PMCID: PMC11560783 DOI: 10.3389/fnhum.2024.1464431] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2024] [Accepted: 10/15/2024] [Indexed: 11/17/2024] Open
Abstract
Background Electroencephalogram (EEG) is widely used in emotion recognition due to its precision and reliability. However, the nonstationarity of EEG signals causes significant differences between individuals or sessions, making it challenging to construct a robust model. Recently, domain adaptation (DA) methods have shown excellent results in cross-subject EEG emotion recognition by aligning marginal distributions. Nevertheless, these methods do not consider emotion category labels, which can lead to label confusion during alignment. Our study aims to alleviate this problem by promoting conditional distribution alignment during domain adaptation to improve cross-subject and cross-session emotion recognition performance. Method This study introduces a multi-source domain adaptation common-branch network for EEG emotion recognition and proposes a novel sample hybridization method. This method enables the introduction of target domain data information by directionally hybridizing source and target domain samples without increasing the overall sample size, thereby enhancing the effectiveness of conditional distribution alignment in domain adaptation. Cross-subject and cross-session experiments were conducted on two publicly available datasets, SEED and SEED-IV, to validate the proposed model. Result In cross-subject emotion recognition, our method achieved an average accuracy of 90.27% on the SEED dataset, with eight out of 15 subjects attaining a recognition accuracy higher than 90%. For the SEED-IV dataset, the recognition accuracy also reached 73.21%. Additionally, in the cross-session experiment, we sequentially used two out of the three session data as source domains and the remaining session as the target domain for emotion recognition. The proposed model yielded average accuracies of 94.16 and 75.05% on the two datasets, respectively. Conclusion Our proposed method aims to alleviate the difficulties of emotion recognition from the limited generalization ability of EEG features across subjects and sessions. Though adapting the multi-source domain adaptation and the sample hybridization method, the proposed method can effectively transfer the emotion-related knowledge of known subjects and achieve accurate emotion recognition on unlabeled subjects.
Collapse
Affiliation(s)
| | | | | | | | - Ming Li
- College of Intelligence Science and Technology, National University of Defense Technology, Changsha, China
| |
Collapse
|
26
|
Zhu L, Xu M, Huang A, Zhang J, Tan X. Multi-source domain transfer network based on subdomain adaptation and minimum class confusion for EEG emotion recognition. Comput Methods Biomech Biomed Engin 2024:1-13. [PMID: 39429223 DOI: 10.1080/10255842.2024.2417212] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2024] [Revised: 09/19/2024] [Accepted: 10/08/2024] [Indexed: 10/22/2024]
Abstract
Electroencephalogram (EEG) signals, which objectively reflect the state of the brain, are widely favored in emotion recognition research. However, the presence of cross-session and cross-subject variation in EEG signals has hindered the practical implementation of EEG-based emotion recognition technologies. In this article, we propose a multi-source domain transfer method based on subdomain adaptation and minimum class confusion (MS-SAMCC) in response to the addressed issue. First, we introduce the mix-up data augmentation technique to generate augmented samples. Next, we propose a minimum class confusion subdomain adaptation method (MCCSA) as a sub-module of the multi-source domain adaptation module. This approach enables global alignment between each source domain and the target domain, while also achieving alignment among individual subdomains within them. Additionally, we employ minimum class confusion (MCC) as a regularizer for this sub-module. We performed experiments on SEED, SEED IV, and FACED datasets. In the cross-subject experiments, our method achieved mean classification accuracies of 87.14% on SEED, 63.24% on SEED IV, and 42.07% on FACED. In the cross-session experiments, our approach obtained average classification accuracies of 94.20% on SEED and 71.66% on SEED IV. These results demonstrate that the MS-SAMCC approach proposed in this study can effectively address EEG-based emotion recognition tasks.
Collapse
Affiliation(s)
- Lei Zhu
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
| | - Mengxuan Xu
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
| | - Aiai Huang
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
| | - Jianhai Zhang
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, China
| | - Xufei Tan
- School of Medicine, Hangzhou City University, Hangzhou, China
| |
Collapse
|
27
|
Gao H, Wang X, Chen Z, Wu M, Cai Z, Zhao L, Li J, Liu C. Graph Convolutional Network With Connectivity Uncertainty for EEG-Based Emotion Recognition. IEEE J Biomed Health Inform 2024; 28:5917-5928. [PMID: 38900625 DOI: 10.1109/jbhi.2024.3416944] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/22/2024]
Abstract
Automatic emotion recognition based on multichannel Electroencephalography (EEG) holds great potential in advancing human-computer interaction. However, several significant challenges persist in existing research on algorithmic emotion recognition. These challenges include the need for a robust model to effectively learn discriminative node attributes over long paths, the exploration of ambiguous topological information in EEG channels and effective frequency bands, and the mapping between intrinsic data qualities and provided labels. To address these challenges, this study introduces the distribution-based uncertainty method to represent spatial dependencies and temporal-spectral relativeness in EEG signals based on Graph Convolutional Network (GCN) architecture that adaptively assigns weights to functional aggregate node features, enabling effective long-path capturing while mitigating over-smoothing phenomena. Moreover, the graph mixup technique is employed to enhance latent connected edges and mitigate noisy label issues. Furthermore, we integrate the uncertainty learning method with deep GCN weights in a one-way learning fashion, termed Connectivity Uncertainty GCN (CU-GCN). We evaluate our approach on two widely used datasets, namely SEED and SEEDIV, for emotion recognition tasks. The experimental results demonstrate the superiority of our methodology over previous methods, yielding positive and significant improvements. Ablation studies confirm the substantial contributions of each component to the overall performance.
Collapse
|
28
|
Fu B, Chu W, Gu C, Liu Y. Cross-Modal Guiding Neural Network for Multimodal Emotion Recognition From EEG and Eye Movement Signals. IEEE J Biomed Health Inform 2024; 28:5865-5876. [PMID: 38917288 DOI: 10.1109/jbhi.2024.3419043] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/27/2024]
Abstract
Multimodal emotion recognition research is gaining attention because of the emerging trend of integrating information from different sensory modalities to improve performance. Electroencephalogram (EEG) signals are considered objective indicators of emotions and provide precise insights despite their complex data collection. In contrast, eye movement signals are more susceptible to environmental and individual differences but offer convenient data collection. Conventional emotion recognition methods typically use separate models for different modalities, potentially overlooking their inherent connections. This study introduces a cross-modal guiding neural network designed to fully leverage the strengths of both modalities. The network includes a dual-branch feature extraction module that simultaneously extracts features from EEG and eye movement signals. In addition, the network includes a feature guidance module that uses EEG features to direct eye movement feature extraction, reducing the impact of subjective factors. This study also introduces a feature reweighting module to explore emotion-related features within eye movement signals, thereby improving emotion classification accuracy. The empirical findings from both the SEED-IV dataset and our collected dataset substantiate the commendable performance of the model, thereby confirming its efficacy.
Collapse
|
29
|
Gong L, Chen W, Zhang D. An Attention-Based Multi-Domain Bi-Hemisphere Discrepancy Feature Fusion Model for EEG Emotion Recognition. IEEE J Biomed Health Inform 2024; 28:5890-5903. [PMID: 38913514 DOI: 10.1109/jbhi.2024.3418010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/26/2024]
Abstract
Electroencephalogram (EEG)-based emotion recognition has become a research hotspot in the field of brain-computer interface. Previous emotion recognition methods have overlooked the fusion of multi-domain emotion-specific information to improve performance, and faced the challenge of insufficient interpretability. In this paper, we proposed a novel EEG emotion recognition model that combined the asymmetry of the brain hemisphere, and the spatial, spectral, and temporal multi-domain properties of EEG signals, aiming to improve emotion recognition performance. Based on the 10-20 standard system, a global spatial projection matrix (GSPM) and a bi-hemisphere discrepancy projection matrix (BDPM) are constructed. A dual-stream spatial-spectral-temporal convolution neural network is designed to extract depth features from the two matrix paradigms. Finally, the transformer-based fusion module is used to learn the dependence of fused features, and to retain the discriminative information. We conducted extensive experiments on the SEED, SEED-IV, and DEAP public datasets, achieving excellent average results of 98.33/2.46 %, 92.15/5.13 %, 97.60/1.68 %(valence), and 97.48/1.42 %(arousal) respectively. Visualization analysis supports the interpretability of the model, and ablation experiments validate the effectiveness of multi-domain and bi-hemisphere discrepancy information fusion.
Collapse
|
30
|
Yin Y, Kong W, Tang J, Li J, Babiloni F. PSPN: Pseudo-Siamese Pyramid Network for multimodal emotion analysis. Cogn Neurodyn 2024; 18:2883-2896. [PMID: 39555297 PMCID: PMC11564494 DOI: 10.1007/s11571-024-10123-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 04/09/2024] [Accepted: 04/28/2024] [Indexed: 11/19/2024] Open
Abstract
Emotion recognition plays an important role in human life and healthcare. The EEG has been extensively researched as an objective indicator of intense emotions. However, current existing methods lack sufficient analysis of shallow and deep EEG features. In addition, human emotions are complex and variable, making it difficult to comprehensively represent emotions using a single-modal signal. As a signal associated with gaze tracking and eye movement detection, Eye-related signals provide various forms of supplementary information for multimodal emotion analysis. Therefore, we propose a Pseudo-Siamese Pyramid Network (PSPN) for multimodal emotion analysis. The PSPN model employs a Depthwise Separable Convolutional Pyramid (DSCP) to extract and integrate intrinsic emotional features at various levels and scales from EEG signals. Simultaneously, we utilize a fully connected subnetwork to extract the external emotional features from eye-related signals. Finally, we introduce a Pseudo-Siamese network that integrates a flexible cross-modal dual-branch subnetwork to collaboratively utilize EEG emotional features and eye-related behavioral features, achieving consistency and complementarity in multimodal emotion recognition. For evaluation, we conducted experiments on the DEAP and SEED-IV public datasets. The experimental results demonstrate that multimodal fusion significantly improves the accuracy of emotion recognition compared to single-modal approaches. Our PSPN model achieved the best accuracy of 96.02% and 96.45% on the valence and arousal dimensions of the DEAP dataset, and 77.81% on the SEED-IV dataset, respectively. Our code link is: https://github.com/Yinyanyan003/PSPN.git.
Collapse
Affiliation(s)
- Yanyan Yin
- College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018 China
- Key Laboratory of Brain Machine Collaborative Intelligenceof Zhejiang Province, Hangzhou, 310018 China
| | - Wanzeng Kong
- College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018 China
- Key Laboratory of Brain Machine Collaborative Intelligenceof Zhejiang Province, Hangzhou, 310018 China
| | - Jiajia Tang
- College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018 China
- Key Laboratory of Brain Machine Collaborative Intelligenceof Zhejiang Province, Hangzhou, 310018 China
| | - Jinghao Li
- College of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018 China
- Key Laboratory of Brain Machine Collaborative Intelligenceof Zhejiang Province, Hangzhou, 310018 China
| | - Fabio Babiloni
- Department of Physiology and Pharmacology, University of Rome “Sapienza”, Rome, 00185 Rome, Italy
| |
Collapse
|
31
|
Zhu L, Yu F, Ding W, Huang A, Ying N, Zhang J. Multiple-source distribution deep adaptive feature norm network for EEG emotion recognition. Cogn Neurodyn 2024; 18:2359-2372. [PMID: 39555265 PMCID: PMC11564436 DOI: 10.1007/s11571-024-10092-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2023] [Revised: 01/31/2024] [Accepted: 02/20/2024] [Indexed: 11/19/2024] Open
Abstract
Electroencephalogram (EEG) emotion recognition plays an important role in human-computer interaction, and a higher recognition accuracy can improve the user experience. In recent years, domain adaptive methods in transfer learning have been used to construct a general emotion recognition model to deal with domain difference among different subjects and sessions. However, it is still challenging to effectively reduce domain difference in domain adaptation. In this paper, we propose a Multiple-Source Distribution Deep Adaptive Feature Norm Network for EEG emotion recognition, which reduce domain difference by improving the transferability of task-specific features. In detail, the domain adaptive method of our model employs a three-layer network topology, inserts Adaptive Feature Norm to self-supervised adjustment between different layers, and combines a multiple-kernel selection approach to mean embedding matching. The method proposed in this paper achieves the best classification performance in the SEED and SEED-IV datasets. In SEED dataset, the average accuracy of cross-subject and cross-session experiments is 85.01 and 91.93%, respectively. In SEED-IV dataset, the average accuracy is 58.81% in cross-subject experiments and 59.51% in cross-session experiments. The experimental results demonstrate that our method can effectively reduce the domain difference and improve the emotion recognition accuracy.
Collapse
Affiliation(s)
- Lei Zhu
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310000 China
| | - Fei Yu
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310000 China
| | - Wangpan Ding
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310000 China
| | - Aiai Huang
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310000 China
| | - Nanjiao Ying
- School of Automation, Hangzhou Dianzi University, Hangzhou, 310000 China
| | - Jianhai Zhang
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, 310000 China
| |
Collapse
|
32
|
Wu M, Teng W, Fan C, Pei S, Li P, Pei G, Li T, Liang W, Lv Z. Multimodal Emotion Recognition Based on EEG and EOG Signals Evoked by the Video-Odor Stimuli. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3496-3505. [PMID: 39255190 DOI: 10.1109/tnsre.2024.3457580] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/12/2024]
Abstract
Affective data is the basis of emotion recognition, which is mainly acquired through extrinsic elicitation. To investigate the enhancing effects of multi-sensory stimuli on emotion elicitation and emotion recognition, we designed an experimental paradigm involving visual, auditory, and olfactory senses. A multimodal emotional dataset (OVPD-II) that employed the video-only or video-odor patterns as the stimuli materials, and recorded the electroencephalogram (EEG) and electrooculogram (EOG) signals, was created. The feedback results reported by subjects after each trial demonstrated that the video-odor pattern outperformed the video-only pattern in evoking individuals' emotions. To further validate the efficiency of the video-odor pattern, the transformer was employed to perform the emotion recognition task, where the highest accuracy reached 86.65% (66.12%) for EEG (EOG) modality with the video-odor pattern, which improved by 1.42% (3.43%) compared with the video-only pattern. What's more, the hybrid fusion (HF) method combined with the transformer and joint training was developed to improve the performance of the emotion recognition task, which achieved classify accuracies of 89.50% and 88.47% for the video-odor and video-only patterns, respectively.
Collapse
|
33
|
Lee MH, Shomanov A, Begim B, Kabidenova Z, Nyssanbay A, Yazici A, Lee SW. EAV: EEG-Audio-Video Dataset for Emotion Recognition in Conversational Contexts. Sci Data 2024; 11:1026. [PMID: 39300129 PMCID: PMC11413008 DOI: 10.1038/s41597-024-03838-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 08/29/2024] [Indexed: 09/22/2024] Open
Abstract
Understanding emotional states is pivotal for the development of next-generation human-machine interfaces. Human behaviors in social interactions have resulted in psycho-physiological processes influenced by perceptual inputs. Therefore, efforts to comprehend brain functions and human behavior could potentially catalyze the development of AI models with human-like attributes. In this study, we introduce a multimodal emotion dataset comprising data from 30-channel electroencephalography (EEG), audio, and video recordings from 42 participants. Each participant engaged in a cue-based conversation scenario, eliciting five distinct emotions: neutral, anger, happiness, sadness, and calmness. Throughout the experiment, each participant contributed 200 interactions, which encompassed both listening and speaking. This resulted in a cumulative total of 8,400 interactions across all participants. We evaluated the baseline performance of emotion recognition for each modality using established deep neural network (DNN) methods. The Emotion in EEG-Audio-Visual (EAV) dataset represents the first public dataset to incorporate three primary modalities for emotion recognition within a conversational context. We anticipate that this dataset will make significant contributions to the modeling of the human emotional process, encompassing both fundamental neuroscience and machine learning viewpoints.
Collapse
Affiliation(s)
- Min-Ho Lee
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Adai Shomanov
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Balgyn Begim
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Zhuldyz Kabidenova
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Aruna Nyssanbay
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Adnan Yazici
- Nazarbayev University, Department of Computer Science, Astana, 010000, Republic of Kazakhstan
| | - Seong-Whan Lee
- Korea University, Department of Artificial Intelligence, Seoul, 02841, Republic of Korea.
| |
Collapse
|
34
|
Pang B, Peng Y, Gao J, Kong W. Semi-supervised bipartite graph construction with active EEG sample selection for emotion recognition. Med Biol Eng Comput 2024; 62:2805-2824. [PMID: 38700614 DOI: 10.1007/s11517-024-03094-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2023] [Accepted: 04/10/2024] [Indexed: 08/18/2024]
Abstract
Electroencephalogram (EEG) signals are derived from the central nervous system and inherently difficult to camouflage, leading to the recent popularity of EEG-based emotion recognition. However, due to the non-stationary nature of EEG, inter-subject variabilities become obstacles for recognition models to well adapt to different subjects. In this paper, we propose a novel approach called semi-supervised bipartite graph construction with active EEG sample selection (SBGASS) for cross-subject emotion recognition, which offers two significant advantages. Firstly, SBGASS adaptively learns a bipartite graph to characterize the underlying relationships between labeled and unlabeled EEG samples, effectively implementing the semantic connection for samples from different subjects. Secondly, we employ active sample selection technique in this paper to reduce the impact of negative samples (outliers or noise in the data) on bipartite graph construction. Drawing from the experimental results with the SEED-IV data set, we have gained the following three insights. (1) SBGASS actively rejects negative labeled samples, which helps mitigate the impact of negative samples when constructing the optimal bipartite graph and improves the model performance. (2) Through the learned optimal bipartite graph in SBGASS, the transferability of labeled EEG samples is quantitatively analyzed, which exhibits a decreasing tendency as the distance between each labeled sample and the corresponding class centroid increases. (3) Besides the improved recognition accuracy, the spatial-frequency patterns in emotion recognition are investigated by the acquired projection matrix.
Collapse
Affiliation(s)
- Bowen Pang
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang, China
| | - Yong Peng
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang, China.
- Zhejiang Key Laboratory of Brain-Machine Collaborative Intelligence, Hangzhou, 310018, Zhejiang, China.
| | - Jian Gao
- Department of Rehabilitation, Hangzhou Mingzhou Brain Rehabilitation Hospital, Hangzhou, 311215, Zhejiang, China
| | - Wanzeng Kong
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang, China
- Zhejiang Key Laboratory of Brain-Machine Collaborative Intelligence, Hangzhou, 310018, Zhejiang, China
| |
Collapse
|
35
|
Fu B, Yu X, Jiang G, Sun N, Liu Y. Enhancing local representation learning through global-local integration with functional connectivity for EEG-based emotion recognition. Comput Biol Med 2024; 179:108857. [PMID: 39018882 DOI: 10.1016/j.compbiomed.2024.108857] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2024] [Revised: 06/21/2024] [Accepted: 07/06/2024] [Indexed: 07/19/2024]
Abstract
Emotion recognition based on electroencephalogram (EEG) signals is crucial in understanding human affective states. Current research has limitations in extracting local features. The representation capabilities of local features are limited, making it difficult to comprehensively capture emotional information. In this study, a novel approach is proposed to enhance local representation learning through global-local integration with functional connectivity for EEG-based emotion recognition. By leveraging the functional connectivity of brain regions, EEG signals are divided into global embeddings that represent comprehensive brain connectivity patterns throughout the entire process and local embeddings that reflect dynamic interactions within specific brain functional networks at particular moments. Firstly, a convolutional feature extraction branch based on the residual network is designed to extract local features from the global embedding. To further improve the representation ability and accuracy of local features, a multidimensional collaborative attention (MCA) module is introduced. Secondly, the local features and patch embedded local embeddings are integrated into the feature coupling module (FCM), which utilizes hierarchical connections and enhanced cross-attention to couple region-level features, thereby enhancing local representation learning. Experimental results on three public datasets show that compared with other methods, this method improves accuracy by 4.92% on the DEAP, by 1.11% on the SEED, and by 7.76% on the SEED-IV, demonstrating its superior performance in emotion recognition tasks.
Collapse
Affiliation(s)
- Baole Fu
- School of Automation, Qingdao University, Qingdao 266071, China; Institute for Future, Qingdao University, Qingdao 266071, China.
| | - Xiangkun Yu
- College of Computer Science & Technology, Qingdao University, Qingdao 266071, China; Institute for Future, Qingdao University, Qingdao 266071, China.
| | - Guijie Jiang
- College of Computer Science & Technology, Qingdao University, Qingdao 266071, China.
| | - Ninghao Sun
- College of Computer Science & Technology, Qingdao University, Qingdao 266071, China.
| | - Yinhua Liu
- School of Automation, Qingdao University, Qingdao 266071, China; Shandong Key Laboratory of Industrial Control Technology, Qingdao 266071, China; Institute for Future, Qingdao University, Qingdao 266071, China.
| |
Collapse
|
36
|
Choi SH. Spiking neural networks for biomedical signal analysis. Biomed Eng Lett 2024; 14:955-966. [PMID: 39220024 PMCID: PMC11362400 DOI: 10.1007/s13534-024-00405-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Revised: 05/15/2024] [Accepted: 06/11/2024] [Indexed: 09/04/2024] Open
Abstract
Artificial intelligence (AI) has had a significant impact on human life because of its pervasiveness across industries and its rapid development. Although AI has achieved superior performance in learning and reasoning, it encounters challenges such as substantial computational demands, privacy concerns, communication delays, and high energy consumption associated with cloud-based models. These limitations have facilitated a paradigm change in on-device AI processing, which offers enhanced privacy, reduced latency, and improved power efficiency through the direct execution of computations on devices. With advancements in neuromorphic systems, spiking neural networks (SNNs), often referred to as the next generation of AI, are currently in focus as on-device AI. These technologies aim to mimic the human brain efficiency and provide promising real-time processing with minimal energy. This study reviewed the application of SNNs in the analysis of biomedical signals (electroencephalograms, electrocardiograms, and electromyograms), and consequently, investigated the distinctive attributes and prospective future paths of SNNs models in the field of biomedical signal analysis.
Collapse
Affiliation(s)
- Sang Ho Choi
- School of Computer and Information Engineering, Kwangwoon University, Seoul, 01897 Korea
| |
Collapse
|
37
|
Farhadi Sedehi J, Jafarnia Dabanloo N, Maghooli K, Sheikhani A. Multimodal insights into granger causality connectivity: Integrating physiological signals and gated eye-tracking data for emotion recognition using convolutional neural network. Heliyon 2024; 10:e36411. [PMID: 39253213 PMCID: PMC11381760 DOI: 10.1016/j.heliyon.2024.e36411] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Revised: 08/14/2024] [Accepted: 08/14/2024] [Indexed: 09/11/2024] Open
Abstract
This study introduces a groundbreaking method to enhance the accuracy and reliability of emotion recognition systems by combining electrocardiogram (ECG) with electroencephalogram (EEG) data, using an eye-tracking gated strategy. Initially, we propose a technique to filter out irrelevant portions of emotional data by employing pupil diameter metrics from eye-tracking data. Subsequently, we introduce an innovative approach for estimating effective connectivity to capture the dynamic interaction between the brain and the heart during emotional states of happiness and sadness. Granger causality (GC) is estimated and utilized to optimize input for a highly effective pre-trained convolutional neural network (CNN), specifically ResNet-18. To assess this methodology, we employed EEG and ECG data from the publicly available MAHNOB-HCI database, using a 5-fold cross-validation approach. Our method achieved an impressive average accuracy and area under the curve (AUC) of 91.00 % and 0.97, respectively, for GC-EEG-ECG images processed with ResNet-18. Comparative analysis with state-of-the-art studies clearly shows that augmenting ECG with EEG and refining data with an eye-tracking strategy significantly enhances emotion recognition performance across various emotions.
Collapse
Affiliation(s)
- Javid Farhadi Sedehi
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Nader Jafarnia Dabanloo
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Keivan Maghooli
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Ali Sheikhani
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| |
Collapse
|
38
|
Zhang Y, Liao Y, Chen W, Zhang X, Huang L. Emotion recognition of EEG signals based on contrastive learning graph convolutional model. J Neural Eng 2024; 21:046060. [PMID: 39151459 DOI: 10.1088/1741-2552/ad7060] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2024] [Accepted: 08/16/2024] [Indexed: 08/19/2024]
Abstract
Objective.Electroencephalogram (EEG) signals offer invaluable insights into the complexities of emotion generation within the brain. Yet, the variability in EEG signals across individuals presents a formidable obstacle for empirical implementations. Our research addresses these challenges innovatively, focusing on the commonalities within distinct subjects' EEG data.Approach.We introduce a novel approach named Contrastive Learning Graph Convolutional Network (CLGCN). This method captures the distinctive features and crucial channel nodes related to individuals' emotional states. Specifically, CLGCN merges the dual benefits of CL's synchronous multisubject data learning and the GCN's proficiency in deciphering brain connectivity matrices. Understanding multifaceted brain functions and their information interchange processes is realized as CLGCN generates a standardized brain network learning matrix during a dataset's learning process.Main results.Our model underwent rigorous testing on the Database for Emotion Analysis using Physiological Signals (DEAP) and SEED datasets. In the five-fold cross-validation used for dependent subject experimental setting, it achieved an accuracy of 97.13% on the DEAP dataset and surpassed 99% on the SEED and SEED_IV datasets. In the incremental learning experiments with the SEED dataset, merely 5% of the data was sufficient to fine-tune the model, resulting in an accuracy of 92.8% for the new subject. These findings validate the model's efficacy.Significance.This work combines CL with GCN, improving the accuracy of decoding emotional states from EEG signals and offering valuable insights into uncovering the underlying mechanisms of emotional processes in the brain.
Collapse
Affiliation(s)
- Yiling Zhang
- College of electronic and optical engineering & college of flexible electronics (future technology), Nanjing University of Posts and Telecommunications, Jiangsu 210023, People's Republic of China
| | - Yuan Liao
- College of electronic and optical engineering & college of flexible electronics (future technology), Nanjing University of Posts and Telecommunications, Jiangsu 210023, People's Republic of China
| | - Wei Chen
- College of electronic and optical engineering & college of flexible electronics (future technology), Nanjing University of Posts and Telecommunications, Jiangsu 210023, People's Republic of China
| | - Xiruo Zhang
- College of electronic and optical engineering & college of flexible electronics (future technology), Nanjing University of Posts and Telecommunications, Jiangsu 210023, People's Republic of China
| | - Liya Huang
- College of electronic and optical engineering & college of flexible electronics (future technology), Nanjing University of Posts and Telecommunications, Jiangsu 210023, People's Republic of China
| |
Collapse
|
39
|
Pfeffer MA, Ling SSH, Wong JKW. Exploring the frontier: Transformer-based models in EEG signal analysis for brain-computer interfaces. Comput Biol Med 2024; 178:108705. [PMID: 38865781 DOI: 10.1016/j.compbiomed.2024.108705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2023] [Revised: 05/01/2024] [Accepted: 06/01/2024] [Indexed: 06/14/2024]
Abstract
This review systematically explores the application of transformer-based models in EEG signal processing and brain-computer interface (BCI) development, with a distinct focus on ensuring methodological rigour and adhering to empirical validations within the existing literature. By examining various transformer architectures, such as the Temporal Spatial Transformer Network (TSTN) and EEG Conformer, this review delineates their capabilities in mitigating challenges intrinsic to EEG data, such as noise and artifacts, and their subsequent implications on decoding and classification accuracies across disparate mental tasks. The analytical scope extends to a meticulous examination of attention mechanisms within transformer models, delineating their role in illuminating critical temporal and spatial EEG features and facilitating interpretability in model decision-making processes. The discourse additionally encapsulates emerging works that substantiate the efficacy of transformer models in noise reduction of EEG signals and diversifying applications beyond the conventional motor imagery paradigm. Furthermore, this review elucidates evident gaps and propounds exploratory avenues in the applications of pre-trained transformers in EEG analysis and the potential expansion into real-time and multi-task BCI applications. Collectively, this review distils extant knowledge, navigates through the empirical findings, and puts forward a structured synthesis, thereby serving as a conduit for informed future research endeavours in transformer-enhanced, EEG-based BCI systems.
Collapse
Affiliation(s)
- Maximilian Achim Pfeffer
- Faculty of Engineering and Information Technology, University of Technology Sydney, CB11 81-113, Broadway, Ultimo, 2007, New South Wales, Australia.
| | - Steve Sai Ho Ling
- Faculty of Engineering and Information Technology, University of Technology Sydney, CB11 81-113, Broadway, Ultimo, 2007, New South Wales, Australia.
| | - Johnny Kwok Wai Wong
- Faculty of Design, Architecture and Building, University of Technology Sydney, 15 Broadway, Ultimo, 2007, New South Wales, Australia.
| |
Collapse
|
40
|
Li C, Li P, Zhang Y, Li N, Si Y, Li F, Cao Z, Chen H, Chen B, Yao D, Xu P. Effective Emotion Recognition by Learning Discriminative Graph Topologies in EEG Brain Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10258-10272. [PMID: 37022389 DOI: 10.1109/tnnls.2023.3238519] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Multichannel electroencephalogram (EEG) is an array signal that represents brain neural networks and can be applied to characterize information propagation patterns for different emotional states. To reveal these inherent spatial graph features and increase the stability of emotion recognition, we propose an effective emotion recognition model that performs multicategory emotion recognition with multiple emotion-related spatial network topology patterns (MESNPs) by learning discriminative graph topologies in EEG brain networks. To evaluate the performance of our proposed MESNP model, we conducted single-subject and multisubject four-class classification experiments on two public datasets, MAHNOB-HCI and DEAP. Compared with existing feature extraction methods, the MESNP model significantly enhances the multiclass emotional classification performance in the single-subject and multisubject conditions. To evaluate the online version of the proposed MESNP model, we designed an online emotion monitoring system. We recruited 14 participants to conduct the online emotion decoding experiments. The average online experimental accuracy of the 14 participants was 84.56%, indicating that our model can be applied in affective brain-computer interface (aBCI) systems. The offline and online experimental results demonstrate that the proposed MESNP model effectively captures discriminative graph topology patterns and significantly improves emotion classification performance. Moreover, the proposed MESNP model provides a new scheme for extracting features from strongly coupled array signals.
Collapse
|
41
|
Wang Y, Chen CB, Imamura T, Tapia IE, Somers VK, Zee PC, Lim DC. A novel methodology for emotion recognition through 62-lead EEG signals: multilevel heterogeneous recurrence analysis. Front Physiol 2024; 15:1425582. [PMID: 39119215 PMCID: PMC11306145 DOI: 10.3389/fphys.2024.1425582] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Accepted: 06/27/2024] [Indexed: 08/10/2024] Open
Abstract
Objective Recognizing emotions from electroencephalography (EEG) signals is a challenging task due to the complex, nonlinear, and nonstationary characteristics of brain activity. Traditional methods often fail to capture these subtle dynamics, while deep learning approaches lack explainability. In this research, we introduce a novel three-phase methodology integrating manifold embedding, multilevel heterogeneous recurrence analysis (MHRA), and ensemble learning to address these limitations in EEG-based emotion recognition. Approach The proposed methodology was evaluated using the SJTU-SEED IV database. We first applied uniform manifold approximation and projection (UMAP) for manifold embedding of the 62-lead EEG signals into a lower-dimensional space. We then developed MHRA to characterize the complex recurrence dynamics of brain activity across multiple transition levels. Finally, we employed tree-based ensemble learning methods to classify four emotions (neutral, sad, fear, happy) based on the extracted MHRA features. Main results Our approach achieved high performance, with an accuracy of 0.7885 and an AUC of 0.7552, outperforming existing methods on the same dataset. Additionally, our methodology provided the most consistent recognition performance across different emotions. Sensitivity analysis revealed specific MHRA metrics that were strongly associated with each emotion, offering valuable insights into the underlying neural dynamics. Significance This study presents a novel framework for EEG-based emotion recognition that effectively captures the complex nonlinear and nonstationary dynamics of brain activity while maintaining explainability. The proposed methodology offers significant potential for advancing our understanding of emotional processing and developing more reliable emotion recognition systems with broad applications in healthcare and beyond.
Collapse
Affiliation(s)
- Yujie Wang
- Department of Industrial and Systems Engineering, University of Miami, Coral Gables, FL, United States
| | - Cheng-Bang Chen
- Department of Industrial and Systems Engineering, University of Miami, Coral Gables, FL, United States
| | - Toshihiro Imamura
- Division of Sleep Medicine, Department of Medicine, University of Pennsylvania, Phialdelphia, PA, United States
- Division of Pulmonary and Sleep Medicine, Children’s Hospital of Philadelphia, Phialdelphia, PA, United States
| | - Ignacio E. Tapia
- Division of Pediatric Pulmonology, Miller School of Medicine, University of Miami, Miami, FL, United States
| | - Virend K. Somers
- Department of Cardiovascular Medicine, Mayo Clinic, Rochester, MN, United States
| | - Phyllis C. Zee
- Center for Circadian and Sleep Medicine, Department of Neurology, Feinberg School of Medicine, Northwestern University, Chicago, IL, United States
| | - Diane C. Lim
- Department of Medicine, Miami VA Medical Center, Miami, FL, United States
- Department of Medicine, Miller School of Medicine, University of Miami, Miami, FL, United States
| |
Collapse
|
42
|
Chen W, Liao Y, Dai R, Dong Y, Huang L. EEG-based emotion recognition using graph convolutional neural network with dual attention mechanism. Front Comput Neurosci 2024; 18:1416494. [PMID: 39099770 PMCID: PMC11294218 DOI: 10.3389/fncom.2024.1416494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 06/26/2024] [Indexed: 08/06/2024] Open
Abstract
EEG-based emotion recognition is becoming crucial in brain-computer interfaces (BCI). Currently, most researches focus on improving accuracy, while neglecting further research on the interpretability of models, we are committed to analyzing the impact of different brain regions and signal frequency bands on emotion generation based on graph structure. Therefore, this paper proposes a method named Dual Attention Mechanism Graph Convolutional Neural Network (DAMGCN). Specifically, we utilize graph convolutional neural networks to model the brain network as a graph to extract representative spatial features. Furthermore, we employ the self-attention mechanism of the Transformer model which allocates more electrode channel weights and signal frequency band weights to important brain regions and frequency bands. The visualization of attention mechanism clearly demonstrates the weight allocation learned by DAMGCN. During the performance evaluation of our model on the DEAP, SEED, and SEED-IV datasets, we achieved the best results on the SEED dataset, showing subject-dependent experiments' accuracy of 99.42% and subject-independent experiments' accuracy of 73.21%. The results are demonstrably superior to the accuracies of most existing models in the realm of EEG-based emotion recognition.
Collapse
Affiliation(s)
| | | | | | | | - Liya Huang
- College of Electronic and Optical Engineering & College of Flexible Electronics (Future Technology), Nanjing University of Posts and Telecommunications, Nanjing, China
| |
Collapse
|
43
|
Wang Z, Yu J, Gao J, Bai Y, Wan Z. MutaPT: A Multi-Task Pre-Trained Transformer for Classifying State of Disorders of Consciousness Using EEG Signal. Brain Sci 2024; 14:688. [PMID: 39061428 PMCID: PMC11274898 DOI: 10.3390/brainsci14070688] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2024] [Revised: 05/22/2024] [Accepted: 05/25/2024] [Indexed: 07/28/2024] Open
Abstract
Deep learning (DL) has been demonstrated to be a valuable tool for classifying state of disorders of consciousness (DOC) using EEG signals. However, the performance of the DL-based DOC state classification is often challenged by the limited size of EEG datasets. To overcome this issue, we introduce multiple open-source EEG datasets to increase data volume and train a novel multi-task pre-training Transformer model named MutaPT. Furthermore, we propose a cross-distribution self-supervised (CDS) pre-training strategy to enhance the model's generalization ability, addressing data distribution shifts across multiple datasets. An EEG dataset of DOC patients is used to validate the effectiveness of our methods for the task of classifying DOC states. Experimental results show the superiority of our MutaPT over several DL models for EEG classification.
Collapse
Affiliation(s)
- Zihan Wang
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Junqi Yu
- School of Information Engineering, Nanchang University, Nanchang 330031, China
| | - Jiahui Gao
- School of Public Policy and Administration, Nanchang University, Nanchang 330031, China
| | - Yang Bai
- Affiliated Rehabilitation Hospital, Jiangxi Medical College, Nanchang University, Nanchang 330031, China
| | - Zhijiang Wan
- School of Information Engineering, Nanchang University, Nanchang 330031, China
- Industrial Institute of Artificial Intelligence, Nanchang University, Nanchang 330031, China
| |
Collapse
|
44
|
Jin X, Yang X, Kong W, Zhu L, Tang J, Peng Y, Ding Y, Zhao Q. TSFAN: tensorized spatial-frequency attention network with domain adaptation for cross-session EEG-based biometric recognition. J Neural Eng 2024; 21:046005. [PMID: 38866001 DOI: 10.1088/1741-2552/ad5761] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Accepted: 06/12/2024] [Indexed: 06/14/2024]
Abstract
Objective.Electroencephalogram (EEG) signals are promising biometrics owning to their invisibility, adapting to the application scenarios with high-security requirements. However, It is challenging to explore EEG identity features without the interference of device and state differences of the subject across sessions. Existing methods treat training sessions as a single domain, affected by the different data distribution among sessions. Although most multi-source unsupervised domain adaptation (MUDA) methods bridge the domain gap between multiple source and target domains individually, relationships among the domain-invariant features of each distribution alignment are neglected.Approach.In this paper, we propose a MUDA method, Tensorized Spatial-Frequency Attention Network (TSFAN), to assist the performance of the target domain for EEG-based biometric recognition. Specifically, significant relationships of domain-invariant features are modeled via a tensorized attention mechanism. It jointly incorporates appropriate common spatial-frequency representations of pairwise source and target but also cross-source domains, without the effect of distribution discrepancy among source domains. Additionally, considering the curse of dimensionality, our TSFAN is approximately represented in Tucker format. Benefiting the low-rank Tucker Network, the TSFAN can scale linearly in the number of domains, providing us the great flexibility to extend TSFAN to the case associated with an arbitrary number of sessions.Main results.Extensive experiments on the representative benchmarks demonstrate the effectiveness of TSFAN in EEG-based biometric recognition, outperforming state-of-the-art approaches, as verified by cross-session validation.Significance.The proposed TSFAN aims to investigate the presence of consistent EEG identity features across sessions. It is achieved by utilizing a novel tensorized attention mechanism that collaborates intra-source transferable information with inter-source interactions, while remaining unaffected by domain shifts in multiple source domains. Furthermore, the electrode selection shows that EEG-based identity features across sessions are distributed across brain regions, and 20 electrodes based on 10-20 standard system are able to extract stable identity information.
Collapse
Affiliation(s)
- Xuanyu Jin
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, People's Republic of China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, People's Republic of China
| | - Xinyu Yang
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, People's Republic of China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, People's Republic of China
| | - Wanzeng Kong
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, People's Republic of China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, People's Republic of China
| | - Li Zhu
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, People's Republic of China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, People's Republic of China
| | - Jiajia Tang
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, People's Republic of China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, People's Republic of China
| | - Yong Peng
- School of Computer Science, Hangzhou Dianzi University, Hangzhou, People's Republic of China
- Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, People's Republic of China
| | - Yu Ding
- Netease Fuxi AI Lab, NetEase, Hangzhou, People's Republic of China
| | - Qibin Zhao
- Tensor Learning Unit, Center for Advanced Intelligence Project, RIKEN, Tokyo, Japan
| |
Collapse
|
45
|
Zhang H, Zuo T, Chen Z, Wang X, Sun PZH. Evolutionary Ensemble Learning for EEG-Based Cross-Subject Emotion Recognition. IEEE J Biomed Health Inform 2024; 28:3872-3881. [PMID: 38954558 DOI: 10.1109/jbhi.2024.3384816] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/04/2024]
Abstract
Electroencephalogram (EEG) has been widely utilized in emotion recognition due to its high temporal resolution and reliability. However, the individual differences and non-stationary characteristics of EEG, along with the complexity and variability of emotions, pose challenges in generalizing emotion recognition models across subjects. In this paper, an end-to-end framework is proposed to improve the performance of cross-subject emotion recognition. A novel evolutionary programming (EP)-based optimization strategy with neural network (NN) as the base classifier termed NN ensemble with EP (EPNNE) is designed for cross-subject emotion recognition. The effectiveness of the proposed method is evaluated on the publicly available DEAP, FACED, SEED, and SEED-IV datasets. Numerical results demonstrate that the proposed method is superior to state-of-the-art cross-subject emotion recognition methods. The proposed end-to-end framework for cross-subject emotion recognition aids biomedical researchers in effectively assessing individual emotional states, thereby enabling efficient treatment and interventions.
Collapse
|
46
|
Goshvarpour A, Goshvarpour A. Lemniscate of Bernoulli's map quantifiers: innovative measures for EEG emotion recognition. Cogn Neurodyn 2024; 18:1061-1077. [PMID: 38826652 PMCID: PMC11143135 DOI: 10.1007/s11571-023-09968-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2022] [Revised: 03/18/2023] [Accepted: 04/05/2023] [Indexed: 06/04/2024] Open
Abstract
Thanks to the advent of affective computing, designing an automatic human emotion recognition system for clinical and non-clinical applications has attracted the attention of many researchers. Currently, multi-channel electroencephalogram (EEG)-based emotion recognition is a fundamental but challenging issue. This experiment envisioned developing a new scheme for automated EEG affect recognition. An innovative nonlinear feature engineering approach was presented based on Lemniscate of Bernoulli's Map (LBM), which belongs to the family of chaotic maps, in line with the EEG's nonlinear nature. As far as the authors know, LBM has not been utilized for biological signal analysis. Next, the map was characterized using several graphical indices. The feature vector was imposed on the feature selection algorithm while evaluating the role of the feature vector dimension on emotion recognition rates. Finally, the efficiency of the features on emotion recognition was appraised using two conventional classifiers and validated using the Database for Emotion Analysis using Physiological signals (DEAP) and SJTU Emotion EEG Dataset-IV (SEED-IV) benchmark databases. The experimental results showed a maximum accuracy of 92.16% for DEAP and 90.7% for SEED-IV. Achieving higher recognition rates compared to the state-of-art EEG emotion recognition systems suggest the proposed method based on LBM could have potential both in characterizing bio-signal dynamics and detecting affect-deficit disorders.
Collapse
Affiliation(s)
- Atefeh Goshvarpour
- Department of Biomedical Engineering, Faculty of Electrical Engineering, Sahand University of Technology, Tabriz, Iran
| | - Ateke Goshvarpour
- Department of Biomedical Engineering, Imam Reza International University, Mashhad, Razavi Khorasan Iran
- Health Technology Research Center, Imam Reza International University, Mashhad, Razavi Khorasan Iran
| |
Collapse
|
47
|
Hamzah HA, Abdalla KK. EEG-based emotion recognition systems; comprehensive study. Heliyon 2024; 10:e31485. [PMID: 38818173 PMCID: PMC11137547 DOI: 10.1016/j.heliyon.2024.e31485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 05/16/2024] [Indexed: 06/01/2024] Open
Abstract
Emotion recognition technology through EEG signal analysis is currently a fundamental concept in artificial intelligence. This recognition has major practical implications in emotional health care, human-computer interaction, and so on. This paper provides a comprehensive study of different methods for extracting electroencephalography (EEG) features for emotion recognition from four different perspectives, including time domain features, frequency domain features, time-frequency features, and nonlinear features. We summarize the current pattern recognition methods adopted in most related works, and with the rapid development of deep learning (DL) attracting the attention of researchers in this field, we pay more attention to deep learning-based studies and analyse the characteristics, advantages, disadvantages, and applicable scenarios. Finally, the current challenges and future development directions in this field were summarized. This paper can help novice researchers in this field gain a systematic understanding of the current status of emotion recognition research based on EEG signals and provide ideas for subsequent related research.
Collapse
Affiliation(s)
- Hussein Ali Hamzah
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| | - Kasim K. Abdalla
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| |
Collapse
|
48
|
Ma W, Zheng Y, Li T, Li Z, Li Y, Wang L. A comprehensive review of deep learning in EEG-based emotion recognition: classifications, trends, and practical implications. PeerJ Comput Sci 2024; 10:e2065. [PMID: 38855206 PMCID: PMC11157589 DOI: 10.7717/peerj-cs.2065] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Accepted: 04/25/2024] [Indexed: 06/11/2024]
Abstract
Emotion recognition utilizing EEG signals has emerged as a pivotal component of human-computer interaction. In recent years, with the relentless advancement of deep learning techniques, using deep learning for analyzing EEG signals has assumed a prominent role in emotion recognition. Applying deep learning in the context of EEG-based emotion recognition carries profound practical implications. Although many model approaches and some review articles have scrutinized this domain, they have yet to undergo a comprehensive and precise classification and summarization process. The existing classifications are somewhat coarse, with insufficient attention given to the potential applications within this domain. Therefore, this article systematically classifies recent developments in EEG-based emotion recognition, providing researchers with a lucid understanding of this field's various trajectories and methodologies. Additionally, it elucidates why distinct directions necessitate distinct modeling approaches. In conclusion, this article synthesizes and dissects the practical significance of EEG signals in emotion recognition, emphasizing its promising avenues for future application.
Collapse
Affiliation(s)
- Weizhi Ma
- School of Information Science and Technology, North China University of Technology, Beijing, China
| | - Yujia Zheng
- School of Information Science and Technology, North China University of Technology, Beijing, China
| | - Tianhao Li
- School of Information Science and Technology, North China University of Technology, Beijing, China
| | - Zhengping Li
- School of Information Science and Technology, North China University of Technology, Beijing, China
| | - Ying Li
- School of Information Science and Technology, North China University of Technology, Beijing, China
| | - Lijun Wang
- School of Information Science and Technology, North China University of Technology, Beijing, China
| |
Collapse
|
49
|
Liang Y, Zhang C, An S, Wang Z, Shi K, Peng T, Ma Y, Xie X, He J, Zheng K. FetchEEG: a hybrid approach combining feature extraction and temporal-channel joint attention for EEG-based emotion classification. J Neural Eng 2024; 21:036011. [PMID: 38701773 DOI: 10.1088/1741-2552/ad4743] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Accepted: 05/03/2024] [Indexed: 05/05/2024]
Abstract
Objective. Electroencephalogram (EEG) analysis has always been an important tool in neural engineering, and the recognition and classification of human emotions are one of the important tasks in neural engineering. EEG data, obtained from electrodes placed on the scalp, represent a valuable resource of information for brain activity analysis and emotion recognition. Feature extraction methods have shown promising results, but recent trends have shifted toward end-to-end methods based on deep learning. However, these approaches often overlook channel representations, and their complex structures pose certain challenges to model fitting.Approach. To address these challenges, this paper proposes a hybrid approach named FetchEEG that combines feature extraction and temporal-channel joint attention. Leveraging the advantages of both traditional feature extraction and deep learning, the FetchEEG adopts a multi-head self-attention mechanism to extract representations between different time moments and channels simultaneously. The joint representations are then concatenated and classified using fully-connected layers for emotion recognition. The performance of the FetchEEG is verified by comparison experiments on a self-developed dataset and two public datasets.Main results. In both subject-dependent and subject-independent experiments, the FetchEEG demonstrates better performance and stronger generalization ability than the state-of-the-art methods on all datasets. Moreover, the performance of the FetchEEG is analyzed for different sliding window sizes and overlap rates in the feature extraction module. The sensitivity of emotion recognition is investigated for three- and five-frequency-band scenarios.Significance. FetchEEG is a novel hybrid method based on EEG for emotion classification, which combines EEG feature extraction with Transformer neural networks. It has achieved state-of-the-art performance on both self-developed datasets and multiple public datasets, with significantly higher training efficiency compared to end-to-end methods, demonstrating its effectiveness and feasibility.
Collapse
Affiliation(s)
- Yu Liang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Chenlong Zhang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Shan An
- JD Health International Inc., Beijing, People's Republic of China
| | - Zaitian Wang
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Kaize Shi
- University of Technology Sydney, Sydney, Australia
| | - Tianhao Peng
- Beihang University, Beijing, People's Republic of China
| | - Yuqing Ma
- Beihang University, Beijing, People's Republic of China
| | - Xiaoyang Xie
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Jian He
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| | - Kun Zheng
- Faculty of Information Technology, Beijing University of Technology, Beijing, People's Republic of China
| |
Collapse
|
50
|
Jin H, Gao Y, Wang T, Gao P. DAST: A Domain-Adaptive Learning Combining Spatio-Temporal Dynamic Attention for Electroencephalography Emotion Recognition. IEEE J Biomed Health Inform 2024; 28:2512-2523. [PMID: 37607151 DOI: 10.1109/jbhi.2023.3307606] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/24/2023]
Abstract
Multimodal emotion recognition with EEG-based have become mainstream in affective computing. However, previous studies mainly focus on perceived emotions (including posture, speech or face expression et al.) of different subjects, while the lack of research on induced emotions (including video or music et al.) limited the development of two-ways emotions. To solve this problem, we propose a multimodal domain adaptive method based on EEG and music called the DAST, which uses spatio-temporal adaptive attention (STA-attention) to globally model the EEG and maps all embeddings dynamically into high-dimensionally space by adaptive space encoder (ASE). Then, adversarial training is performed with domain discriminator and ASE to learn invariant emotion representations. Furthermore, we conduct extensive experiments on the DEAP dataset, and the results show that our method can further explore the relationship between induced and perceived emotions, and provide a reliable reference for exploring the potential correlation between EEG and music stimulation.
Collapse
|