1
|
Gunda NK, Khalaf MI, Bhatnagar S, Quraishi A, Gudala L, Venkata AKP, Alghayadh FY, Alsubai S, Bhatnagar V. Lightweight attention mechanisms for EEG emotion recognition for brain computer interface. J Neurosci Methods 2024; 410:110223. [PMID: 39032522 DOI: 10.1016/j.jneumeth.2024.110223] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2024] [Revised: 06/18/2024] [Accepted: 07/17/2024] [Indexed: 07/23/2024]
Abstract
BACKGROUND In the realm of brain-computer interfaces (BCI), identifying emotions from electroencephalogram (EEG) data is a difficult endeavor because of the volume of data, the intricacy of the signals, and the several channels that make up the signals. NEW METHODS Using dual-stream structure scaling and multiple attention mechanisms (LDMGEEG), a lightweight network is provided to maximize the accuracy and performance of EEG-based emotion identification. Reducing the number of computational parameters while maintaining the current level of classification accuracy is the aim. This network employs a symmetric dual-stream architecture to assess separately time-domain and frequency-domain spatio-temporal maps constructed using differential entropy features of EEG signals as inputs. RESULT The experimental results show that after significantly lowering the number of parameters, the model achieved the best possible performance in the field, with a 95.18 % accuracy on the SEED dataset. COMPARISON WITH EXISTING METHODS Moreover, it reduced the number of parameters by 98 % when compared to existing models. CONCLUSION The proposed method distinct channel-time/frequency-space multiple attention and post-attention methods enhance the model's ability to aggregate features and result in lightweight performance.
Collapse
Affiliation(s)
- Naresh Kumar Gunda
- Information Technology Management, Campbellsville Univeristy, Campbellsville, KY, United States.
| | - Mohammed I Khalaf
- Department of computer science, Al Maarif University College, Al Anbar 31001, Iraq.
| | - Shaleen Bhatnagar
- Department of Computer Science and Engineering, Manipal Institute of Technology Bengaluru, Manipal Academy of Higher Education, Manipal, India.
| | - Aadam Quraishi
- M. D. Research, Interventional Treatment Institute, Al Anbar, TX, United States.
| | | | | | - Faisal Yousef Alghayadh
- Computer Science and Information Systems Department, College of Applied Sciences, AlMaarefa University, Riyadh, Saudi Arabia.
| | - Shtwai Alsubai
- Department of Computer Science, College of Computer Engineering and Sciences in Al-Kharj, Prince Sattam bin Abdulaziz University, P.O. Box 151, Al-Kharj 11942, Saudi Arabia.
| | - Vaibhav Bhatnagar
- Department of Computer Applications, Manipal University Jaipur, India.
| |
Collapse
|
2
|
Li J, Wang L, Zhang Z, Feng Y, Huang M, Liang D. Analysis and recognition of a novel experimental paradigm for musical emotion brain-computer interface. Brain Res 2024; 1839:149039. [PMID: 38815645 DOI: 10.1016/j.brainres.2024.149039] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Revised: 05/17/2024] [Accepted: 05/27/2024] [Indexed: 06/01/2024]
Abstract
Musical emotions have received increasing attention over the years. To better recognize the emotions by brain-computer interface (BCI), the random music-playing and sequential music-playing experimental paradigms are proposed and compared in this paper. Two experimental paradigms consist of three positive pieces, three neutral pieces and three negative pieces of music. Ten subjects participate in two experimental paradigms. The features of electroencephalography (EEG) signals are firstly analyzed in the time, frequency and spatial domains. To improve the effect of emotion recognition, a recognition model is proposed with the optimal channels selecting by Pearson's correlation coefficient, and the feature fusion combining differential entropy and wavelet packet energy. According to the analysis results, the features of sequential music-playing experimental paradigm are more different among three emotions. The classification results of sequential music-playing experimental paradigm are also better, and its average results of positive, neutral and negative emotions are 78.53%, 72.81% and 77.35%, respectively. The more obvious the changes of EEG induced by the emotions, the higher the classification accuracy will be. After analyzing two experimental paradigms, a better way for music to induce the emotions can be explored. Therefore, our research offers a novel perspective on affective BCIs.
Collapse
Affiliation(s)
- Jin Li
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
| | - Li Wang
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China.
| | - Zhun Zhang
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
| | - Yujie Feng
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
| | - Mingyang Huang
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
| | - Danni Liang
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou 510006, China
| |
Collapse
|
3
|
Nikouei M, Abdali-Mohammadi F. A novel method for modeling effective connections between brain regions based on EEG signals and graph neural networks for motor imagery detection. Comput Methods Biomech Biomed Engin 2024; 27:1430-1447. [PMID: 37548428 DOI: 10.1080/10255842.2023.2244110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Revised: 06/07/2023] [Accepted: 07/28/2023] [Indexed: 08/08/2023]
Abstract
Classified as biomedical signal processing, cerebral signal processing plays a key role in human-computer interaction (HCI) and medical diagnosis. The motor imagery (MI) problem is an important research area in this field. Accurate solutions to this problem will greatly affect real-world applications. Most of the proposed methods are based on raw signal processing techniques. Known as prior knowledge, the structural-functional information and interregional connections can improve signal processing accuracy. It is possible to correctly perceive the generated signals by considering the brain structure (i.e. anatomical units), the source of signals, and the structural-functional dependence of different brain regions (i.e. effective connection) that are the semantic generators of signals. This study employed electroencephalograph (EEG) signals based on the activity of brain regions (cortex) and effective connections between brain regions based on dynamic causal modeling to solve the MI problem. EEG signals, as well as effective connections between brain regions to improve the interpretability of MI action, were fed into the architecture of Graph Convolutional Neural Network (GCN). The proposed model allowed GCN to extract more discriminative features. The results indicated that the proposed method was successful in developing a model with a MI detection accuracy of 93.73%.
Collapse
Affiliation(s)
- Mahya Nikouei
- Department of Computer Engineering and Information Technology, Razi University, Kermanshah, Iran
| | - Fardin Abdali-Mohammadi
- Department of Computer Engineering and Information Technology, Razi University, Kermanshah, Iran
| |
Collapse
|
4
|
Tong C, Ding Y, Zhang Z, Zhang H, JunLiang Lim K, Guan C. TASA: Temporal Attention With Spatial Autoencoder Network for Odor-Induced Emotion Classification Using EEG. IEEE Trans Neural Syst Rehabil Eng 2024; 32:1944-1954. [PMID: 38722724 DOI: 10.1109/tnsre.2024.3399326] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/21/2024]
Abstract
The olfactory system enables humans to smell different odors, which are closely related to emotions. The high temporal resolution and non-invasiveness of Electroencephalogram (EEG) make it suitable to objectively study human preferences for odors. Effectively learning the temporal dynamics and spatial information from EEG is crucial for detecting odor-induced emotional valence. In this paper, we propose a deep learning architecture called Temporal Attention with Spatial Autoencoder Network (TASA) for predicting odor-induced emotions using EEG. TASA consists of a filter-bank layer, a spatial encoder, a time segmentation layer, a Long Short-Term Memory (LSTM) module, a multi-head self-attention (MSA) layer, and a fully connected layer. We improve upon the previous work by utilizing a two-phase learning framework, using the autoencoder module to learn the spatial information among electrodes by reconstructing the given input with a latent representation in the spatial dimension, which aims to minimize information loss compared to spatial filtering with CNN. The second improvement is inspired by the continuous nature of the olfactory process; we propose to use LSTM-MSA in TASA to capture its temporal dynamics by learning the intercorrelation among the time segments of the EEG. TASA is evaluated on an existing olfactory EEG dataset and compared with several existing deep learning architectures to demonstrate its effectiveness in predicting olfactory-triggered emotional responses. Interpretability analyses with DeepLIFT also suggest that TASA learns spatial-spectral features that are relevant to olfactory-induced emotion recognition.
Collapse
|
5
|
Xu F, Pan D, Zheng H, Ouyang Y, Jia Z, Zeng H. EESCN: A novel spiking neural network method for EEG-based emotion recognition. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 243:107927. [PMID: 38000320 DOI: 10.1016/j.cmpb.2023.107927] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Revised: 10/16/2023] [Accepted: 11/13/2023] [Indexed: 11/26/2023]
Abstract
BACKGROUND AND OBJECTIVE Although existing artificial neural networks have achieved good results in electroencephalograph (EEG) emotion recognition, further improvements are needed in terms of bio-interpretability and robustness. In this research, we aim to develop a highly efficient and high-performance method for emotion recognition based on EEG. METHODS We propose an Emo-EEGSpikeConvNet (EESCN), a novel emotion recognition method based on spiking neural network (SNN). It consists of a neuromorphic data generation module and a NeuroSpiking framework. The neuromorphic data generation module converts EEG data into 2D frame format as input to the NeuroSpiking framework, while the NeuroSpiking framework is used to extract spatio-temporal features of EEG for classification. RESULTS EESCN achieves high emotion recognition accuracies on DEAP and SEED-IV datasets, ranging from 94.56% to 94.81% on DEAP and a mean accuracy of 79.65% on SEED-IV. Compared to existing SNN methods, EESCN significantly improves EEG emotion recognition performance. In addition, it also has the advantages of faster running speed and less memory footprint. CONCLUSIONS EESCN has shown excellent performance and efficiency in EEG-based emotion recognition with potential for practical applications requiring portability and resource constraints.
Collapse
Affiliation(s)
- FeiFan Xu
- Hangzhou Dianzi University, School of Computer Science and Technology, HangZhou, ZheJiang, China.
| | - Deng Pan
- Hangzhou Dianzi University, School of Computer Science and Technology, HangZhou, ZheJiang, China.
| | - Haohao Zheng
- Hangzhou Dianzi University, School of Computer Science and Technology, HangZhou, ZheJiang, China.
| | - Yu Ouyang
- Hangzhou Dianzi University, School of Computer Science and Technology, HangZhou, ZheJiang, China.
| | - Zhe Jia
- Hangzhou Dianzi University, School of Computer Science and Technology, HangZhou, ZheJiang, China.
| | - Hong Zeng
- Hangzhou Dianzi University, School of Computer Science and Technology, HangZhou, ZheJiang, China; Key Laboratory of Brain Machine Collaborative of Zhejiang Province, HangZhou, ZheJiang, China.
| |
Collapse
|
6
|
An Y, Hu S, Liu S, Li B. BiTCAN: A emotion recognition network based on saliency in brain cognition. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:21537-21562. [PMID: 38124609 DOI: 10.3934/mbe.2023953] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/23/2023]
Abstract
In recent years, with the continuous development of artificial intelligence and brain-computer interfaces, emotion recognition based on electroencephalogram (EEG) signals has become a prosperous research direction. Due to saliency in brain cognition, we construct a new spatio-temporal convolutional attention network for emotion recognition named BiTCAN. First, in the proposed method, the original EEG signals are de-baselined, and the two-dimensional mapping matrix sequence of EEG signals is constructed by combining the electrode position. Second, on the basis of the two-dimensional mapping matrix sequence, the features of saliency in brain cognition are extracted by using the Bi-hemisphere discrepancy module, and the spatio-temporal features of EEG signals are captured by using the 3-D convolution module. Finally, the saliency features and spatio-temporal features are fused into the attention module to further obtain the internal spatial relationships between brain regions, and which are input into the classifier for emotion recognition. Many experiments on DEAP and SEED (two public datasets) show that the accuracies of the proposed algorithm on both are higher than 97%, which is superior to most existing emotion recognition algorithms.
Collapse
Affiliation(s)
- Yanling An
- Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
| | - Shaohai Hu
- Institute of Information Science, Beijing Jiaotong University, Beijing 100044, China
| | - Shuaiqi Liu
- College of Electronic and Information Engineering, Hebei University, Baoding 071000, China
- Machine Vision Technology Innovation Center of Hebei Province, Baoding 071000, China
- The State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| | - Bing Li
- The State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| |
Collapse
|
7
|
Jiang H, Shen F, Chen L, Peng Y, Guo H, Gao H. Joint domain symmetry and predictive balance for cross-dataset EEG emotion recognition. J Neurosci Methods 2023; 400:109978. [PMID: 37806390 DOI: 10.1016/j.jneumeth.2023.109978] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 09/16/2023] [Accepted: 09/26/2023] [Indexed: 10/10/2023]
Abstract
BACKGROUND Cross-dataset EEG emotion recognition is an extremely challenging task, since data distributions of EEG from different datasets are greatly different, which makes the universal models yield unsatisfactory results. Although there are many methods have been proposed to reduce cross-dataset distribution discrepancies, they still neglected the following two problems. (1) Label space inconsistency: emotional label spaces of subjects from different datasets are different; (2) Uncertainty propagation: the uncertainty of misclassified emotion samples will propagate between datasets. NEW METHOD To solve these problems, we propose a novel method called domain symmetry and predictive balance (DSPB). For the problem of label space inconsistency, a domain symmetry module is designed to make label spaces of source and target domain to be the same, which randomly selects samples from the source domain and put into the target domain. For the problem of uncertainty propagation, a predictive balance module is proposed to reduce the prediction score of incorrect samples and then effectively reduce distribution differences between EEG from different datasets. RESULTS Experimental results show that our method achieve 61.48% average accuracies on the three cross-dataset tasks. Moreover, we find that gamma is the most relevant to emotion recognition among the five frequency bands, and the prefrontal and temporal brain regions are the channels carrying the most emotional information among the 62 brain channels. COMPARISON WITH EXISTING METHODS Compared with the partial domain adaptation method (SPDA) and the unsupervised domain adaptation (MS-MDA), our method improves average accuracies by 15.60% and 23.11%, respectively. CONCLUSION Besides, data distributions of EEG from different datasets but with the same emotional labels have been well aligned, which demonstrates the effectiveness of DSPB.
Collapse
Affiliation(s)
- Haiting Jiang
- College of Physics and Electronic Information Engineering, Zhejiang Normal University, Jin Hua, 321004, China
| | - Fangyao Shen
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jin Hua, 321004, China
| | - Lina Chen
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jin Hua, 321004, China.
| | - Yong Peng
- School of Computer Science and Technology, Hangzhou Dianzi University, Hang Zhou, 310018, China
| | - Hongjie Guo
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jin Hua, 321004, China
| | - Hong Gao
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jin Hua, 321004, China
| |
Collapse
|
8
|
Huynh VQ, Van Huynh T. Application of cluster repeated mini-batch training method to classify electroencephalography for grab and lift tasks. Med Eng Phys 2023; 120:104041. [PMID: 37838395 DOI: 10.1016/j.medengphy.2023.104041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2023] [Revised: 08/21/2023] [Accepted: 08/23/2023] [Indexed: 10/16/2023]
Abstract
Modern deep neural network training is based on mini-batch stochastic gradient optimization. While using extensive mini-batches improves the computational parallelism, the small batch training proved that it delivers improved generalization performance and allows a significantly smaller memory, which might also improve machine throughput. However, mini-batch size and characteristics, a key factor for training deep neural networks, has not been sufficiently investigated in training correlated group features and looping with highly complex ones. In addition, the unsupervised learning method clusters the data into different groups with similar properties to make the training process more stable and faster. Then, the supervised learning algorithm was applied with the cluster repeated mini-batch training (CRMT) methods. The CRMT algorithm changed the random minibatch characteristics in the training step into training in order of clusters. Specifically, the self-organizing maps (SOM) were used to cluster the information into n groups based on the dataset's labels Then, neural network models (ANN) were trained with each cluster using the cluster repeated mini-batch training method. Experiments conducted on EEG datasets demonstrate the survey of the proposed method and optimize it. In addition, the results in our research outperform other state-of-the-art methods.
Collapse
Affiliation(s)
- Viet Quoc Huynh
- Faculty of Physics and Engineering Physics, University of Science, Ho Chi Minh City, Vietnam; Vietnam National University, Ho Chi Minh City, Vietnam
| | - Tuan Van Huynh
- Faculty of Physics and Engineering Physics, University of Science, Ho Chi Minh City, Vietnam; Vietnam National University, Ho Chi Minh City, Vietnam.
| |
Collapse
|
9
|
Pan H, Yang H, Xie L, Wang Z. Multi-scale fusion visual attention network for facial micro-expression recognition. Front Neurosci 2023; 17:1216181. [PMID: 37575295 PMCID: PMC10412924 DOI: 10.3389/fnins.2023.1216181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Accepted: 06/26/2023] [Indexed: 08/15/2023] Open
Abstract
Introduction Micro-expressions are facial muscle movements that hide genuine emotions. In response to the challenge of micro-expression low-intensity, recent studies have attempted to locate localized areas of facial muscle movement. However, this ignores the feature redundancy caused by the inaccurate locating of the regions of interest. Methods This paper proposes a novel multi-scale fusion visual attention network (MFVAN), which learns multi-scale local attention weights to mask regions of redundancy features. Specifically, this model extracts the multi-scale features of the apex frame in the micro-expression video clips by convolutional neural networks. The attention mechanism focuses on the weights of local region features in the multi-scale feature maps. Then, we mask operate redundancy regions in multi-scale features and fuse local features with high attention weights for micro-expression recognition. The self-supervision and transfer learning reduce the influence of individual identity attributes and increase the robustness of multi-scale feature maps. Finally, the multi-scale classification loss, mask loss, and removing individual identity attributes loss joint to optimize the model. Results The proposed MFVAN method is evaluated on SMIC, CASME II, SAMM, and 3DB-Combined datasets that achieve state-of-the-art performance. The experimental results show that focusing on local at the multi-scale contributes to micro-expression recognition. Discussion This paper proposed MFVAN model is the first to combine image generation with visual attention mechanisms to solve the combination challenge problem of individual identity attribute interference and low-intensity facial muscle movements. Meanwhile, the MFVAN model reveal the impact of individual attributes on the localization of local ROIs. The experimental results show that a multi-scale fusion visual attention network contributes to micro-expression recognition.
Collapse
Affiliation(s)
- Hang Pan
- Department of Computer Science, Changzhi University, Changzhi, China
| | - Hongling Yang
- Department of Computer Science, Changzhi University, Changzhi, China
| | - Lun Xie
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, China
| | - Zhiliang Wang
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, China
| |
Collapse
|
10
|
Li JW, Lin D, Che Y, Lv JJ, Chen RJ, Wang LJ, Zeng XX, Ren JC, Zhao HM, Lu X. An innovative EEG-based emotion recognition using a single channel-specific feature from the brain rhythm code method. Front Neurosci 2023; 17:1221512. [PMID: 37547144 PMCID: PMC10397731 DOI: 10.3389/fnins.2023.1221512] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Accepted: 06/30/2023] [Indexed: 08/08/2023] Open
Abstract
Introduction Efficiently recognizing emotions is a critical pursuit in brain-computer interface (BCI), as it has many applications for intelligent healthcare services. In this work, an innovative approach inspired by the genetic code in bioinformatics, which utilizes brain rhythm code features consisting of δ, θ, α, β, or γ, is proposed for electroencephalography (EEG)-based emotion recognition. Methods These features are first extracted from the sequencing technique. After evaluating them using four conventional machine learning classifiers, an optimal channel-specific feature that produces the highest accuracy in each emotional case is identified, so emotion recognition through minimal data is realized. By doing so, the complexity of emotion recognition can be significantly reduced, making it more achievable for practical hardware setups. Results The best classification accuracies achieved for the DEAP and MAHNOB datasets range from 83-92%, and for the SEED dataset, it is 78%. The experimental results are impressive, considering the minimal data employed. Further investigation of the optimal features shows that their representative channels are primarily on the frontal region, and associated rhythmic characteristics are typical of multiple kinds. Additionally, individual differences are found, as the optimal feature varies with subjects. Discussion Compared to previous studies, this work provides insights into designing portable devices, as only one electrode is appropriate to generate satisfactory performances. Consequently, it would advance the understanding of brain rhythms, which offers an innovative solution for classifying EEG signals in diverse BCI applications, including emotion recognition.
Collapse
Affiliation(s)
- Jia Wen Li
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
- Engineering Research Center of Big Data Application in Private Health Medicine, Fujian Province University, Putian, China
- Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, Wuhan University of Science and Technology, Wuhan, China
- Guangxi Key Lab of Multi-Source Information Mining and Security, Guangxi Normal University, Guilin, China
| | - Di Lin
- Engineering Research Center of Big Data Application in Private Health Medicine, Fujian Province University, Putian, China
- New Engineering Industry College, Putian University, Putian, China
| | - Yan Che
- Engineering Research Center of Big Data Application in Private Health Medicine, Fujian Province University, Putian, China
- New Engineering Industry College, Putian University, Putian, China
| | - Ju Jian Lv
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Rong Jun Chen
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Lei Jun Wang
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Xian Xian Zeng
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Jin Chang Ren
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
- National Subsea Centre, Robert Gordon University, Aberdeen, United Kingdom
| | - Hui Min Zhao
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Xu Lu
- School of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| |
Collapse
|
11
|
Wang X, Xu B, Zhang W, Wang J, Deng L, Ping J, Hu C, Li H. Recognizing emotions induced by wearable haptic vibration using noninvasive electroencephalogram. Front Neurosci 2023; 17:1219553. [PMID: 37483356 PMCID: PMC10357513 DOI: 10.3389/fnins.2023.1219553] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Accepted: 06/20/2023] [Indexed: 07/25/2023] Open
Abstract
The integration of haptic technology into affective computing has led to a new field known as affective haptics. Nonetheless, the mechanism underlying the interaction between haptics and emotions remains unclear. In this paper, we proposed a novel haptic pattern with adaptive vibration intensity and rhythm according to the volume, and applied it into the emotional experiment paradigm. To verify its superiority, the proposed haptic pattern was compared with an existing haptic pattern by combining them with conventional visual-auditory stimuli to induce emotions (joy, sadness, fear, and neutral), and the subjects' EEG signals were collected simultaneously. The features of power spectral density (PSD), differential entropy (DE), differential asymmetry (DASM), and differential caudality (DCAU) were extracted, and the support vector machine (SVM) was utilized to recognize four target emotions. The results demonstrated that haptic stimuli enhanced the activity of the lateral temporal and prefrontal areas of the emotion-related brain regions. Moreover, the classification accuracy of the existing constant haptic pattern and the proposed adaptive haptic pattern increased by 7.71 and 8.60%, respectively. These findings indicate that flexible and varied haptic patterns can enhance immersion and fully stimulate target emotions, which are of great importance for wearable haptic interfaces and emotion communication through haptics.
Collapse
Affiliation(s)
- Xin Wang
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Baoguo Xu
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Wenbin Zhang
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Jiajin Wang
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Leying Deng
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Jingyu Ping
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| | - Cong Hu
- Guangxi Key Laboratory of Automatic Detecting Technology and Instruments, Guilin University of Electronic Technology, Guilin, China
| | - Huijun Li
- The State Key Laboratory of Digital Medical Engineering, Jiangsu Key Laboratory of Remote Measurement and Control, School of Instrument Science and Engineering, Southeast University, Nanjing, China
| |
Collapse
|
12
|
Zhang Y, Peng Y, Li J, Kong W. SIFIAE: An adaptive emotion recognition model with EEG feature-label inconsistency consideration. J Neurosci Methods 2023:109909. [PMID: 37399992 DOI: 10.1016/j.jneumeth.2023.109909] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Revised: 06/06/2023] [Accepted: 06/26/2023] [Indexed: 07/05/2023]
Abstract
BACKGROUND A common but easily overlooked affective overlap problem has not been received enough attention in electroencephalogram (EEG)-based emotion recognition research. In real life, affective overlap refers to the current emotional state of human being is sometimes influenced easily by his/her historical mood. In stimulus-evoked EEG collection experiment, due to the short rest interval in consecutive trials, the inner mechanisms of neural responses make subjects cannot switch their emotion state easily and quickly, which might lead to the affective overlap. For example, we might be still in sad state to some extent even if we are watching a comedy because we just saw a tragedy before. In pattern recognition, affective overlap usually means that there exists the feature-label inconsistency in EEG data. NEW METHODS To alleviate the impact of inconsistent EEG data, we introduce a variable to adaptively explore the sample inconsistency in emotion recognition model development. Then, we propose a semi-supervised emotion recognition model for joint sample inconsistency and feature importance exploration (SIFIAE). Accordingly, an efficient optimization method to SIFIAE model is proposed. RESULTS Extensive experiments on the SEED-V dataset demonstrate the effectiveness of SIFIAE. Specifically, SIFIAE achieves 69.10%, 67.01%, 71.5%, 73.26%, 72.07% and 71.35% average accuracies in six cross-session emotion recognition tasks. CONCLUSION The results illustrated that the sample weights have a rising trend in the beginning of most trials, which coincides with the affective overlap hypothesis. The feature importance factor indicated the critical bands and channels are more obvious compared with some models without considering EEG feature-label inconsistency.
Collapse
Affiliation(s)
- Yikai Zhang
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang Province, China
| | - Yong Peng
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang Province, China; Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, 310018, Zhejiang Province, China.
| | - Junhua Li
- School of Computer Science and Electronic Engineering, University of Essex, Colchester, CO4 3SQ, UK
| | - Wanzeng Kong
- School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou, 310018, Zhejiang Province, China; Key Laboratory of Brain Machine Collaborative Intelligence of Zhejiang Province, Hangzhou, 310018, Zhejiang Province, China
| |
Collapse
|
13
|
Li F, Luo J, Wang L, Liu W, Sang X. GCF 2-Net: global-aware cross-modal feature fusion network for speech emotion recognition. Front Neurosci 2023; 17:1183132. [PMID: 37214410 PMCID: PMC10192703 DOI: 10.3389/fnins.2023.1183132] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Accepted: 04/13/2023] [Indexed: 05/24/2023] Open
Abstract
Emotion recognition plays an essential role in interpersonal communication. However, existing recognition systems use only features of a single modality for emotion recognition, ignoring the interaction of information from the different modalities. Therefore, in our study, we propose a global-aware Cross-modal feature Fusion Network (GCF2-Net) for recognizing emotion. We construct a residual cross-modal fusion attention module (ResCMFA) to fuse information from multiple modalities and design a global-aware module to capture global details. More specifically, we first use transfer learning to extract wav2vec 2.0 features and text features fused by the ResCMFA module. Then, cross-modal fusion features are fed into the global-aware module to capture the most essential emotional information globally. Finally, the experiment results have shown that our proposed method has significant advantages than state-of-the-art methods on the IEMOCAP and MELD datasets, respectively.
Collapse
Affiliation(s)
- Feng Li
- Department of Computer Science and Technology, Anhui University of Finance and Economics, Anhui, China
- School of Information Science and Technology, University of Science and Technology of China, Anhui, China
| | - Jiusong Luo
- Department of Computer Science and Technology, Anhui University of Finance and Economics, Anhui, China
| | - Lingling Wang
- Department of Computer Science and Technology, Anhui University of Finance and Economics, Anhui, China
| | - Wei Liu
- Department of Computer Science and Technology, Anhui University of Finance and Economics, Anhui, China
| | - Xiaoshuang Sang
- Department of Computer Science and Technology, Anhui University of Finance and Economics, Anhui, China
| |
Collapse
|
14
|
Skaramagkas V, Ktistakis E, Manousos D, Kazantzaki E, Tachos NS, Tripoliti E, Fotiadis DI, Tsiknakis M. eSEE-d: Emotional State Estimation Based on Eye-Tracking Dataset. Brain Sci 2023; 13:brainsci13040589. [PMID: 37190554 DOI: 10.3390/brainsci13040589] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Revised: 03/22/2023] [Accepted: 03/29/2023] [Indexed: 04/03/2023] Open
Abstract
Affective state estimation is a research field that has gained increased attention from the research community in the last decade. Two of the main catalysts for this are the advancement in the data analysis using artificial intelligence and the availability of high-quality video. Unfortunately, benchmarks and public datasets are limited, thus making the development of new methodologies and the implementation of comparative studies essential. The current work presents the eSEE-d database, which is a resource to be used for emotional State Estimation based on Eye-tracking data. Eye movements of 48 participants were recorded as they watched 10 emotion-evoking videos, each of them followed by a neutral video. Participants rated four emotions (tenderness, anger, disgust, sadness) on a scale from 0 to 10, which was later translated in terms of emotional arousal and valence levels. Furthermore, each participant filled three self-assessment questionnaires. An extensive analysis of the participants’ answers to the questionnaires’ self-assessment scores as well as their ratings during the experiments is presented. Moreover, eye and gaze features were extracted from the low-level eye-recorded metrics, and their correlations with the participants’ ratings are investigated. Finally, we take on the challenge to classify arousal and valence levels based solely on eye and gaze features, leading to promising results. In particular, the Deep Multilayer Perceptron (DMLP) network we developed achieved an accuracy of 92% in distinguishing positive valence from non-positive and 81% in distinguishing low arousal from medium arousal. The dataset is made publicly available.
Collapse
|