1
|
Yu H, Xiong X, Zhou J, Qian R, Sha K. CATM: A Multi-Feature-Based Cross-Scale Attentional Convolutional EEG Emotion Recognition Model. SENSORS (BASEL, SWITZERLAND) 2024; 24:4837. [PMID: 39123882 PMCID: PMC11314657 DOI: 10.3390/s24154837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/24/2024] [Revised: 07/21/2024] [Accepted: 07/23/2024] [Indexed: 08/12/2024]
Abstract
Aiming at the problem that existing emotion recognition methods fail to make full use of the information in the time, frequency, and spatial domains in the EEG signals, which leads to the low accuracy of EEG emotion classification, this paper proposes a multi-feature, multi-frequency band-based cross-scale attention convolutional model (CATM). The model is mainly composed of a cross-scale attention module, a frequency-space attention module, a feature transition module, a temporal feature extraction module, and a depth classification module. First, the cross-scale attentional convolution module extracts spatial features at different scales for the preprocessed EEG signals; then, the frequency-space attention module assigns higher weights to important channels and spatial locations; next, the temporal feature extraction module extracts temporal features of the EEG signals; and, finally, the depth classification module categorizes the EEG signals into emotions. We evaluated the proposed method on the DEAP dataset with accuracies of 99.70% and 99.74% in the valence and arousal binary classification experiments, respectively; the accuracy in the valence-arousal four-classification experiment was 97.27%. In addition, considering the application of fewer channels, we also conducted 5-channel experiments, and the binary classification accuracies of valence and arousal were 97.96% and 98.11%, respectively. The valence-arousal four-classification accuracy was 92.86%. The experimental results show that the method proposed in this paper exhibits better results compared to other recent methods, and also achieves better results in few-channel experiments.
Collapse
Affiliation(s)
| | | | - Jianhua Zhou
- Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China; (H.Y.); (X.X.); (R.Q.); (K.S.)
| | | | | |
Collapse
|
2
|
李 湘, 王 丹, 张 柏, 范 超, 陈 佳, 许 萌, 陈 远. [A review on electroencephalogram based channel selection]. SHENG WU YI XUE GONG CHENG XUE ZA ZHI = JOURNAL OF BIOMEDICAL ENGINEERING = SHENGWU YIXUE GONGCHENGXUE ZAZHI 2024; 41:398-405. [PMID: 38686423 PMCID: PMC11058489 DOI: 10.7507/1001-5515.202308034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 08/20/2023] [Revised: 02/08/2024] [Indexed: 05/02/2024]
Abstract
The electroencephalogram (EEG) signal is the key signal carrier of the brain-computer interface (BCI) system. The EEG data collected by the whole-brain electrode arrangement is conducive to obtaining higher information representation. Personalized electrode layout, while ensuring the accuracy of EEG signal decoding, can also shorten the calibration time of BCI and has become an important research direction. This paper reviews the EEG signal channel selection methods in recent years, conducts a comparative analysis of the combined effects of different channel selection methods and different classification algorithms, obtains the commonly used channel combinations in motor imagery, P300 and other paradigms in BCI, and explains the application scenarios of the channel selection method in different paradigms are discussed, in order to provide stronger support for a more accurate and portable BCI system.
Collapse
Affiliation(s)
- 湘喆 李
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| | - 丹 王
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| | - 柏雯 张
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| | - 超杰 范
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| | - 佳明 陈
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| | - 萌 许
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| | - 远方 陈
- 北京工业大学 信息学部(北京 100124)Faculty of information Technology, Beijing University of Technology, Beijing 100124, P.R. China
| |
Collapse
|
3
|
Davis JJJ, Schübeler F, Kozma R. Information-Theoretical Analysis of the Cycle of Creation of Knowledge and Meaning in Brains under Multiple Cognitive Modalities. SENSORS (BASEL, SWITZERLAND) 2024; 24:1605. [PMID: 38475141 DOI: 10.3390/s24051605] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/25/2024] [Revised: 02/21/2024] [Accepted: 02/27/2024] [Indexed: 03/14/2024]
Abstract
It is of great interest to develop advanced sensory technologies allowing non-invasive monitoring of neural correlates of cognitive processing in people performing everyday tasks. A lot of progress has been reported in recent years in this research area using scalp EEG arrays, but the high level of noise in the electrode signals poses a lot of challenges. This study presents results of detailed statistical analysis of experimental data on the cycle of creation of knowledge and meaning in human brains under multiple cognitive modalities. We measure brain dynamics using a HydroCel Geodesic Sensor Net, 128-electrode dense-array electroencephalography (EEG). We compute a pragmatic information (PI) index derived from analytic amplitude and phase, by Hilbert transforming the EEG signals of 20 participants in six modalities, which combine various audiovisual stimuli, leading to different mental states, including relaxed and cognitively engaged conditions. We derive several relevant measures to classify different brain states based on the PI indices. We demonstrate significant differences between engaged brain states that require sensory information processing to create meaning and knowledge for intentional action, and relaxed-meditative brain states with less demand on psychophysiological resources. We also point out that different kinds of meanings may lead to different brain dynamics and behavioral responses.
Collapse
Affiliation(s)
- Joshua J J Davis
- Dodd-Walls Centre for Photonics and Quantum Technologies, Department of Physics & Ian Kirk's Lab., Centre for Brain Research, The University of Auckland, Auckland 1142, New Zealand
| | | | - Robert Kozma
- Department of Mathematics, University of Memphis, Memphis, TN 38152, USA
- School of Informatics, Obuda University, H-1034 Budapest, Hungary
- Kozmos Research Laboratories, Boston, MA 02215, USA
| |
Collapse
|
4
|
Jafari M, Shoeibi A, Khodatars M, Bagherzadeh S, Shalbaf A, García DL, Gorriz JM, Acharya UR. Emotion recognition in EEG signals using deep learning methods: A review. Comput Biol Med 2023; 165:107450. [PMID: 37708717 DOI: 10.1016/j.compbiomed.2023.107450] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 08/03/2023] [Accepted: 09/01/2023] [Indexed: 09/16/2023]
Abstract
Emotions are a critical aspect of daily life and serve a crucial role in human decision-making, planning, reasoning, and other mental states. As a result, they are considered a significant factor in human interactions. Human emotions can be identified through various sources, such as facial expressions, speech, behavior (gesture/position), or physiological signals. The use of physiological signals can enhance the objectivity and reliability of emotion detection. Compared with peripheral physiological signals, electroencephalogram (EEG) recordings are directly generated by the central nervous system and are closely related to human emotions. EEG signals have the great spatial resolution that facilitates the evaluation of brain functions, making them a popular modality in emotion recognition studies. Emotion recognition using EEG signals presents several challenges, including signal variability due to electrode positioning, individual differences in signal morphology, and lack of a universal standard for EEG signal processing. Moreover, identifying the appropriate features for emotion recognition from EEG data requires further research. Finally, there is a need to develop more robust artificial intelligence (AI) including conventional machine learning (ML) and deep learning (DL) methods to handle the complex and diverse EEG signals associated with emotional states. This paper examines the application of DL techniques in emotion recognition from EEG signals and provides a detailed discussion of relevant articles. The paper explores the significant challenges in emotion recognition using EEG signals, highlights the potential of DL techniques in addressing these challenges, and suggests the scope for future research in emotion recognition using DL techniques. The paper concludes with a summary of its findings.
Collapse
Affiliation(s)
- Mahboobeh Jafari
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Afshin Shoeibi
- Data Science and Computational Intelligence Institute, University of Granada, Spain.
| | - Marjane Khodatars
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Sara Bagherzadeh
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Ahmad Shalbaf
- Department of Biomedical Engineering and Medical Physics, School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - David López García
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Juan M Gorriz
- Data Science and Computational Intelligence Institute, University of Granada, Spain; Department of Psychiatry, University of Cambridge, UK
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
5
|
Peng G, Zhao K, Zhang H, Xu D, Kong X. Temporal relative transformer encoding cooperating with channel attention for EEG emotion analysis. Comput Biol Med 2023; 154:106537. [PMID: 36682180 DOI: 10.1016/j.compbiomed.2023.106537] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2022] [Revised: 12/18/2022] [Accepted: 01/10/2023] [Indexed: 01/18/2023]
Abstract
Electroencephalogram (EEG)-based emotion computing has become a hot topic of brain-computer fusion. EEG signals have inherent temporal and spatial characteristics. However, existing studies did not fully consider the two properties. In addition, the position encoding mechanism in the vanilla transformer cannot effectively encode the continuous temporal character of the emotion. A temporal relative (TR) encoding mechanism is proposed to encode the temporal EEG signals for constructing the temporality self-attention in the transformer. To explore the contribution of each EEG channel corresponding to the electrode on the cerebral cortex to emotion analysis, a channel-attention (CA) mechanism is presented. The temporality self-attention mechanism cooperates with the channel-attention mechanism to utilize the temporal and spatial information of EEG signals simultaneously by preprocessing. Exhaustive experiments are conducted on the DEAP dataset, including the binary classification on valence, arousal, dominance, and liking. Furthermore, the discrete emotion category classification task is also conducted by mapping the dimensional annotations of DEAP into discrete emotion categories (5-class). Experimental results demonstrate that our model outperforms the advanced methods for all classification tasks.
Collapse
Affiliation(s)
- Guoqin Peng
- Yunnan University, Kunming, 650500, China; Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, 310013, China; Department of Psychiatry of Sir Run Run Shaw Hospital, Zhejiang University School of Medicine, Hangzhou, 310013, China.
| | | | - Hao Zhang
- Yunnan University, Kunming, 650500, China
| | - Dan Xu
- Yunnan University, Kunming, 650500, China.
| | - Xiangzhen Kong
- Department of Psychology and Behavioral Sciences, Zhejiang University, Hangzhou, 310013, China; Department of Psychiatry of Sir Run Run Shaw Hospital, Zhejiang University School of Medicine, Hangzhou, 310013, China.
| |
Collapse
|
6
|
Electroencephalography based emotion detection using ensemble classification and asymmetric brain activity. J Affect Disord 2022; 319:416-427. [PMID: 36162677 DOI: 10.1016/j.jad.2022.09.054] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/09/2022] [Revised: 08/07/2022] [Accepted: 09/20/2022] [Indexed: 11/22/2022]
Abstract
Over the past decade, emotion detection using rhythmic brain activity has become a critical area of research. The asymmetrical brain activity has garnered the most significant level of research attention due to its implications for the study of emotions, including hemispheric asymmetry or, more generally, asymmetrical brain activity. This study aimed at enhancing the accuracy of emotion detection using Electroencephalography (EEG) brain signals. This happens by identifying electrodes where relevant brain activity changes occur during the emotions and by defining pairs of relevant electrodes having asymmetric brain activities during emotions. Experimental results showed that the proposed method is highly competitive compared with existing studies of multi-class emotion recognition. These results were improved by processing not the whole EEG signals but by focusing on fragments of the signals, called epochs, which represent the instants where the excitation is maximum during emotions. The epochs were extracted using the zero-time windowing method and the numerator group-delay function.
Collapse
|
7
|
Nalwaya A, Das K, Pachori RB. Automated Emotion Identification Using Fourier-Bessel Domain-Based Entropies. ENTROPY (BASEL, SWITZERLAND) 2022; 24:1322. [PMID: 37420342 DOI: 10.3390/e24101322] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/07/2022] [Revised: 09/09/2022] [Accepted: 09/16/2022] [Indexed: 07/09/2023]
Abstract
Human dependence on computers is increasing day by day; thus, human interaction with computers must be more dynamic and contextual rather than static or generalized. The development of such devices requires knowledge of the emotional state of the user interacting with it; for this purpose, an emotion recognition system is required. Physiological signals, specifically, electrocardiogram (ECG) and electroencephalogram (EEG), were studied here for the purpose of emotion recognition. This paper proposes novel entropy-based features in the Fourier-Bessel domain instead of the Fourier domain, where frequency resolution is twice that of the latter. Further, to represent such non-stationary signals, the Fourier-Bessel series expansion (FBSE) is used, which has non-stationary basis functions, making it more suitable than the Fourier representation. EEG and ECG signals are decomposed into narrow-band modes using FBSE-based empirical wavelet transform (FBSE-EWT). The proposed entropies of each mode are computed to form the feature vector, which are further used to develop machine learning models. The proposed emotion detection algorithm is evaluated using publicly available DREAMER dataset. K-nearest neighbors (KNN) classifier provides accuracies of 97.84%, 97.91%, and 97.86% for arousal, valence, and dominance classes, respectively. Finally, this paper concludes that the obtained entropy features are suitable for emotion recognition from given physiological signals.
Collapse
Affiliation(s)
- Aditya Nalwaya
- Department of Electrical Engineering, Indian Institute of Technology Indore, Indore 453552, India
| | - Kritiprasanna Das
- Department of Electrical Engineering, Indian Institute of Technology Indore, Indore 453552, India
| | - Ram Bilas Pachori
- Department of Electrical Engineering, Indian Institute of Technology Indore, Indore 453552, India
| |
Collapse
|
8
|
Khoeun R, Chophuk P, Chinnasarn K. Emotion Recognition for Partial Faces Using a Feature Vector Technique. SENSORS (BASEL, SWITZERLAND) 2022; 22:4633. [PMID: 35746415 PMCID: PMC9230947 DOI: 10.3390/s22124633] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Revised: 06/10/2022] [Accepted: 06/17/2022] [Indexed: 06/15/2023]
Abstract
Wearing a facial mask is indispensable in the COVID-19 pandemic; however, it has tremendous effects on the performance of existing facial emotion recognition approaches. In this paper, we propose a feature vector technique comprising three main steps to recognize emotions from facial mask images. First, a synthetic mask is used to cover the facial input image. With only the upper part of the image showing, and including only the eyes, eyebrows, a portion of the bridge of the nose, and the forehead, the boundary and regional representation technique is applied. Second, a feature extraction technique based on our proposed rapid landmark detection method employing the infinity shape is utilized to flexibly extract a set of feature vectors that can effectively indicate the characteristics of the partially occluded masked face. Finally, those features, including the location of the detected landmarks and the Histograms of the Oriented Gradients, are brought into the classification process by adopting CNN and LSTM; the experimental results are then evaluated using images from the CK+ and RAF-DB data sets. As the result, our proposed method outperforms existing cutting-edge approaches and demonstrates better performance, achieving 99.30% and 95.58% accuracy on CK+ and RAF-DB, respectively.
Collapse
|