1
|
Narimisaei J, Naeim M, Imannezhad S, Samian P, Sobhani M. Exploring emotional intelligence in artificial intelligence systems: a comprehensive analysis of emotion recognition and response mechanisms. Ann Med Surg (Lond) 2024; 86:4657-4663. [PMID: 39118764 PMCID: PMC11305735 DOI: 10.1097/ms9.0000000000002315] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2024] [Accepted: 06/17/2024] [Indexed: 08/10/2024] Open
Abstract
This study aims to dissect the current state of emotion recognition and response mechanisms in artificial intelligence (AI) systems, exploring the progress made, challenges faced, and implicit operations of integrating emotional intelligence into AI. This study utilized a comprehensive review approach to investigate the integration of emotional intelligence (EI) into artificial intelligence (AI) systems, concentrating on emotion recognition and response mechanisms. The review process entailed formulating research questions, systematically searching academic databases such as PubMed, Scopus, and Web of Science, critically evaluating relevant literature, synthesizing the data, and presenting the findings in a comprehensive format. The study highlights the advancements in emotion recognition models, including the use of deep literacy ways and multimodal data emulsion. It discusses the challenges in emotion recognition, similar to variability in mortal expressions and the need for real-time processing. The integration of contextual information and individual traits is emphasized as enhancing the understanding of mortal feelings. The study also addresses ethical enterprises, similar as sequestration and impulses in training data. The integration of emotional intelligence into AI systems presents openings to revise mortal-computer relations. Emotion recognition and response mechanisms have made significant progress, but challenges remain. Unborn exploration directions include enhancing the robustness and interpretability of emotion recognition models, exploring cross-cultural and environment-apprehensive emotion understanding, and addressing long-term emotion shadowing and adaption. By further exploring emotional intelligence in AI systems, further compassionate and responsive machines can be developed, enabling deeper emotional connections with humans.
Collapse
Affiliation(s)
- Jale Narimisaei
- Department of Computer, Energy and Data Science Faculty, Behbahan Khatam Alanbia University of Technology, Behbahan
| | - Mahdi Naeim
- Department of Research, Psychology and Counseling Organization, Tehran
| | - Shima Imannezhad
- Department of Pediatrics, Faculty of Medicine, Mashhad University of Medical Sciences, Mashhad
| | - Pooya Samian
- Department of Educational Sciences, Faculty of Educational Sciences and Psychology, Shahid Madani University of Azerbaijan, Tabriz, Iran
| | | |
Collapse
|
2
|
Hamzah HA, Abdalla KK. EEG-based emotion recognition systems; comprehensive study. Heliyon 2024; 10:e31485. [PMID: 38818173 PMCID: PMC11137547 DOI: 10.1016/j.heliyon.2024.e31485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 05/16/2024] [Indexed: 06/01/2024] Open
Abstract
Emotion recognition technology through EEG signal analysis is currently a fundamental concept in artificial intelligence. This recognition has major practical implications in emotional health care, human-computer interaction, and so on. This paper provides a comprehensive study of different methods for extracting electroencephalography (EEG) features for emotion recognition from four different perspectives, including time domain features, frequency domain features, time-frequency features, and nonlinear features. We summarize the current pattern recognition methods adopted in most related works, and with the rapid development of deep learning (DL) attracting the attention of researchers in this field, we pay more attention to deep learning-based studies and analyse the characteristics, advantages, disadvantages, and applicable scenarios. Finally, the current challenges and future development directions in this field were summarized. This paper can help novice researchers in this field gain a systematic understanding of the current status of emotion recognition research based on EEG signals and provide ideas for subsequent related research.
Collapse
Affiliation(s)
- Hussein Ali Hamzah
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| | - Kasim K. Abdalla
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| |
Collapse
|
3
|
Qiao Y, Mu J, Xie J, Hu B, Liu G. Music emotion recognition based on temporal convolutional attention network using EEG. Front Hum Neurosci 2024; 18:1324897. [PMID: 38617132 PMCID: PMC11010638 DOI: 10.3389/fnhum.2024.1324897] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Accepted: 03/08/2024] [Indexed: 04/16/2024] Open
Abstract
Music is one of the primary ways to evoke human emotions. However, the feeling of music is subjective, making it difficult to determine which emotions music triggers in a given individual. In order to correctly identify emotional problems caused by different types of music, we first created an electroencephalogram (EEG) data set stimulated by four different types of music (fear, happiness, calm, and sadness). Secondly, the differential entropy features of EEG were extracted, and then the emotion recognition model CNN-SA-BiLSTM was established to extract the temporal features of EEG, and the recognition performance of the model was improved by using the global perception ability of the self-attention mechanism. The effectiveness of the model was further verified by the ablation experiment. The classification accuracy of this method in the valence and arousal dimensions is 93.45% and 96.36%, respectively. By applying our method to a publicly available EEG dataset DEAP, we evaluated the generalization and reliability of our method. In addition, we further investigate the effects of different EEG bands and multi-band combinations on music emotion recognition, and the results confirm relevant neuroscience studies. Compared with other representative music emotion recognition works, this method has better classification performance, and provides a promising framework for the future research of emotion recognition system based on brain computer interface.
Collapse
Affiliation(s)
- Yinghao Qiao
- School of Electronic and Information Engineering, Southwest University, Chongqing, China
- Institute of Affective Computing and Information Processing, Southwest University, Chongqing, China
- Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, Chongqing, China
| | - Jiajia Mu
- School of Electronic and Information Engineering, Southwest University, Chongqing, China
- Institute of Affective Computing and Information Processing, Southwest University, Chongqing, China
- Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, Chongqing, China
| | - Jialan Xie
- School of Electronic and Information Engineering, Southwest University, Chongqing, China
- Institute of Affective Computing and Information Processing, Southwest University, Chongqing, China
- Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, Chongqing, China
| | - Binghui Hu
- School of Electronic and Information Engineering, Southwest University, Chongqing, China
- Institute of Affective Computing and Information Processing, Southwest University, Chongqing, China
- Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, Chongqing, China
| | - Guangyuan Liu
- School of Electronic and Information Engineering, Southwest University, Chongqing, China
- Institute of Affective Computing and Information Processing, Southwest University, Chongqing, China
- Chongqing Key Laboratory of Nonlinear Circuits and Intelligent Information Processing, Southwest University, Chongqing, China
| |
Collapse
|
4
|
Yang L, Wang Z, Wang G, Liang L, Liu M, Wang J. Brain-inspired modular echo state network for EEG-based emotion recognition. Front Neurosci 2024; 18:1305284. [PMID: 38495107 PMCID: PMC10940514 DOI: 10.3389/fnins.2024.1305284] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2023] [Accepted: 01/10/2024] [Indexed: 03/19/2024] Open
Abstract
Previous studies have successfully applied a lightweight recurrent neural network (RNN) called Echo State Network (ESN) for EEG-based emotion recognition. These studies use intrinsic plasticity (IP) and synaptic plasticity (SP) to tune the hidden reservoir layer of ESN, yet they require extra training procedures and are often computationally complex. Recent neuroscientific research reveals that the brain is modular, consisting of internally dense and externally sparse subnetworks. Furthermore, it has been proved that this modular topology facilitates information processing efficiency in both biological and artificial neural networks (ANNs). Motivated by these findings, we propose Modular Echo State Network (M-ESN), where the hidden layer of ESN is directly initialized to a more efficient modular structure. In this paper, we first describe our novel implementation method, which enables us to find the optimal module numbers, local and global connectivity. Then, the M-ESN is benchmarked on the DEAP dataset. Lastly, we explain why network modularity improves model performance. We demonstrate that modular organization leads to a more diverse distribution of node degrees, which increases network heterogeneity and subsequently improves classification accuracy. On the emotion arousal, valence, and stress/calm classification tasks, our M-ESN outperforms regular ESN by 5.44, 5.90, and 5.42%, respectively, while this difference when comparing with adaptation rules tuned ESNs are 0.77, 5.49, and 0.95%. Notably, our results are obtained using M-ESN with a much smaller reservoir size and simpler training process.
Collapse
Affiliation(s)
- Liuyi Yang
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Zhaoze Wang
- School of Engineering and Applied Science, University of Pennsylvania, Pennsylvania, PA, United States
| | - Guoyu Wang
- Department of Auromation, Tiangong University, Tianjin, China
| | - Lixin Liang
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Meng Liu
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| | - Junsong Wang
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, China
| |
Collapse
|
5
|
Liu R, Chao Y, Ma X, Sha X, Sun L, Li S, Chang S. ERTNet: an interpretable transformer-based framework for EEG emotion recognition. Front Neurosci 2024; 18:1320645. [PMID: 38298914 PMCID: PMC10827927 DOI: 10.3389/fnins.2024.1320645] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2023] [Accepted: 01/02/2024] [Indexed: 02/02/2024] Open
Abstract
Background Emotion recognition using EEG signals enables clinicians to assess patients' emotional states with precision and immediacy. However, the complexity of EEG signal data poses challenges for traditional recognition methods. Deep learning techniques effectively capture the nuanced emotional cues within these signals by leveraging extensive data. Nonetheless, most deep learning techniques lack interpretability while maintaining accuracy. Methods We developed an interpretable end-to-end EEG emotion recognition framework rooted in the hybrid CNN and transformer architecture. Specifically, temporal convolution isolates salient information from EEG signals while filtering out potential high-frequency noise. Spatial convolution discerns the topological connections between channels. Subsequently, the transformer module processes the feature maps to integrate high-level spatiotemporal features, enabling the identification of the prevailing emotional state. Results Experiments' results demonstrated that our model excels in diverse emotion classification, achieving an accuracy of 74.23% ± 2.59% on the dimensional model (DEAP) and 67.17% ± 1.70% on the discrete model (SEED-V). These results surpass the performances of both CNN and LSTM-based counterparts. Through interpretive analysis, we ascertained that the beta and gamma bands in the EEG signals exert the most significant impact on emotion recognition performance. Notably, our model can independently tailor a Gaussian-like convolution kernel, effectively filtering high-frequency noise from the input EEG data. Discussion Given its robust performance and interpretative capabilities, our proposed framework is a promising tool for EEG-driven emotion brain-computer interface.
Collapse
Affiliation(s)
- Ruixiang Liu
- School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| | - Yihu Chao
- School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| | - Xuerui Ma
- School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| | - Xianzheng Sha
- School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| | - Limin Sun
- Shanghai Institute of Microsystem and Information Technology, Chinese Academy of Sciences, Shanghai, China
| | - Shuo Li
- School of Life Sciences, China Medical University, Shenyang, Liaoning, China
| | - Shijie Chang
- School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| |
Collapse
|
6
|
Wang X, Liesaputra V, Liu Z, Wang Y, Huang Z. An in-depth survey on Deep Learning-based Motor Imagery Electroencephalogram (EEG) classification. Artif Intell Med 2024; 147:102738. [PMID: 38184362 DOI: 10.1016/j.artmed.2023.102738] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2022] [Revised: 10/16/2023] [Accepted: 11/28/2023] [Indexed: 01/08/2024]
Abstract
Electroencephalogram (EEG)-based Brain-Computer Interfaces (BCIs) build a communication path between human brain and external devices. Among EEG-based BCI paradigms, the most commonly used one is motor imagery (MI). As a hot research topic, MI EEG-based BCI has largely contributed to medical fields and smart home industry. However, because of the low signal-to-noise ratio (SNR) and the non-stationary characteristic of EEG data, it is difficult to correctly classify different types of MI-EEG signals. Recently, the advances in Deep Learning (DL) significantly facilitate the development of MI EEG-based BCIs. In this paper, we provide a systematic survey of DL-based MI-EEG classification methods. Specifically, we first comprehensively discuss several important aspects of DL-based MI-EEG classification, covering input formulations, network architectures, public datasets, etc. Then, we summarize problems in model performance comparison and give guidelines to future studies for fair performance comparison. Next, we fairly evaluate the representative DL-based models using source code released by the authors and meticulously analyse the evaluation results. By performing ablation study on the network architecture, we found that (1) effective feature fusion is indispensable for multi-stream CNN-based models. (2) LSTM should be combined with spatial feature extraction techniques to obtain good classification performance. (3) the use of dropout contributes little to improving the model performance, and that (4) adding fully connected layers to the models significantly increases their parameters but it might not improve their performance. Finally, we raise several open issues in MI-EEG classification and provide possible future research directions.
Collapse
Affiliation(s)
- Xianheng Wang
- Department of Computer Science, University of Otago, Dunedin, New Zealand.
| | | | - Zhaobin Liu
- College of Information Science and Technology, Dalian Maritime University, Liaoning, PR China
| | - Yi Wang
- Laboratory for Circuit and Behavioral Physiology, RIKEN Center for Brain Science, Wako-shi, Saitama, Japan
| | - Zhiyi Huang
- Department of Computer Science, University of Otago, Dunedin, New Zealand.
| |
Collapse
|
7
|
Wang Q, Pan M, Zang Z, Li DDU. Quantification of blood flow index in diffuse correlation spectroscopy using a robust deep learning method. JOURNAL OF BIOMEDICAL OPTICS 2024; 29:015004. [PMID: 38283935 PMCID: PMC10821781 DOI: 10.1117/1.jbo.29.1.015004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 12/22/2023] [Accepted: 01/02/2024] [Indexed: 01/30/2024]
Abstract
Significance Diffuse correlation spectroscopy (DCS) is a powerful, noninvasive optical technique for measuring blood flow. Traditionally the blood flow index (BFi) is derived through nonlinear least-square fitting the measured intensity autocorrelation function (ACF). However, the fitting process is computationally intensive, susceptible to measurement noise, and easily influenced by optical properties (absorption coefficient μ a and reduced scattering coefficient μ s ' ) and scalp and skull thicknesses. Aim We aim to develop a data-driven method that enables rapid and robust analysis of multiple-scattered light's temporal ACFs. Moreover, the proposed method can be applied to a range of source-detector distances instead of being limited to a specific source-detector distance. Approach We present a deep learning architecture with one-dimensional convolution neural networks, called DCS neural network (DCS-NET), for BFi and coherent factor (β ) estimation. This DCS-NET was performed using simulated DCS data based on a three-layer brain model. We quantified the impact from physiologically relevant optical property variations, layer thicknesses, realistic noise levels, and multiple source-detector distances (5, 10, 15, 20, 25, and 30 mm) on BFi and β estimations among DCS-NET, semi-infinite, and three-layer fitting models. Results DCS-NET shows a much faster analysis speed, around 17,000-fold and 32-fold faster than the traditional three-layer and semi-infinite models, respectively. It offers higher intrinsic sensitivity to deep tissues compared with fitting methods. DCS-NET shows excellent anti-noise features and is less sensitive to variations of μ a and μ s ' at a source-detector separation of 30 mm. Also, we have demonstrated that relative BFi (rBFi) can be extracted by DCS-NET with a much lower error of 8.35%. By contrast, the semi-infinite and three-layer fitting models result in significant errors in rBFi of 43.76% and 19.66%, respectively. Conclusions DCS-NET can robustly quantify blood flow measurements at considerable source-detector distances, corresponding to much deeper biological tissues. It has excellent potential for hardware implementation, promising continuous real-time blood flow measurements.
Collapse
Affiliation(s)
- Quan Wang
- University of Strathclyde, Department of Biomedical Engineering, Faculty of Engineering, Glasgow, United Kingdom
| | - Mingliang Pan
- University of Strathclyde, Department of Biomedical Engineering, Faculty of Engineering, Glasgow, United Kingdom
| | - Zhenya Zang
- University of Strathclyde, Department of Biomedical Engineering, Faculty of Engineering, Glasgow, United Kingdom
| | - David Day-Uei Li
- University of Strathclyde, Department of Biomedical Engineering, Faculty of Engineering, Glasgow, United Kingdom
| |
Collapse
|
8
|
Zhang H, Zhou QQ, Chen H, Hu XQ, Li WG, Bai Y, Han JX, Wang Y, Liang ZH, Chen D, Cong FY, Yan JQ, Li XL. The applied principles of EEG analysis methods in neuroscience and clinical neurology. Mil Med Res 2023; 10:67. [PMID: 38115158 PMCID: PMC10729551 DOI: 10.1186/s40779-023-00502-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 11/23/2023] [Indexed: 12/21/2023] Open
Abstract
Electroencephalography (EEG) is a non-invasive measurement method for brain activity. Due to its safety, high resolution, and hypersensitivity to dynamic changes in brain neural signals, EEG has aroused much interest in scientific research and medical fields. This article reviews the types of EEG signals, multiple EEG signal analysis methods, and the application of relevant methods in the neuroscience field and for diagnosing neurological diseases. First, three types of EEG signals, including time-invariant EEG, accurate event-related EEG, and random event-related EEG, are introduced. Second, five main directions for the methods of EEG analysis, including power spectrum analysis, time-frequency analysis, connectivity analysis, source localization methods, and machine learning methods, are described in the main section, along with different sub-methods and effect evaluations for solving the same problem. Finally, the application scenarios of different EEG analysis methods are emphasized, and the advantages and disadvantages of similar methods are distinguished. This article is expected to assist researchers in selecting suitable EEG analysis methods based on their research objectives, provide references for subsequent research, and summarize current issues and prospects for the future.
Collapse
Affiliation(s)
- Hao Zhang
- School of Systems Science, Beijing Normal University, Beijing, 100875, China
| | - Qing-Qi Zhou
- College of Electrical and Control Engineering, North China University of Technology, Beijing, 100041, China
| | - He Chen
- School of Automation Science and Engineering, South China University of Technology, Guangzhou, 510641, China
| | - Xiao-Qing Hu
- Department of Psychology, the State Key Laboratory of Brain and Cognitive Sciences, The University of Hong Kong, Hong Kong SAR, 999077, China
- HKU-Shenzhen Institute of Research and Innovation, Shenzhen, 518057, Guangdong, China
| | - Wei-Guang Li
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, 999077, China
| | - Yang Bai
- Department of Rehabilitation Medicine, the First Affiliated Hospital of Nanchang University, Nanchang, 330006, China
- Rehabilitation Medicine Clinical Research Center of Jiangxi Province, Nanchang, 330006, China
| | - Jun-Xia Han
- Beijing Key Laboratory of Learning and Cognition, School of Psychology, Capital Normal University, Beijing, 100048, China
| | - Yao Wang
- School of Communication Science, Beijing Language and Culture University, Beijing, 100083, China
| | - Zhen-Hu Liang
- Institute of Electrical Engineering, Yanshan University, Qinhuangdao, 066004, Hebei, China.
| | - Dan Chen
- School of Computer Science, Wuhan University, Wuhan, 430072, China.
| | - Feng-Yu Cong
- School of Biomedical Engineering, Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, 116081, Liaoning, China.
| | - Jia-Qing Yan
- College of Electrical and Control Engineering, North China University of Technology, Beijing, 100041, China.
| | - Xiao-Li Li
- School of Automation Science and Engineering, South China University of Technology, Guangzhou, 510641, China.
- Guangdong Artificial Intelligence and Digital Economy Laboratory (Guangzhou), Guangzhou, 510335, China.
| |
Collapse
|
9
|
Tao J, Dan Y, Zhou D. Local domain generalization with low-rank constraint for EEG-based emotion recognition. Front Neurosci 2023; 17:1213099. [PMID: 38027525 PMCID: PMC10662311 DOI: 10.3389/fnins.2023.1213099] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Accepted: 10/04/2023] [Indexed: 12/01/2023] Open
Abstract
As an important branch in the field of affective computing, emotion recognition based on electroencephalography (EEG) faces a long-standing challenge due to individual diversities. To conquer this challenge, domain adaptation (DA) or domain generalization (i.e., DA without target domain in the training stage) techniques have been introduced into EEG-based emotion recognition to eliminate the distribution discrepancy between different subjects. The preceding DA or domain generalization (DG) methods mainly focus on aligning the global distribution shift between source and target domains, yet without considering the correlations between the subdomains within the source domain and the target domain of interest. Since the ignorance of the fine-grained distribution information in the source may still bind the DG expectation on EEG datasets with multimodal structures, multiple patches (or subdomains) should be reconstructed from the source domain, on which multi-classifiers could be learned collaboratively. It is expected that accurately aligning relevant subdomains by excavating multiple distribution patterns within the source domain could further boost the learning performance of DG/DA. Therefore, we propose in this work a novel DG method for EEG-based emotion recognition, i.e., Local Domain Generalization with low-rank constraint (LDG). Specifically, the source domain is firstly partitioned into multiple local domains, each of which contains only one positive sample and its positive neighbors and k2 negative neighbors. Multiple subject-invariant classifiers on different subdomains are then co-learned in a unified framework by minimizing local regression loss with low-rank regularization for considering the shared knowledge among local domains. In the inference stage, the learned local classifiers are discriminatively selected according to their importance of adaptation. Extensive experiments are conducted on two benchmark databases (DEAP and SEED) under two cross-validation evaluation protocols, i.e., cross-subject within-dataset and cross-dataset within-session. The experimental results under the 5-fold cross-validation demonstrate the superiority of the proposed method compared with several state-of-the-art methods.
Collapse
Affiliation(s)
- Jianwen Tao
- Institute of Artificial Intelligence Application, Ningbo Polytechnic, Zhejiang, China
| | - Yufang Dan
- Institute of Artificial Intelligence Application, Ningbo Polytechnic, Zhejiang, China
| | - Di Zhou
- Industrial Technological Institute of Intelligent Manufacturing, Sichuan University of Arts and Science, Dazhou, China
| |
Collapse
|
10
|
Jafari M, Shoeibi A, Khodatars M, Bagherzadeh S, Shalbaf A, García DL, Gorriz JM, Acharya UR. Emotion recognition in EEG signals using deep learning methods: A review. Comput Biol Med 2023; 165:107450. [PMID: 37708717 DOI: 10.1016/j.compbiomed.2023.107450] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 08/03/2023] [Accepted: 09/01/2023] [Indexed: 09/16/2023]
Abstract
Emotions are a critical aspect of daily life and serve a crucial role in human decision-making, planning, reasoning, and other mental states. As a result, they are considered a significant factor in human interactions. Human emotions can be identified through various sources, such as facial expressions, speech, behavior (gesture/position), or physiological signals. The use of physiological signals can enhance the objectivity and reliability of emotion detection. Compared with peripheral physiological signals, electroencephalogram (EEG) recordings are directly generated by the central nervous system and are closely related to human emotions. EEG signals have the great spatial resolution that facilitates the evaluation of brain functions, making them a popular modality in emotion recognition studies. Emotion recognition using EEG signals presents several challenges, including signal variability due to electrode positioning, individual differences in signal morphology, and lack of a universal standard for EEG signal processing. Moreover, identifying the appropriate features for emotion recognition from EEG data requires further research. Finally, there is a need to develop more robust artificial intelligence (AI) including conventional machine learning (ML) and deep learning (DL) methods to handle the complex and diverse EEG signals associated with emotional states. This paper examines the application of DL techniques in emotion recognition from EEG signals and provides a detailed discussion of relevant articles. The paper explores the significant challenges in emotion recognition using EEG signals, highlights the potential of DL techniques in addressing these challenges, and suggests the scope for future research in emotion recognition using DL techniques. The paper concludes with a summary of its findings.
Collapse
Affiliation(s)
- Mahboobeh Jafari
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Afshin Shoeibi
- Data Science and Computational Intelligence Institute, University of Granada, Spain.
| | - Marjane Khodatars
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Sara Bagherzadeh
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Ahmad Shalbaf
- Department of Biomedical Engineering and Medical Physics, School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - David López García
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Juan M Gorriz
- Data Science and Computational Intelligence Institute, University of Granada, Spain; Department of Psychiatry, University of Cambridge, UK
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
11
|
Misthos LM, Krassanakis V, Merlemis N, Kesidis AL. Modeling the Visual Landscape: A Review on Approaches, Methods and Techniques. SENSORS (BASEL, SWITZERLAND) 2023; 23:8135. [PMID: 37836966 PMCID: PMC10574952 DOI: 10.3390/s23198135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Revised: 08/14/2023] [Accepted: 09/21/2023] [Indexed: 10/15/2023]
Abstract
Modeling the perception and evaluation of landscapes from the human perspective is a desirable goal for several scientific domains and applications. Human vision is the dominant sense, and human eyes are the sensors for apperceiving the environmental stimuli of our surroundings. Therefore, exploring the experimental recording and measurement of the visual landscape can reveal crucial aspects about human visual perception responses while viewing the natural or man-made landscapes. Landscape evaluation (or assessment) is another dimension that refers mainly to preferences of the visual landscape, involving human cognition as well, in ways that are often unpredictable. Yet, landscape can be approached by both egocentric (i.e., human view) and exocentric (i.e., bird's eye view) perspectives. The overarching approach of this review article lies in systematically presenting the different ways for modeling and quantifying the two 'modalities' of human perception and evaluation, under the two geometric perspectives, suggesting integrative approaches on these two 'diverging' dualities. To this end, several pertinent traditions/approaches, sensor-based experimental methods and techniques (e.g., eye tracking, fMRI, and EEG), and metrics are adduced and described. Essentially, this review article acts as a 'guide-map' for the delineation of the different activities related to landscape experience and/or management and to the valid or potentially suitable types of stimuli, sensors techniques, and metrics for each activity. Throughout our work, two main research directions are identified: (1) one that attempts to transfer the visual landscape experience/management from the one perspective to the other (and vice versa); (2) another one that aims to anticipate the visual perception of different landscapes and establish connections between perceptual processes and landscape preferences. As it appears, the research in the field is rapidly growing. In our opinion, it can be greatly advanced and enriched using integrative, interdisciplinary approaches in order to better understand the concepts and the mechanisms by which the visual landscape, as a complex set of stimuli, influences visual perception, potentially leading to more elaborate outcomes such as the anticipation of landscape preferences. As an effect, such approaches can support a rigorous, evidence-based, and socially just framework towards landscape management, protection, and decision making, based on a wide spectrum of well-suited and advanced sensor-based technologies.
Collapse
Affiliation(s)
- Loukas-Moysis Misthos
- Department of Surveying and Geoinformatics Engineering, University of West Attica, GR-12243 Athens, Greece; (L.-M.M.); (V.K.); (N.M.)
- Department of Public and One Health, University of Thessaly, GR-43100 Karditsa, Greece
| | - Vassilios Krassanakis
- Department of Surveying and Geoinformatics Engineering, University of West Attica, GR-12243 Athens, Greece; (L.-M.M.); (V.K.); (N.M.)
| | - Nikolaos Merlemis
- Department of Surveying and Geoinformatics Engineering, University of West Attica, GR-12243 Athens, Greece; (L.-M.M.); (V.K.); (N.M.)
| | - Anastasios L. Kesidis
- Department of Surveying and Geoinformatics Engineering, University of West Attica, GR-12243 Athens, Greece; (L.-M.M.); (V.K.); (N.M.)
| |
Collapse
|
12
|
Li R, Li W, Yue K, Zhang R, Li Y. Automatic snoring detection using a hybrid 1D-2D convolutional neural network. Sci Rep 2023; 13:14009. [PMID: 37640790 PMCID: PMC10462688 DOI: 10.1038/s41598-023-41170-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2022] [Accepted: 08/23/2023] [Indexed: 08/31/2023] Open
Abstract
Snoring, as a prevalent symptom, seriously interferes with life quality of patients with sleep disordered breathing only (simple snorers), patients with obstructive sleep apnea (OSA) and their bed partners. Researches have shown that snoring could be used for screening and diagnosis of OSA. Therefore, accurate detection of snoring sounds from sleep respiratory audio at night has been one of the most important parts. Considered that the snoring is somewhat dangerously overlooked around the world, an automatic and high-precision snoring detection algorithm is required. In this work, we designed a non-contact data acquire equipment to record nocturnal sleep respiratory audio of subjects in their private bedrooms, and proposed a hybrid convolutional neural network (CNN) model for the automatic snore detection. This model consists of a one-dimensional (1D) CNN processing the original signal and a two-dimensional (2D) CNN representing images mapped by the visibility graph method. In our experiment, our algorithm achieves an average classification accuracy of 89.3%, an average sensitivity of 89.7%, an average specificity of 88.5%, and an average AUC of 0.947, which surpasses some state-of-the-art models trained on our data. In conclusion, our results indicate that the proposed method in this study could be effective and significance for massive screening of OSA patients in daily life. And our work provides an alternative framework for time series analysis.
Collapse
Affiliation(s)
- Ruixue Li
- Key Laboratory of RF Circuits and Systems, Hangzhou Dianzi University, Hangzhou, Zhejiang, China
| | - Wenjun Li
- Key Laboratory of RF Circuits and Systems, Hangzhou Dianzi University, Hangzhou, Zhejiang, China.
| | - Keqiang Yue
- Key Laboratory of RF Circuits and Systems, Hangzhou Dianzi University, Hangzhou, Zhejiang, China
| | - Rulin Zhang
- Key Laboratory of RF Circuits and Systems, Hangzhou Dianzi University, Hangzhou, Zhejiang, China
| | - Yilin Li
- Key Laboratory of RF Circuits and Systems, Hangzhou Dianzi University, Hangzhou, Zhejiang, China
| |
Collapse
|
13
|
Xu H, Cao K, Chen H, Abudusalamu A, Wu W, Xue Y. Emotional brain network decoded by biological spiking neural network. Front Neurosci 2023; 17:1200701. [PMID: 37496741 PMCID: PMC10366476 DOI: 10.3389/fnins.2023.1200701] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2023] [Accepted: 05/30/2023] [Indexed: 07/28/2023] Open
Abstract
Introduction Emotional disorders are essential manifestations of many neurological and psychiatric diseases. Nowadays, researchers try to explore bi-directional brain-computer interface techniques to help the patients. However, the related functional brain areas and biological markers are still unclear, and the dynamic connection mechanism is also unknown. Methods To find effective regions related to different emotion recognition and intervention, our research focuses on finding emotional EEG brain networks using spiking neural network algorithm with binary coding. We collected EEG data while human participants watched emotional videos (fear, sadness, happiness, and neutrality), and analyzed the dynamic connections between the electrodes and the biological rhythms of different emotions. Results The analysis has shown that the local high-activation brain network of fear and sadness is mainly in the parietal lobe area. The local high-level brain network of happiness is in the prefrontal-temporal lobe-central area. Furthermore, the α frequency band could effectively represent negative emotions, while the α frequency band could be used as a biological marker of happiness. The decoding accuracy of the three emotions reached 86.36%, 95.18%, and 89.09%, respectively, fully reflecting the excellent emotional decoding performance of the spiking neural network with self- backpropagation. Discussion The introduction of the self-backpropagation mechanism effectively improves the performance of the spiking neural network model. Different emotions exhibit distinct EEG networks and neuro-oscillatory-based biological markers. These emotional brain networks and biological markers may provide important hints for brain-computer interface technique exploration to help related brain disease recovery.
Collapse
Affiliation(s)
- Hubo Xu
- National Institute on Drug Dependence and Beijing Key Laboratory of Drug Dependence, Peking University, Beijing, China
- Department of Pharmacology, School of Basic Medical Sciences, Peking University, Beijing, China
| | - Kexin Cao
- National Institute on Drug Dependence and Beijing Key Laboratory of Drug Dependence, Peking University, Beijing, China
- Department of Pharmacology, School of Basic Medical Sciences, Peking University, Beijing, China
| | - Hongguang Chen
- NHC Key Laboratory of Mental Health (Peking University), National Clinical Research Center for Mental Disorders (Peking University Sixth Hospital), Peking University Institute of Mental Health, Peking University Sixth Hospital, Beijing, China
| | - Awuti Abudusalamu
- National Institute on Drug Dependence and Beijing Key Laboratory of Drug Dependence, Peking University, Beijing, China
- Department of Pharmacology, School of Basic Medical Sciences, Peking University, Beijing, China
| | - Wei Wu
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Yanxue Xue
- National Institute on Drug Dependence and Beijing Key Laboratory of Drug Dependence, Peking University, Beijing, China
- Chinese Institute for Brain Research, Beijing, China
- Key Laboratory for Neuroscience, Ministry of Education/National Health Commission, Peking University, Beijing, China
| |
Collapse
|
14
|
Gong L, Li M, Zhang T, Chen W. EEG emotion recognition using attention-based convolutional transformer neural network. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104835] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/12/2023]
|
15
|
Hosseini MSK, Firoozabadi SM, Badie K, Azadfallah P. Personality-Based Emotion Recognition Using EEG Signals with a CNN-LSTM Network. Brain Sci 2023; 13:947. [PMID: 37371425 DOI: 10.3390/brainsci13060947] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2023] [Revised: 05/27/2023] [Accepted: 05/30/2023] [Indexed: 06/29/2023] Open
Abstract
The accurate detection of emotions has significant implications in healthcare, psychology, and human-computer interaction. Integrating personality information into emotion recognition can enhance its utility in various applications. The present study introduces a novel deep learning approach to emotion recognition, which utilizes electroencephalography (EEG) signals and the Big Five personality traits. The study recruited 60 participants and recorded their EEG data while they viewed unique sequence stimuli designed to effectively capture the dynamic nature of human emotions and personality traits. A pre-trained convolutional neural network (CNN) was used to extract emotion-related features from the raw EEG data. Additionally, a long short-term memory (LSTM) network was used to extract features related to the Big Five personality traits. The network was able to accurately predict personality traits from EEG data. The extracted features were subsequently used in a novel network to predict emotional states within the arousal and valence dimensions. The experimental results showed that the proposed classifier outperformed common classifiers, with a high accuracy of 93.97%. The findings suggest that incorporating personality traits as features in the designed network, for emotion recognition, leads to higher accuracy, highlighting the significance of examining these traits in the analysis of emotions.
Collapse
Affiliation(s)
| | - Seyed Mohammad Firoozabadi
- Department of Medical Physics, Faculty of Medicine, Tarbiat Modares University, Tehran 14117-13116, Iran
| | - Kambiz Badie
- Content & E-Services Research Group, IT Research Faculty, ICT Research Institute, Tehran 14399-55471, Iran
| | - Parviz Azadfallah
- Department of Psychology, Faculty of Humanities, Tarbiat Modares University, Tehran 14117-13116, Iran
| |
Collapse
|
16
|
Cai Y, Li X, Li J. Emotion Recognition Using Different Sensors, Emotion Models, Methods and Datasets: A Comprehensive Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23052455. [PMID: 36904659 PMCID: PMC10007272 DOI: 10.3390/s23052455] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/31/2023] [Revised: 02/18/2023] [Accepted: 02/21/2023] [Indexed: 06/12/2023]
Abstract
In recent years, the rapid development of sensors and information technology has made it possible for machines to recognize and analyze human emotions. Emotion recognition is an important research direction in various fields. Human emotions have many manifestations. Therefore, emotion recognition can be realized by analyzing facial expressions, speech, behavior, or physiological signals. These signals are collected by different sensors. Correct recognition of human emotions can promote the development of affective computing. Most existing emotion recognition surveys only focus on a single sensor. Therefore, it is more important to compare different sensors or unimodality and multimodality. In this survey, we collect and review more than 200 papers on emotion recognition by literature research methods. We categorize these papers according to different innovations. These articles mainly focus on the methods and datasets used for emotion recognition with different sensors. This survey also provides application examples and developments in emotion recognition. Furthermore, this survey compares the advantages and disadvantages of different sensors for emotion recognition. The proposed survey can help researchers gain a better understanding of existing emotion recognition systems, thus facilitating the selection of suitable sensors, algorithms, and datasets.
Collapse
|
17
|
Quiles Pérez M, Martínez Beltrán ET, López Bernal S, Martínez Pérez G, Huertas Celdrán A. Analyzing the impact of Driving tasks when detecting emotions through brain–computer interfaces. Neural Comput Appl 2023. [DOI: 10.1007/s00521-023-08343-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/25/2023]
Abstract
AbstractTraffic accidents are the leading cause of death among young people, a problem that today costs an enormous number of victims. Several technologies have been proposed to prevent accidents, being brain–computer interfaces (BCIs) one of the most promising. In this context, BCIs have been used to detect emotional states, concentration issues, or stressful situations, which could play a fundamental role in the road since they are directly related to the drivers’ decisions. However, there is no extensive literature applying BCIs to detect subjects’ emotions in driving scenarios. In such a context, there are some challenges to be solved, such as (i) the impact of performing a driving task on the emotion detection and (ii) which emotions are more detectable in driving scenarios. To improve these challenges, this work proposes a framework focused on detecting emotions using electroencephalography with machine learning and deep learning algorithms. In addition, a use case has been designed where two scenarios are presented. The first scenario consists in listening to sounds as the primary task to perform, while in the second scenario listening to sound becomes a secondary task, being the primary task using a driving simulator. In this way, it is intended to demonstrate whether BCIs are useful in this driving scenario. The results improve those existing in the literature, achieving 99% accuracy for the detection of two emotions (non-stimuli and angry), 93% for three emotions (non-stimuli, angry and neutral) and 75% for four emotions (non-stimuli, angry, neutral and joy).
Collapse
|
18
|
Dimauro G, Griseta ME, Camporeale MG, Clemente F, Guarini A, Maglietta R. An intelligent non-invasive system for automated diagnosis of anemia exploiting a novel dataset. Artif Intell Med 2023; 136:102477. [PMID: 36710064 DOI: 10.1016/j.artmed.2022.102477] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2022] [Revised: 12/19/2022] [Accepted: 12/19/2022] [Indexed: 12/27/2022]
Abstract
Anemia is a condition in which the oxygen-carrying capacity of red blood cells is insufficient to meet the body's physiological needs. It affects billions of people worldwide. An early diagnosis of this disease could prevent the advancement of other disorders. Traditional methods used to detect anemia consist of venipuncture, which requires a patient to frequently undergo laboratory tests. Therefore, anemia diagnosis using noninvasive and cost-effective methods is an open challenge. The pallor of the fingertips, palms, nail beds, and eye conjunctiva can be observed to establish whether a patient suffers from anemia. This article addresses the above challenges by presenting a novel intelligent system, based on machine learning, that supports the automated diagnosis of anemia. This system is innovative from different points of view. Specifically, it has been trained on a dataset that contains eye conjunctiva photos of Indian and Italian patients. This dataset, which was created using a very strict experimental set, is now made available to the Scientific Community. Moreover, compared to previous systems in the literature, the proposed system uses a low-cost device, which makes it suitable for widespread use. The performance of the learning algorithms utilizing two different areas of the mucous membrane of the eye is discussed. In particular, the RUSBoost algorithm, when appropriately trained on palpebral conjunctiva images, shows good performance in classifying anemic and nonanemic patients. The results are very robust, even when considering different ethnicities.
Collapse
Affiliation(s)
- Giovanni Dimauro
- Department of Computer Science, University of Bari 'Aldo Moro', Bari, Italy.
| | - Maria Elena Griseta
- Institute of Intelligent Industrial Technologies and Systems for Advanced Manufacturing, National Research Council of Italy, Bari, Italy.
| | | | - Felice Clemente
- Haematology Dept. of National Cancer Institute 'Giovanni Paolo II', Bari, Italy.
| | - Attilio Guarini
- Haematology Dept. of National Cancer Institute 'Giovanni Paolo II', Bari, Italy.
| | - Rosalia Maglietta
- Institute of Intelligent Industrial Technologies and Systems for Advanced Manufacturing, National Research Council of Italy, Bari, Italy.
| |
Collapse
|
19
|
Li J, Pan W, Huang H, Pan J, Wang F. STGATE: Spatial-temporal graph attention network with a transformer encoder for EEG-based emotion recognition. Front Hum Neurosci 2023; 17:1169949. [PMID: 37125349 PMCID: PMC10133470 DOI: 10.3389/fnhum.2023.1169949] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Accepted: 03/27/2023] [Indexed: 05/02/2023] Open
Abstract
Electroencephalogram (EEG) is a crucial and widely utilized technique in neuroscience research. In this paper, we introduce a novel graph neural network called the spatial-temporal graph attention network with a transformer encoder (STGATE) to learn graph representations of emotion EEG signals and improve emotion recognition performance. In STGATE, a transformer-encoder is applied for capturing time-frequency features which are fed into a spatial-temporal graph attention for emotion classification. Using a dynamic adjacency matrix, the proposed STGATE adaptively learns intrinsic connections between different EEG channels. To evaluate the cross-subject emotion recognition performance, leave-one-subject-out experiments are carried out on three public emotion recognition datasets, i.e., SEED, SEED-IV, and DREAMER. The proposed STGATE model achieved a state-of-the-art EEG-based emotion recognition performance accuracy of 90.37% in SEED, 76.43% in SEED-IV, and 76.35% in DREAMER dataset, respectively. The experiments demonstrated the effectiveness of the proposed STGATE model for cross-subject EEG emotion recognition and its potential for graph-based neuroscience research.
Collapse
|
20
|
Zhang X, Li H, Dong R, Lu Z, Li C. Electroencephalogram and surface electromyogram fusion-based precise detection of lower limb voluntary movement using convolution neural network-long short-term memory model. Front Neurosci 2022; 16:954387. [PMID: 36213740 PMCID: PMC9538146 DOI: 10.3389/fnins.2022.954387] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2022] [Accepted: 08/26/2022] [Indexed: 11/13/2022] Open
Abstract
The electroencephalogram (EEG) and surface electromyogram (sEMG) fusion has been widely used in the detection of human movement intention for human–robot interaction, but the internal relationship of EEG and sEMG signals is not clear, so their fusion still has some shortcomings. A precise fusion method of EEG and sEMG using the CNN-LSTM model was investigated to detect lower limb voluntary movement in this study. At first, the EEG and sEMG signal processing of each stage was analyzed so that the response time difference between EEG and sEMG can be estimated to detect lower limb voluntary movement, and it can be calculated by the symbolic transfer entropy. Second, the data fusion and feature of EEG and sEMG were both used for obtaining a data matrix of the model, and a hybrid CNN-LSTM model was established for the EEG and sEMG-based decoding model of lower limb voluntary movement so that the estimated value of time difference was about 24 ∼ 26 ms, and the calculated value was between 25 and 45 ms. Finally, the offline experimental results showed that the accuracy of data fusion was significantly higher than feature fusion-based accuracy in 5-fold cross-validation, and the average accuracy of EEG and sEMG data fusion was more than 95%; the improved average accuracy for eliminating the response time difference between EEG and sEMG was about 0.7 ± 0.26% in data fusion. In the meantime, the online average accuracy of data fusion-based CNN-LSTM was more than 87% in all subjects. These results demonstrated that the time difference had an influence on the EEG and sEMG fusion to detect lower limb voluntary movement, and the proposed CNN-LSTM model can achieve high performance. This work provides a stable and reliable basis for human–robot interaction of the lower limb exoskeleton.
Collapse
Affiliation(s)
- Xiaodong Zhang
- School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an, Shaanxi, China
- Shaanxi Key Laboratory of Intelligent Robots, Xi’an Jiaotong University, Xi’an, Shaanxi, China
- Wearable Human Enhancement Technology Innovation Center, Xi’an Jiaotong University, Xi’an, Shaanxi, China
| | - Hanzhe Li
- School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an, Shaanxi, China
- Wearable Human Enhancement Technology Innovation Center, Xi’an Jiaotong University, Xi’an, Shaanxi, China
- *Correspondence: Hanzhe Li,
| | - Runlin Dong
- School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an, Shaanxi, China
| | - Zhufeng Lu
- School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an, Shaanxi, China
| | - Cunxin Li
- School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an, Shaanxi, China
| |
Collapse
|
21
|
Jiang L, Siriaraya P, Choi D, Zeng F, Kuwahara N. Electroencephalogram signals emotion recognition based on convolutional neural network-recurrent neural network framework with channel-temporal attention mechanism for older adults. Front Aging Neurosci 2022; 14:945024. [PMID: 36212045 PMCID: PMC9535340 DOI: 10.3389/fnagi.2022.945024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Accepted: 08/29/2022] [Indexed: 11/13/2022] Open
Abstract
Reminiscence and conversation between older adults and younger volunteers using past photographs are very effective in improving the emotional state of older adults and alleviating depression. However, we need to evaluate the emotional state of the older adult while conversing on the past photographs. While electroencephalogram (EEG) has a significantly stronger association with emotion than other physiological signals, the challenge is to eliminate muscle artifacts in the EEG during speech as well as to reduce the number of dry electrodes to improve user comfort while maintaining high emotion recognition accuracy. Therefore, we proposed the CTA-CNN-Bi-LSTM emotion recognition framework. EEG signals of eight channels (P3, P4, F3, F4, F7, F8, T7, and T8) were first implemented in the MEMD-CCA method on three brain regions separately (Frontal, Temporal, Parietal) to remove the muscle artifacts then were fed into the Channel-Temporal attention module to get the weights of channels and temporal points most relevant to the positive, negative and neutral emotions to recode the EEG data. A Convolutional Neural Networks (CNNs) module then extracted the spatial information in the new EEG data to obtain the spatial feature maps which were then sequentially inputted into a Bi-LSTM module to learn the bi-directional temporal information for emotion recognition. Finally, we designed four group experiments to demonstrate that the proposed CTA-CNN-Bi-LSTM framework outperforms the previous works. And the highest average recognition accuracy of the positive, negative, and neutral emotions achieved 98.75%.
Collapse
Affiliation(s)
- Lei Jiang
- Graduate School of Science and Technology, Kyoto Institute of Technology, Kyoto, Japan
| | - Panote Siriaraya
- Graduate School of Science and Technology, Kyoto Institute of Technology, Kyoto, Japan
| | - Dongeun Choi
- Faculty of Informatics, The University of Fukuchiyama, Kyoto, Japan
| | - Fangmeng Zeng
- College of Textile Science and Engineering, Zhejiang Sci-Tech University, Hangzhou, China
| | - Noriaki Kuwahara
- Graduate School of Science and Technology, Kyoto Institute of Technology, Kyoto, Japan
- *Correspondence: Noriaki Kuwahara,
| |
Collapse
|
22
|
Wang Y, Zhang L, Xia P, Wang P, Chen X, Du L, Fang Z, Du M. EEG-Based Emotion Recognition Using a 2D CNN with Different Kernels. Bioengineering (Basel) 2022; 9:bioengineering9060231. [PMID: 35735474 PMCID: PMC9219701 DOI: 10.3390/bioengineering9060231] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2022] [Revised: 05/21/2022] [Accepted: 05/23/2022] [Indexed: 11/16/2022] Open
Abstract
Emotion recognition is receiving significant attention in research on health care and Human-Computer Interaction (HCI). Due to the high correlation with emotion and the capability to affect deceptive external expressions such as voices and faces, Electroencephalogram (EEG) based emotion recognition methods have been globally accepted and widely applied. Recently, great improvements have been made in the development of machine learning for EEG-based emotion detection. However, there are still some major disadvantages in previous studies. Firstly, traditional machine learning methods require extracting features manually which is time-consuming and rely heavily on human experts. Secondly, to improve the model accuracies, many researchers used user-dependent models that lack generalization and universality. Moreover, there is still room for improvement in the recognition accuracies in most studies. Therefore, to overcome these shortcomings, an EEG-based novel deep neural network is proposed for emotion classification in this article. The proposed 2D CNN uses two convolutional kernels of different sizes to extract emotion-related features along both the time direction and the spatial direction. To verify the feasibility of the proposed model, the pubic emotion dataset DEAP is used in experiments. The results show accuracies of up to 99.99% and 99.98 for arousal and valence binary classification, respectively, which are encouraging for research and applications in the emotion recognition field.
Collapse
Affiliation(s)
- Yuqi Wang
- Institute of Microelectronics of Chinese Academy of Sciences, Beijing 100029, China; (Y.W.); (L.Z.)
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
| | - Lijun Zhang
- Institute of Microelectronics of Chinese Academy of Sciences, Beijing 100029, China; (Y.W.); (L.Z.)
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
| | - Pan Xia
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
- Aerospace Information Research Institute, Chinese Academy of Sciences (AIRCAS), Beijing 100190, China
| | - Peng Wang
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
- Aerospace Information Research Institute, Chinese Academy of Sciences (AIRCAS), Beijing 100190, China
| | - Xianxiang Chen
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
- Aerospace Information Research Institute, Chinese Academy of Sciences (AIRCAS), Beijing 100190, China
| | - Lidong Du
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
- Aerospace Information Research Institute, Chinese Academy of Sciences (AIRCAS), Beijing 100190, China
| | - Zhen Fang
- University of Chinese Academy of Sciences, Beijing 100049, China; (P.X.); (P.W.); (X.C.); (L.D.)
- Aerospace Information Research Institute, Chinese Academy of Sciences (AIRCAS), Beijing 100190, China
- Personalized Management of Chronic Respiratory Disease, Chinese Academy of Medical Sciences, Beijing 100190, China
- Correspondence: (Z.F.); (M.D.)
| | - Mingyan Du
- China Beijing Luhe Hospital, Capital Medical University, Beijing 101199, China
- Correspondence: (Z.F.); (M.D.)
| |
Collapse
|
23
|
Dar MN, Akram MU, Yuvaraj R, Gul Khawaja S, Murugappan M. EEG-based emotion charting for Parkinson's disease patients using Convolutional Recurrent Neural Networks and cross dataset learning. Comput Biol Med 2022; 144:105327. [PMID: 35303579 DOI: 10.1016/j.compbiomed.2022.105327] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2021] [Revised: 01/30/2022] [Accepted: 02/14/2022] [Indexed: 01/04/2023]
Abstract
Electroencephalogram (EEG) based emotion classification reflects the actual and intrinsic emotional state, resulting in more reliable, natural, and meaningful human-computer interaction with applications in entertainment consumption behavior, interactive brain-computer interface, and monitoring of psychological health of patients in the domain of e-healthcare. Challenges of EEG-based emotion recognition in real-world applications are variations among experimental settings and cognitive health conditions. Parkinson's Disease (PD) is the second most common neurodegenerative disorder, resulting in impaired recognition and expression of emotions. The deficit of emotional expression poses challenges for the healthcare services provided to PD patients. This study proposes 1D-CRNN-ELM architecture, which combines one-dimensional Convolutional Recurrent Neural Network (1D-CRNN) with an Extreme Learning Machine (ELM), robust for the emotion detection of PD patients, also available for cross dataset learning with various emotions and experimental settings. In the proposed framework, after EEG preprocessing, the trained CRNN can use as a feature extractor with ELM as the classifier, and again this trained CRNN can be used for learning of new emotions set with fine-tuning of other datasets. This paper also applied cross dataset learning of emotions by training with PD patients datasets and fine-tuning with publicly available datasets of AMIGOS and SEED-IV, and vice versa. Random splitting of train and test data with 80 - 20 ratio resulted in an accuracy of 97.75% for AMIGOS, 83.20% for PD, and 86.00% for HC with six basic emotion classes. Fine-tuning of trained architecture with four emotions of the SEED-IV dataset results in 92.5% accuracy. To validate the generalization of our results, leave one subject (patient) out cross-validation is also incorporated with mean accuracies of 95.84% for AMIGOS, 75.09% for PD, 77.85% for HC, and 84.97% for SEED-IV is achieved. Only a 1 - sec segment of EEG signal from 14 channels is enough to detect emotions with this performance. The proposed method outperforms state-of-the-art studies to classify EEG-based emotions with publicly available datasets, provide cross dataset learning, and validate the robustness of the deep learning framework for real-world application of psychological healthcare monitoring of Parkinson's disease patients.
Collapse
Affiliation(s)
- Muhammad Najam Dar
- National University of Sciences and Technology, Islamabad, Postcode: 44000, Pakistan.
| | - Muhammad Usman Akram
- National University of Sciences and Technology, Islamabad, Postcode: 44000, Pakistan.
| | - Rajamanickam Yuvaraj
- Nanyang Technological University (NTU), 639798, Singapore; Science of Learning in Education (SoLE), Office of Education Research (OER), National Institute of Education (NIE), 637616, Singapore.
| | - Sajid Gul Khawaja
- National University of Sciences and Technology, Islamabad, Postcode: 44000, Pakistan.
| | - M Murugappan
- Department of Electronics and Communication Engineering, Kuwait College of Science and Technology, Al-Jahra, Kuwait.
| |
Collapse
|
24
|
Maithri M, Raghavendra U, Gudigar A, Samanth J, Murugappan M, Chakole Y, Acharya UR. Automated emotion recognition: Current trends and future perspectives. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 215:106646. [PMID: 35093645 DOI: 10.1016/j.cmpb.2022.106646] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/08/2021] [Revised: 12/25/2021] [Accepted: 01/16/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND Human emotions greatly affect the actions of a person. The automated emotion recognition has applications in multiple domains such as health care, e-learning, surveillance, etc. The development of computer-aided diagnosis (CAD) tools has led to the automated recognition of human emotions. OBJECTIVE This review paper provides an insight into various methods employed using electroencephalogram (EEG), facial, and speech signals coupled with multi-modal emotion recognition techniques. In this work, we have reviewed most of the state-of-the-art papers published on this topic. METHOD This study was carried out by considering the various emotion recognition (ER) models proposed between 2016 and 2021. The papers were analysed based on methods employed, classifier used and performance obtained. RESULTS There is a significant rise in the application of deep learning techniques for ER. They have been widely applied for EEG, speech, facial expression, and multimodal features to develop an accurate ER model. CONCLUSION Our study reveals that most of the proposed machine and deep learning-based systems have yielded good performances for automated ER in a controlled environment. However, there is a need to obtain high performance for ER even in an uncontrolled environment.
Collapse
Affiliation(s)
- M Maithri
- Department of Mechatronics, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - U Raghavendra
- Department of Instrumentation and Control Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Anjan Gudigar
- Department of Instrumentation and Control Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India.
| | - Jyothi Samanth
- Department of Cardiovascular Technology, Manipal College of Health Professions, Manipal Academy of Higher Education, Manipal, Karnataka 576104, India
| | - Murugappan Murugappan
- Department of Electronics and Communication Engineering, Kuwait College of Science and Technology, 13133, Kuwait
| | - Yashas Chakole
- Department of Instrumentation and Control Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - U Rajendra Acharya
- School of Engineering, Ngee Ann Polytechnic, Clementi 599489, Singapore; Department of Biomedical Informatics and Medical Engineering, Asia University, Taichung, Taiwan; Department of Biomedical Engineering, School of Science and Technology, SUSS University, Singapore
| |
Collapse
|
25
|
Recognition of the Mental Workloads of Pilots in the Cockpit Using EEG Signals. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12052298] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
The commercial flightdeck is a naturally multi-tasking work environment, one in which interruptions are frequent come in various forms, contributing in many cases to aviation incident reports. Automatic characterization of pilots’ workloads is essential to preventing these kind of incidents. In addition, minimizing the physiological sensor network as much as possible remains both a challenge and a requirement. Electroencephalogram (EEG) signals have shown high correlations with specific cognitive and mental states, such as workload. However, there is not enough evidence in the literature to validate how well models generalize in cases of new subjects performing tasks with workloads similar to the ones included during the model’s training. In this paper, we propose a convolutional neural network to classify EEG features across different mental workloads in a continuous performance task test that partly measures working memory and working memory capacity. Our model is valid at the general population level and it is able to transfer task learning to pilot mental workload recognition in a simulated operational environment.
Collapse
|
26
|
Abstract
As a long-standing research topic in the field of brain–computer interface, emotion recognition still suffers from low recognition accuracy. In this research, we present a novel model named DE-CNN-BiLSTM deeply integrating the complexity of EEG signals, the spatial structure of brain and temporal contexts of emotion formation. Firstly, we extract the complexity properties of the EEG signal by calculating Differential Entropy in different time slices of different frequency bands to obtain 4D feature tensors according to brain location. Subsequently, the 4D tensors are input into the Convolutional Neural Network to learn brain structure and output time sequences; after that Bidirectional Long-Short Term Memory is used to learn past and future information of the time sequences. Compared with the existing emotion recognition models, the new model can decode the EEG signal deeply and extract key emotional features to improve accuracy. The simulation results show the algorithm achieves an average accuracy of 94% for DEAP dataset and 94.82% for SEED dataset, confirming its high accuracy and strong robustness.
Collapse
|
27
|
Jurczak M, Kołodziej M, Majkowski A. Implementation of a Convolutional Neural Network for Eye Blink Artifacts Removal From the Electroencephalography Signal. Front Neurosci 2022; 16:782367. [PMID: 35221897 PMCID: PMC8874023 DOI: 10.3389/fnins.2022.782367] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/24/2021] [Accepted: 01/10/2022] [Indexed: 01/01/2023] Open
Abstract
Electroencephalography (EEG) signals are disrupted by technical and physiological artifacts. One of the most common artifacts is the natural activity that results from the movement of the eyes and the blinking of the subject. Eye blink artifacts (EB) spread across the entire head surface and make EEG signal analysis difficult. Methods for the elimination of electrooculography (EOG) artifacts, such as independent component analysis (ICA) and regression, are known. The aim of this article was to implement the convolutional neural network (CNN) to eliminate eye blink artifacts. To train the CNN, a method for augmenting EEG signals was proposed. The results obtained from the CNN were compared with the results of the ICA and regression methods for the generated and real EEG signals. The results obtained indicate a much better performance of the CNN in the task of removing eye-blink artifacts, in particular for the electrodes located in the central part of the head.
Collapse
|
28
|
Chang Y, Stevenson C, Chen IC, Lin DS, Ko LW. Neurological state changes indicative of ADHD in children learned via EEG-based LSTM networks. J Neural Eng 2022; 19. [PMID: 35081524 DOI: 10.1088/1741-2552/ac4f07] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2021] [Accepted: 01/26/2022] [Indexed: 11/12/2022]
Abstract
Attention-deficit/hyperactivity disorder (ADHD) is a neurodevelopmental disorder that pervasively interferes with the lives of individuals starting in childhood. OBJECTIVE To address the subjectivity of current diagnostic approaches, many studies have been dedicated to efforts to identify the differences between ADHD and neurotypical (NT) individuals using EEG and continuous performance tests (CPT). APPROACH In this study, we proposed EEG-based long short-term memory (LSTM) networks that utilize deep learning techniques with learning the cognitive state transition to discriminate between ADHD and NT children via EEG signal processing. A total of thirty neurotypical children and thirty ADHD children participated in CPT tests while being monitored with EEG. Several architectures of deep and machine learning were applied to three EEG data segments including resting state, cognitive execution, and a period containing a fusion of those. MAIN RESULTS The experimental results indicated that EEG-based LSTM networks produced the best performance with an average accuracy of 90.50 ± 0.81 % in comparison with the deep neural networks, the convolutional neural networks, and the support vector machines with learning the cognitive state transition of EEG data. Novel observations of individual neural markers showed that the beta power activity of the O1 and O2 sites contributed the most to the classifications, subjects exhibited decreased beta power in the ADHD group, and had larger decreases during cognitive execution. SIGNIFICANCE These findings showed that the proposed EEG-based LSTM networks are capable of extracting the varied temporal characteristics of high-resolution electrophysiological signals to differentiate between ADHD and NT children, and brought a new insight to facilitate the diagnosis of ADHD. The registration numbers of the institutional review boards are 16MMHIS021 and EC1070401-F.
Collapse
Affiliation(s)
- Yang Chang
- Institute of Bioinformatics and Systems Biology, National Yang Ming Chiao Tung University, Rm. 742, Bio-ICT Building, No. 75, Bo'ai St., East Dist., Hsinchu City 300 , Taiwan (R.O.C.), Hsinchu, 300, TAIWAN
| | - Cory Stevenson
- Institute of Bioinformatics and Systems Biology, National Yang Ming Chiao Tung University, Rm. 742, Bio-ICT Building, No. 75, Bo'ai St., East Dist., Hsinchu City 300 , Taiwan (R.O.C.), Hsinchu, 300, TAIWAN
| | - I-Chun Chen
- Department of Physical Medicine and Rehabilitation, Ton-Yen General Hospital, No. 69, Xianzheng 2nd Rd., Zhubei City, Hsinchu County 302, Taiwan (R.O.C.), Hsinchu, 302, TAIWAN
| | - Dar-Shong Lin
- Department of Pediatrics, Mackay Memorial Hospital, No. 92, Sec. 2, Zhongshan N. Rd., Zhongshan Dist., Taipei City 104, Taiwan (R.O.C.), Taipei, 104, TAIWAN
| | - Li-Wei Ko
- Institute of Bioinformatics and Systems Biology, National Chiao Tung University, Rm. 742, Bio-ICT Building, No. 75, Bo'ai St., East Dist., Hsinchu City 300 , Taiwan (R.O.C.), Hsinchu, 300, TAIWAN
| |
Collapse
|
29
|
Khan HA, Ul Ain R, Kamboh AM, Butt HT, Shafait S, Alamgir W, Stricker D, Shafait F. The NMT Scalp EEG Dataset: An Open-Source Annotated Dataset of Healthy and Pathological EEG Recordings for Predictive Modeling. Front Neurosci 2022; 15:755817. [PMID: 35069095 PMCID: PMC8766964 DOI: 10.3389/fnins.2021.755817] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2021] [Accepted: 11/30/2021] [Indexed: 11/13/2022] Open
Abstract
Electroencephalogram (EEG) is widely used for the diagnosis of neurological conditions like epilepsy, neurodegenerative illnesses and sleep related disorders. Proper interpretation of EEG recordings requires the expertise of trained neurologists, a resource which is scarce in the developing world. Neurologists spend a significant portion of their time sifting through EEG recordings looking for abnormalities. Most recordings turn out to be completely normal, owing to the low yield of EEG tests. To minimize such wastage of time and effort, automatic algorithms could be used to provide pre-diagnostic screening to separate normal from abnormal EEG. Data driven machine learning offers a way forward however, design and verification of modern machine learning algorithms require properly curated labeled datasets. To avoid bias, deep learning based methods must be trained on large datasets from diverse sources. This work presents a new open-source dataset, named the NMT Scalp EEG Dataset, consisting of 2,417 recordings from unique participants spanning almost 625 h. Each recording is labeled as normal or abnormal by a team of qualified neurologists. Demographic information such as gender and age of the patient are also included. Our dataset focuses on the South Asian population. Several existing state-of-the-art deep learning architectures developed for pre-diagnostic screening of EEG are implemented and evaluated on the NMT, and referenced against baseline performance on the well-known Temple University Hospital EEG Abnormal Corpus. Generalization of deep learning based architectures across the NMT and the reference datasets is also investigated. The NMT dataset is being released to increase the diversity of EEG datasets and to overcome the scarcity of accurately annotated publicly available datasets for EEG research.
Collapse
Affiliation(s)
- Hassan Aqeel Khan
- College of Computer Science and Engineering, University of Jeddah, Jeddah, Saudi Arabia
- *Correspondence: Hassan Aqeel Khan
| | - Rahat Ul Ain
- School of Electrical Engineering and Computer Science, National University of Sciences and Technology (NUST), Islamabad, Pakistan
| | - Awais Mehmood Kamboh
- College of Computer Science and Engineering, University of Jeddah, Jeddah, Saudi Arabia
- School of Electrical Engineering and Computer Science, National University of Sciences and Technology (NUST), Islamabad, Pakistan
| | - Hammad Tanveer Butt
- Department of Computer Science, University of Kaiserslautern, Kaiserslautern, Germany
- College of Electrical and Mechanical Engineering, National University of Sciences and Technology (NUST), Islamabad, Pakistan
| | - Saima Shafait
- Pak-Emirates Military Hospital, Rawalpindi, Pakistan
| | - Wasim Alamgir
- Pak-Emirates Military Hospital, Rawalpindi, Pakistan
| | - Didier Stricker
- Department of Computer Science, University of Kaiserslautern, Kaiserslautern, Germany
- German Research Center for Artificial Intelligence (DFKI), Kaiserslautern, Germany
| | - Faisal Shafait
- School of Electrical Engineering and Computer Science, National University of Sciences and Technology (NUST), Islamabad, Pakistan
- Deep Learning Laboratory, National Center of Artificial Intelligence, Islamabad, Pakistan
| |
Collapse
|
30
|
Huang X, Xu Y, Hua J, Yi W, Yin H, Hu R, Wang S. A Review on Signal Processing Approaches to Reduce Calibration Time in EEG-Based Brain-Computer Interface. Front Neurosci 2021; 15:733546. [PMID: 34489636 PMCID: PMC8417074 DOI: 10.3389/fnins.2021.733546] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2021] [Accepted: 07/30/2021] [Indexed: 11/26/2022] Open
Abstract
In an electroencephalogram- (EEG-) based brain–computer interface (BCI), a subject can directly communicate with an electronic device using his EEG signals in a safe and convenient way. However, the sensitivity to noise/artifact and the non-stationarity of EEG signals result in high inter-subject/session variability. Therefore, each subject usually spends long and tedious calibration time in building a subject-specific classifier. To solve this problem, we review existing signal processing approaches, including transfer learning (TL), semi-supervised learning (SSL), and a combination of TL and SSL. Cross-subject TL can transfer amounts of labeled samples from different source subjects for the target subject. Moreover, Cross-session/task/device TL can reduce the calibration time of the subject for the target session, task, or device by importing the labeled samples from the source sessions, tasks, or devices. SSL simultaneously utilizes the labeled and unlabeled samples from the target subject. The combination of TL and SSL can take advantage of each other. For each kind of signal processing approaches, we introduce their concepts and representative methods. The experimental results show that TL, SSL, and their combination can obtain good classification performance by effectively utilizing the samples available. In the end, we draw a conclusion and point to research directions in the future.
Collapse
Affiliation(s)
- Xin Huang
- Software College, Jiangxi Normal University, Nanchang, China
| | - Yilu Xu
- School of Software, Jiangxi Agricultural University, Nanchang, China
| | - Jing Hua
- School of Software, Jiangxi Agricultural University, Nanchang, China
| | - Wenlong Yi
- School of Software, Jiangxi Agricultural University, Nanchang, China
| | - Hua Yin
- School of Software, Jiangxi Agricultural University, Nanchang, China
| | - Ronghua Hu
- School of Mechatronics Engineering, Nanchang University, Nanchang, China
| | - Shiyi Wang
- Youth League Committee, Jiangxi University of Traditional Chinese Medicine, Nanchang, China
| |
Collapse
|
31
|
De Filippi E, Wolter M, Melo BRP, Tierra-Criollo CJ, Bortolini T, Deco G, Moll J. Classification of Complex Emotions Using EEG and Virtual Environment: Proof of Concept and Therapeutic Implication. Front Hum Neurosci 2021; 15:711279. [PMID: 34512297 PMCID: PMC8427812 DOI: 10.3389/fnhum.2021.711279] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2021] [Accepted: 07/29/2021] [Indexed: 11/29/2022] Open
Abstract
During the last decades, neurofeedback training for emotional self-regulation has received significant attention from scientific and clinical communities. Most studies have investigated emotions using functional magnetic resonance imaging (fMRI), including the real-time application in neurofeedback training. However, the electroencephalogram (EEG) is a more suitable tool for therapeutic application. Our study aims at establishing a method to classify discrete complex emotions (e.g., tenderness and anguish) elicited through a near-immersive scenario that can be later used for EEG-neurofeedback. EEG-based affective computing studies have mainly focused on emotion classification based on dimensions, commonly using passive elicitation through single-modality stimuli. Here, we integrated both passive and active elicitation methods. We recorded electrophysiological data during emotion-evoking trials, combining emotional self-induction with a multimodal virtual environment. We extracted correlational and time-frequency features, including frontal-alpha asymmetry (FAA), using Complex Morlet Wavelet convolution. Thinking about future real-time applications, we performed within-subject classification using 1-s windows as samples and we applied trial-specific cross-validation. We opted for a traditional machine-learning classifier with low computational complexity and sufficient validation in online settings, the Support Vector Machine. Results of individual-based cross-validation using the whole feature sets showed considerable between-subject variability. The individual accuracies ranged from 59.2 to 92.9% using time-frequency/FAA and 62.4 to 92.4% using correlational features. We found that features of the temporal, occipital, and left-frontal channels were the most discriminative between the two emotions. Our results show that the suggested pipeline is suitable for individual-based classification of discrete emotions, paving the way for future personalized EEG-neurofeedback training.
Collapse
Affiliation(s)
- Eleonora De Filippi
- Computational Neuroscience Group, Center for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Mara Wolter
- Cognitive Neuroscience and Neuroinformatics Unit, D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | - Bruno R. P. Melo
- Cognitive Neuroscience and Neuroinformatics Unit, D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
- Biomedical Engineering Program, Instituto Alberto Luiz Coimbra de Pós-Graduação e Pesquisa de Engenharia, Federal University of Rio de Janeiro, Rio de Janeiro, Brazil
| | - Carlos J. Tierra-Criollo
- Biomedical Engineering Program, Instituto Alberto Luiz Coimbra de Pós-Graduação e Pesquisa de Engenharia, Federal University of Rio de Janeiro, Rio de Janeiro, Brazil
| | - Tiago Bortolini
- Cognitive Neuroscience and Neuroinformatics Unit, D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
| | - Gustavo Deco
- Computational Neuroscience Group, Center for Brain and Cognition, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
- Institució Catalana de la Recerca i Estudis Avançats, Barcelona, Spain
- Department of Neuropsychology, Max Planck Institute for Human Cognitive and Brain Sciences, Leipzig, Germany
- Turner Institute for Brain and Mental Health, Monash University, Melbourne, VIC, Australia
| | - Jorge Moll
- Cognitive Neuroscience and Neuroinformatics Unit, D'Or Institute for Research and Education (IDOR), Rio de Janeiro, Brazil
- Scients Institute, Palo Alto, CA, United States
| |
Collapse
|
32
|
López-Hernández JL, González-Carrasco I, López-Cuadrado JL, Ruiz-Mezcua B. Framework for the Classification of Emotions in People With Visual Disabilities Through Brain Signals. Front Neuroinform 2021; 15:642766. [PMID: 34025381 PMCID: PMC8137841 DOI: 10.3389/fninf.2021.642766] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2020] [Accepted: 03/26/2021] [Indexed: 11/13/2022] Open
Abstract
Nowadays, the recognition of emotions in people with sensory disabilities still represents a challenge due to the difficulty of generalizing and modeling the set of brain signals. In recent years, the technology that has been used to study a person's behavior and emotions based on brain signals is the brain-computer interface (BCI). Although previous works have already proposed the classification of emotions in people with sensory disabilities using machine learning techniques, a model of recognition of emotions in people with visual disabilities has not yet been evaluated. Consequently, in this work, the authors present a twofold framework focused on people with visual disabilities. Firstly, auditory stimuli have been used, and a component of acquisition and extraction of brain signals has been defined. Secondly, analysis techniques for the modeling of emotions have been developed, and machine learning models for the classification of emotions have been defined. Based on the results, the algorithm with the best performance in the validation is random forest (RF), with an accuracy of 85 and 88% in the classification for negative and positive emotions, respectively. According to the results, the framework is able to classify positive and negative emotions, but the experimentation performed also shows that the framework performance depends on the number of features in the dataset and the quality of the Electroencephalogram (EEG) signals is a determining factor.
Collapse
|
33
|
Ahmad IS, Zhang S, Saminu S, Wang L, Isselmou AEK, Cai Z, Javaid I, Kamhi S, Kulsum U. Deep Learning Based on CNN for Emotion Recognition Using EEG Signal. WSEAS TRANSACTIONS ON SIGNAL PROCESSING 2021; 17:28-40. [DOI: 10.37394/232014.2021.17.4] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/02/2023]
Abstract
Emotion recognition based on brain-computer interface (BCI) has attracted important research attention despite its difficulty. It plays a vital role in human cognition and helps in making the decision. Many researchers use electroencephalograms (EEG) signals to study emotion because of its easy and convenient. Deep learning has been employed for the emotion recognition system. It recognizes emotion into single or multi-models, with visual or music stimuli shown on a screen. In this article, the convolutional neural network (CNN) model is introduced to simultaneously learn the feature and recognize the emotion of positive, neutral, and negative states of pure EEG signals single model based on the SJTU emotion EEG dataset (SEED) with ResNet50 and Adam optimizer. The dataset is shuffle, divided into training and testing, and then fed to the CNN model. The negative emotion has the highest accuracy of 94.86% fellow by neutral emotion with 94.29% and positive emotion with 93.25% respectively. With average accuracy of 94.13%. The results showed excellent classification ability of the model and can improve emotion recognition.
Collapse
Affiliation(s)
- Isah Salim Ahmad
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Shuai Zhang
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Sani Saminu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Lingyue Wang
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Abd El Kader Isselmou
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Ziliang Cai
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Imran Javaid
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Souha Kamhi
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| | - Ummay Kulsum
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, Hebei University of Technology, Tianjin, 300130, P.r. China
| |
Collapse
|