1
|
Li D, Xie L, Wang Z, Yang H. Brain Emotion Perception Inspired EEG Emotion Recognition With Deep Reinforcement Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:12979-12992. [PMID: 37126638 DOI: 10.1109/tnnls.2023.3265730] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/03/2023]
Abstract
Inspired by the well-known Papez circuit theory and neuroscience knowledge of reinforcement learning, a double dueling deep Q network (DQN) is built incorporating the electroencephalogram (EEG) signals of the frontal lobe as prior information, which is named frontal lobe double dueling DQN (FLD3QN). The framework of FLD3QN is constructed in accord with the brain emotion mechanism which takes the frontal lobe and the thalamus as the core, in which the part of the Papez circuit is simulated by the bifrontal lobe residual convolution neural network (BiFRCNN). Moreover, a step penalty factor is designed to constrain the number of mistakes of the agent. The ablation studies results on the public EEG emotion dataset DEAP verified the important roles of the frontal lobe and the Papez circuit in modeling the procedure of learning rewards during the perception of emotions, with a great increase in the average accuracies by 25.24% and 23.31% in valence and arousal dimensions.
Collapse
|
2
|
Farhadi Sedehi J, Jafarnia Dabanloo N, Maghooli K, Sheikhani A. Multimodal insights into granger causality connectivity: Integrating physiological signals and gated eye-tracking data for emotion recognition using convolutional neural network. Heliyon 2024; 10:e36411. [PMID: 39253213 PMCID: PMC11381760 DOI: 10.1016/j.heliyon.2024.e36411] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Revised: 08/14/2024] [Accepted: 08/14/2024] [Indexed: 09/11/2024] Open
Abstract
This study introduces a groundbreaking method to enhance the accuracy and reliability of emotion recognition systems by combining electrocardiogram (ECG) with electroencephalogram (EEG) data, using an eye-tracking gated strategy. Initially, we propose a technique to filter out irrelevant portions of emotional data by employing pupil diameter metrics from eye-tracking data. Subsequently, we introduce an innovative approach for estimating effective connectivity to capture the dynamic interaction between the brain and the heart during emotional states of happiness and sadness. Granger causality (GC) is estimated and utilized to optimize input for a highly effective pre-trained convolutional neural network (CNN), specifically ResNet-18. To assess this methodology, we employed EEG and ECG data from the publicly available MAHNOB-HCI database, using a 5-fold cross-validation approach. Our method achieved an impressive average accuracy and area under the curve (AUC) of 91.00 % and 0.97, respectively, for GC-EEG-ECG images processed with ResNet-18. Comparative analysis with state-of-the-art studies clearly shows that augmenting ECG with EEG and refining data with an eye-tracking strategy significantly enhances emotion recognition performance across various emotions.
Collapse
Affiliation(s)
- Javid Farhadi Sedehi
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Nader Jafarnia Dabanloo
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Keivan Maghooli
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Ali Sheikhani
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| |
Collapse
|
3
|
Guo X, Zhang L, Tian Z. Judgment Prediction Based on Tensor Decomposition With Optimized Neural Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:11116-11127. [PMID: 37028331 DOI: 10.1109/tnnls.2023.3248275] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
In the field of smart justice, handling legal cases through artificial intelligence technology is a research hotspot. Traditional judgment prediction methods are mainly based on feature models and classification algorithms. The former is difficult to describe cases from multiple angles and capture the correlation information between different case modules, while requires a wealth of legal expertise and manual labeling. The latter is unable to accurately extract the most useful information from case documents and produce fine-grained predictions. This article proposes a judgment prediction method based on tensor decomposition with optimized neural networks, which consists of OTenr, GTend, and RnEla. OTenr represents cases as normalized tensors. GTend decomposes normalized tensors into core tensors using the guidance tensor. RnEla intervenes in a case modeling process in GTend by optimizing the guidance tensor, so that core tensors represent tensor structural and elemental information, which is most conducive to improving the accuracy of judgment prediction. RnEla consists of the similarity correlation Bi-LSTM and optimized Elastic-Net regression. RnEla takes the similarity between cases as an important factor for judgment prediction. Experimental results on real legal case dataset show that the accuracy of our method is higher than that of the previous judgment prediction methods.
Collapse
|
4
|
Li C, Li P, Zhang Y, Li N, Si Y, Li F, Cao Z, Chen H, Chen B, Yao D, Xu P. Effective Emotion Recognition by Learning Discriminative Graph Topologies in EEG Brain Networks. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10258-10272. [PMID: 37022389 DOI: 10.1109/tnnls.2023.3238519] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Multichannel electroencephalogram (EEG) is an array signal that represents brain neural networks and can be applied to characterize information propagation patterns for different emotional states. To reveal these inherent spatial graph features and increase the stability of emotion recognition, we propose an effective emotion recognition model that performs multicategory emotion recognition with multiple emotion-related spatial network topology patterns (MESNPs) by learning discriminative graph topologies in EEG brain networks. To evaluate the performance of our proposed MESNP model, we conducted single-subject and multisubject four-class classification experiments on two public datasets, MAHNOB-HCI and DEAP. Compared with existing feature extraction methods, the MESNP model significantly enhances the multiclass emotional classification performance in the single-subject and multisubject conditions. To evaluate the online version of the proposed MESNP model, we designed an online emotion monitoring system. We recruited 14 participants to conduct the online emotion decoding experiments. The average online experimental accuracy of the 14 participants was 84.56%, indicating that our model can be applied in affective brain-computer interface (aBCI) systems. The offline and online experimental results demonstrate that the proposed MESNP model effectively captures discriminative graph topology patterns and significantly improves emotion classification performance. Moreover, the proposed MESNP model provides a new scheme for extracting features from strongly coupled array signals.
Collapse
|
5
|
Goshvarpour A, Goshvarpour A. Lemniscate of Bernoulli's map quantifiers: innovative measures for EEG emotion recognition. Cogn Neurodyn 2024; 18:1061-1077. [PMID: 38826652 PMCID: PMC11143135 DOI: 10.1007/s11571-023-09968-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2022] [Revised: 03/18/2023] [Accepted: 04/05/2023] [Indexed: 06/04/2024] Open
Abstract
Thanks to the advent of affective computing, designing an automatic human emotion recognition system for clinical and non-clinical applications has attracted the attention of many researchers. Currently, multi-channel electroencephalogram (EEG)-based emotion recognition is a fundamental but challenging issue. This experiment envisioned developing a new scheme for automated EEG affect recognition. An innovative nonlinear feature engineering approach was presented based on Lemniscate of Bernoulli's Map (LBM), which belongs to the family of chaotic maps, in line with the EEG's nonlinear nature. As far as the authors know, LBM has not been utilized for biological signal analysis. Next, the map was characterized using several graphical indices. The feature vector was imposed on the feature selection algorithm while evaluating the role of the feature vector dimension on emotion recognition rates. Finally, the efficiency of the features on emotion recognition was appraised using two conventional classifiers and validated using the Database for Emotion Analysis using Physiological signals (DEAP) and SJTU Emotion EEG Dataset-IV (SEED-IV) benchmark databases. The experimental results showed a maximum accuracy of 92.16% for DEAP and 90.7% for SEED-IV. Achieving higher recognition rates compared to the state-of-art EEG emotion recognition systems suggest the proposed method based on LBM could have potential both in characterizing bio-signal dynamics and detecting affect-deficit disorders.
Collapse
Affiliation(s)
- Atefeh Goshvarpour
- Department of Biomedical Engineering, Faculty of Electrical Engineering, Sahand University of Technology, Tabriz, Iran
| | - Ateke Goshvarpour
- Department of Biomedical Engineering, Imam Reza International University, Mashhad, Razavi Khorasan Iran
- Health Technology Research Center, Imam Reza International University, Mashhad, Razavi Khorasan Iran
| |
Collapse
|
6
|
Cai L, Wei X, Qing Y, Lu M, Yi G, Wang J, Dong Y. Assessment of impaired consciousness using EEG-based connectivity features and convolutional neural networks. Cogn Neurodyn 2024; 18:919-930. [PMID: 38826674 PMCID: PMC11143130 DOI: 10.1007/s11571-023-09944-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2022] [Revised: 12/18/2022] [Accepted: 02/10/2023] [Indexed: 03/05/2023] Open
Abstract
Growing electroencephalogram (EEG) studies have linked the abnormities of functional brain networks with disorders of consciousness (DOC). However, due to network data's high-dimensional and non-Euclidean properties, it is difficult to exploit the brain connectivity information that can effectively detect the consciousness levels of DOC patients via deep learning. To take maximum advantage of network information in assessing impaired consciousness, we utilized the functional connectivity with convolutional neural network (CNN) and employed three rearrangement schemes to improve the evaluation performance of brain networks. In addition, the gradient-weighted class activation mapping (Grad-CAM) was adopted to visualize the classification contributions of connections among different areas. We demonstrated that the classification performance was significantly enhanced by applying network rearrangement techniques compared to those obtained by the original connectivity matrix (with an accuracy of 75.0%). The highest classification accuracy (87.2%) was achieved by rearranging the alpha network based on the anatomical regions. The inter-region connections (i.e., frontal-parietal and frontal-occipital connectivity) played dominant roles in the classification of patients with different consciousness states. The effectiveness of functional connectivity in revealing individual differences in brain activity was further validated by the correlation between behavioral performance and connections among specific regions. These findings suggest that our proposed assessment model could detect the residual consciousness of patients.
Collapse
Affiliation(s)
- Lihui Cai
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Xile Wei
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Yang Qing
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Meili Lu
- School of Information Technology Engineering, Tianjin University of Technology and Education, Tianjin, China
| | - Guosheng Yi
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Jiang Wang
- School of Electrical and Information Engineering, Tianjin University, Tianjin, China
| | - Yueqing Dong
- Xincheng Hospital of Tianjin University, Tianjin, China
| |
Collapse
|
7
|
Nia AF, Tang V, Talou GM, Billinghurst M. Synthesizing affective neurophysiological signals using generative models: A review paper. J Neurosci Methods 2024; 406:110129. [PMID: 38614286 DOI: 10.1016/j.jneumeth.2024.110129] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 01/04/2024] [Accepted: 04/03/2024] [Indexed: 04/15/2024]
Abstract
The integration of emotional intelligence in machines is an important step in advancing human-computer interaction. This demands the development of reliable end-to-end emotion recognition systems. However, the scarcity of public affective datasets presents a challenge. In this literature review, we emphasize the use of generative models to address this issue in neurophysiological signals, particularly Electroencephalogram (EEG) and Functional Near-Infrared Spectroscopy (fNIRS). We provide a comprehensive analysis of different generative models used in the field, examining their input formulation, deployment strategies, and methodologies for evaluating the quality of synthesized data. This review serves as a comprehensive overview, offering insights into the advantages, challenges, and promising future directions in the application of generative models in emotion recognition systems. Through this review, we aim to facilitate the progression of neurophysiological data augmentation, thereby supporting the development of more efficient and reliable emotion recognition systems.
Collapse
Affiliation(s)
- Alireza F Nia
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand.
| | - Vanessa Tang
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand
| | - Gonzalo Maso Talou
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand
| | - Mark Billinghurst
- Auckland Bioengineering Institute, 70 Symonds Street, Auckland, 1010, New Zealand
| |
Collapse
|
8
|
Hamzah HA, Abdalla KK. EEG-based emotion recognition systems; comprehensive study. Heliyon 2024; 10:e31485. [PMID: 38818173 PMCID: PMC11137547 DOI: 10.1016/j.heliyon.2024.e31485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Accepted: 05/16/2024] [Indexed: 06/01/2024] Open
Abstract
Emotion recognition technology through EEG signal analysis is currently a fundamental concept in artificial intelligence. This recognition has major practical implications in emotional health care, human-computer interaction, and so on. This paper provides a comprehensive study of different methods for extracting electroencephalography (EEG) features for emotion recognition from four different perspectives, including time domain features, frequency domain features, time-frequency features, and nonlinear features. We summarize the current pattern recognition methods adopted in most related works, and with the rapid development of deep learning (DL) attracting the attention of researchers in this field, we pay more attention to deep learning-based studies and analyse the characteristics, advantages, disadvantages, and applicable scenarios. Finally, the current challenges and future development directions in this field were summarized. This paper can help novice researchers in this field gain a systematic understanding of the current status of emotion recognition research based on EEG signals and provide ideas for subsequent related research.
Collapse
Affiliation(s)
- Hussein Ali Hamzah
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| | - Kasim K. Abdalla
- Electrical Engineering Department, College of Engineering, University of Babylon, Iraq
| |
Collapse
|
9
|
Zhang J, Zhang K, An Y, Luo H, Yin S. An Integrated Multitasking Intelligent Bearing Fault Diagnosis Scheme Based on Representation Learning Under Imbalanced Sample Condition. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:6231-6242. [PMID: 37018605 DOI: 10.1109/tnnls.2022.3232147] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Accurate bearing fault diagnosis is of great significance of the safety and reliability of rotary mechanical system. In practice, the sample proportion between faulty data and healthy data in rotating mechanical system is imbalanced. Furthermore, there are commonalities between the bearing fault detection, classification, and identification tasks. Based on these observations, this article proposes a novel integrated multitasking intelligent bearing fault diagnosis scheme with the aid of representation learning under imbalanced sample condition, which realizes bearing fault detection, classification, and unknown fault identification. Specifically, in the unsupervised condition, a bearing fault detection approach based on modified denoising autoencoder (DAE) with self-attention mechanism for bottleneck layer (MDAE-SAMB) is proposed in the integrated scheme, which only uses the healthy data for training. The self-attention mechanism is introduced into the neurons in the bottleneck layer, which can assign different weights to the neurons in the bottleneck layer. Moreover, the transfer learning based on representation learning is proposed for few-shot fault classification. Only a few fault samples are used for offline training, and high-accuracy online bearing fault classification is achieved. Finally, according to the known fault data, the unknown bearing faults can be effectively identified. A bearing dataset generated by rotor dynamics experiment rig (RDER) and a public bearing dataset demonstrates the applicability of the proposed integrated fault diagnosis scheme.
Collapse
|
10
|
An P, Zhao J, Du B, Zhao W, Zhang T, Yuan Z. Amplitude-Time Dual-View Fused EEG Temporal Feature Learning for Automatic Sleep Staging. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:6492-6506. [PMID: 36215384 DOI: 10.1109/tnnls.2022.3210384] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Electroencephalogram (EEG) plays an important role in studying brain function and human cognitive performance, and the recognition of EEG signals is vital to develop an automatic sleep staging system. However, due to the complex nonstationary characteristics and the individual difference between subjects, how to obtain the effective signal features of the EEG for practical application is still a challenging task. In this article, we investigate the EEG feature learning problem and propose a novel temporal feature learning method based on amplitude-time dual-view fusion for automatic sleep staging. First, we explore the feature extraction ability of convolutional neural networks for the EEG signal from the perspective of interpretability and construct two new representation signals for the raw EEG from the views of amplitude and time. Then, we extract the amplitude-time signal features that reflect the transformation between different sleep stages from the obtained representation signals by using conventional 1-D CNNs. Furthermore, a hybrid dilation convolution module is used to learn the long-term temporal dependency features of EEG signals, which can overcome the shortcoming that the small-scale convolution kernel can only learn the local signal variation information. Finally, we conduct attention-based feature fusion for the learned dual-view signal features to further improve sleep staging performance. To evaluate the performance of the proposed method, we test 30-s-epoch EEG signal samples for healthy subjects and subjects with mild sleep disorders. The experimental results from the most commonly used datasets show that the proposed method has better sleep staging performance and has the potential for the development and application of an EEG-based automatic sleep staging system.
Collapse
|
11
|
Du X, Wang X, Zhu L, Ding X, Lv Y, Qiu S, Liu Q. Electroencephalographic Signal Data Augmentation Based on Improved Generative Adversarial Network. Brain Sci 2024; 14:367. [PMID: 38672017 PMCID: PMC11047879 DOI: 10.3390/brainsci14040367] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Revised: 04/03/2024] [Accepted: 04/08/2024] [Indexed: 04/28/2024] Open
Abstract
EEG signals combined with deep learning play an important role in the study of human-computer interaction. However, the limited dataset makes it challenging to study EEG signals using deep learning methods. Inspired by the GAN network in image generation, this paper presents an improved generative adversarial network model L-C-WGAN-GP to generate artificial EEG data to augment training sets and improve the application of BCI in various fields. The generator consists of a long short-term memory (LSTM) network and the discriminator consists of a convolutional neural network (CNN) which uses the gradient penalty-based Wasserstein distance as the loss function in model training. The model can learn the statistical features of EEG signals and generate EEG data that approximate real samples. In addition, the performance of the compressed sensing reconstruction model can be improved by using augmented datasets. Experiments show that, compared with the existing advanced data amplification techniques, the proposed model produces EEG signals closer to the real EEG signals as measured by RMSE, FD and WTD indicators. In addition, in the compressed reconstruction of EEG signals, adding the new data reduces the loss by about 15% compared with the original data, which greatly improves the reconstruction accuracy of the EEG signals' compressed sensing.
Collapse
Affiliation(s)
- Xiuli Du
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| | - Xinyue Wang
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| | - Luyao Zhu
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| | - Xiaohui Ding
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| | - Yana Lv
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| | - Shaoming Qiu
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| | - Qingli Liu
- School of Information Engineering, Dalian University, Dalian 116622, China; (X.D.); (X.W.); (L.Z.); (Y.L.); (S.Q.); (Q.L.)
- Communication and Network Laboratory, Dalian University, Dalian 116622, China
| |
Collapse
|
12
|
Chen Z, Yang R, Huang M, Li F, Lu G, Wang Z. EEGProgress: A fast and lightweight progressive convolution architecture for EEG classification. Comput Biol Med 2024; 169:107901. [PMID: 38159400 DOI: 10.1016/j.compbiomed.2023.107901] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2023] [Revised: 12/11/2023] [Accepted: 12/23/2023] [Indexed: 01/03/2024]
Abstract
Because of the intricate topological structure and connection of the human brain, extracting deep spatial features from electroencephalograph (EEG) signals is a challenging and time-consuming task. The extraction of topological spatial information plays a crucial role in EEG classification, and the architecture of the spatial convolution greatly affects the performance and complexity of convolutional neural network (CNN) based EEG classification models. In this study, a progressive convolution CNN architecture named EEGProgress is proposed, aiming to efficiently extract the topological spatial information of EEG signals from multi-scale levels (electrode, brain region, hemisphere, global) with superior speed. To achieve this, the raw EEG data is permuted using the empirical topological permutation rule, integrating the EEG data with numerous topological properties. Subsequently, the spatial features are extracted by a progressive feature extractor including prior, electrode, region, and hemisphere convolution blocks, progressively extracting the deep spatial features with reduced parameters and speed. Finally, the comparison and ablation experiments under both cross-subject and within-subject scenarios are conducted on a public dataset to verify the performance of the proposed EEGProgress and the effectiveness of the topological permutation. The results demonstrate the superior feature extraction ability of the proposed EEGProgress, with an average increase of 4.02% compared to other CNN-based EEG classification models under both cross-subject and within-subject scenarios. Furthermore, with the obtained average testing time, FLOPs, and parameters, the proposed EEGProgress outperforms other comparison models in terms of model complexity.
Collapse
Affiliation(s)
- Zhige Chen
- School of Advanced Technology, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China; School of Electrical Engineering, Electronics and Computer Science, University of Liverpool, Liverpool L69 3BX, United Kingdom
| | - Rui Yang
- School of Advanced Technology, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China
| | - Mengjie Huang
- Design School, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China.
| | - Fumin Li
- School of Advanced Technology, Xi'an Jiaotong-Liverpool University, Suzhou 215123, China; School of Electrical Engineering, Electronics and Computer Science, University of Liverpool, Liverpool L69 3BX, United Kingdom
| | - Guoping Lu
- School of Electrical Engineering, Nantong University, Nantong 226019, China
| | - Zidong Wang
- Department of Computer Science, Brunel University London, Uxbridge, Middlesex UB8 3PH, United Kingdom
| |
Collapse
|
13
|
Zhou XH, Xie XL, Liu SQ, Ni ZL, Zhou YJ, Li RQ, Gui MJ, Fan CC, Feng ZQ, Bian GB, Hou ZG. Learning Skill Characteristics From Manipulations. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:9727-9741. [PMID: 35333726 DOI: 10.1109/tnnls.2022.3160159] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Percutaneous coronary intervention (PCI) has increasingly become the main treatment for coronary artery disease. The procedure requires high experienced skills and dexterous manipulations. However, there are few techniques to model PCI skill so far. In this study, a learning framework with local and ensemble learning is proposed to learn skill characteristics of different skill-level subjects from their PCI manipulations. Ten interventional cardiologists (four experts and six novices) were recruited to deliver a medical guidewire to two target arteries on a porcine model for in vivo studies. Simultaneously, translation and twist manipulations of thumb, forefinger, and wrist are acquired with electromagnetic (EM) and fiber-optic bend (FOB) sensors, respectively. These behavior data are then processed with wavelet packet decomposition (WPD) under 1-10 levels for feature extraction. The feature vectors are further fed into three candidate individual classifiers in the local learning layer. Furthermore, the local learning results from different manipulation behaviors are fused in the ensemble learning layer with three rule-based ensemble learning algorithms. In subject-dependent skill characteristics learning, the ensemble learning can achieve 100% accuracy, significantly outperforming the best local result (90%). Furthermore, ensemble learning can also maintain 73% accuracy in subject-independent schemes. These promising results demonstrate the great potential of the proposed method to facilitate skill learning in surgical robotics and skill assessment in clinical practice.
Collapse
|
14
|
Jafari M, Shoeibi A, Khodatars M, Bagherzadeh S, Shalbaf A, García DL, Gorriz JM, Acharya UR. Emotion recognition in EEG signals using deep learning methods: A review. Comput Biol Med 2023; 165:107450. [PMID: 37708717 DOI: 10.1016/j.compbiomed.2023.107450] [Citation(s) in RCA: 9] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 08/03/2023] [Accepted: 09/01/2023] [Indexed: 09/16/2023]
Abstract
Emotions are a critical aspect of daily life and serve a crucial role in human decision-making, planning, reasoning, and other mental states. As a result, they are considered a significant factor in human interactions. Human emotions can be identified through various sources, such as facial expressions, speech, behavior (gesture/position), or physiological signals. The use of physiological signals can enhance the objectivity and reliability of emotion detection. Compared with peripheral physiological signals, electroencephalogram (EEG) recordings are directly generated by the central nervous system and are closely related to human emotions. EEG signals have the great spatial resolution that facilitates the evaluation of brain functions, making them a popular modality in emotion recognition studies. Emotion recognition using EEG signals presents several challenges, including signal variability due to electrode positioning, individual differences in signal morphology, and lack of a universal standard for EEG signal processing. Moreover, identifying the appropriate features for emotion recognition from EEG data requires further research. Finally, there is a need to develop more robust artificial intelligence (AI) including conventional machine learning (ML) and deep learning (DL) methods to handle the complex and diverse EEG signals associated with emotional states. This paper examines the application of DL techniques in emotion recognition from EEG signals and provides a detailed discussion of relevant articles. The paper explores the significant challenges in emotion recognition using EEG signals, highlights the potential of DL techniques in addressing these challenges, and suggests the scope for future research in emotion recognition using DL techniques. The paper concludes with a summary of its findings.
Collapse
Affiliation(s)
- Mahboobeh Jafari
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Afshin Shoeibi
- Data Science and Computational Intelligence Institute, University of Granada, Spain.
| | - Marjane Khodatars
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Sara Bagherzadeh
- Department of Biomedical Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Ahmad Shalbaf
- Department of Biomedical Engineering and Medical Physics, School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - David López García
- Data Science and Computational Intelligence Institute, University of Granada, Spain
| | - Juan M Gorriz
- Data Science and Computational Intelligence Institute, University of Granada, Spain; Department of Psychiatry, University of Cambridge, UK
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
15
|
Yao Q, Gu H, Wang S, Liang G, Zhao X, Li X. Exploring EEG characteristics of multi-level mental stress based on human-machine system. J Neural Eng 2023; 20:056023. [PMID: 37729925 DOI: 10.1088/1741-2552/acfbba] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2023] [Accepted: 09/20/2023] [Indexed: 09/22/2023]
Abstract
Objective.The understanding of cognitive states is important for the development of human-machine systems (HMSs), and one of the fundamental but challenging issues is the understanding and assessment of the operator's mental stress state in real task scenarios.Approach.In this paper, a virtual unmanned vehicle (UAV) driving task with multi-challenge-level was created to explore the operator's mental stress, and the human brain activity during the task was tracked in real time via electroencephalography (EEG). A mental stress analysis dataset for the virtual UAV task was then developed and used to explore the neural activation patterns associated with mental stress activity. Finally, a multiple attention-based convolutional neural network (MACN) was constructed for automatic stress assessment using the extracted stress-sensitive neural activation features.Main Results.The statistical results of EEG power spectral density (PSD) showed that frontal theta-PSD decreased with increasing task difficulty, and central beta-PSD increased with increasing task difficulty, indicating that neural patterns showed different trends under different levels of mental stress. The performance of the proposed MACN was evaluated based on the dimensional model, and results showed that average three-class classification accuracies of 89.49%/89.88% were respectively achieved for arousal/valence.Significance.The results of this paper suggest that objective assessment of mental stress in a HMS based on a virtual UAV scenario is feasible, and the proposed method provides a promising solution for cognitive computing and applications in human-machine tasks.
Collapse
Affiliation(s)
- Qunli Yao
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing 100875, People's Republic of China
| | - Heng Gu
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing 100875, People's Republic of China
| | - Shaodi Wang
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing 100875, People's Republic of China
| | - Guanhao Liang
- Center for Cognition and Neuroergonomics, Beijing Normal University, Zhuhai 519087, People's Republic of China
| | - Xiaochuan Zhao
- Institute of Computer Applied Technology of China North Industries Group Corporation Limited, Beijing 100821, People's Republic of China
| | - Xiaoli Li
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing 100875, People's Republic of China
| |
Collapse
|
16
|
Li L, Tang W, Yang H, Xue C. Classification of User Emotional Experiences on B2C Websites Utilizing Infrared Thermal Imaging. SENSORS (BASEL, SWITZERLAND) 2023; 23:7991. [PMID: 37766045 PMCID: PMC10534612 DOI: 10.3390/s23187991] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/14/2023] [Revised: 09/18/2023] [Accepted: 09/18/2023] [Indexed: 09/29/2023]
Abstract
The acquisition of physiological signals for analyzing emotional experiences has been intrusive, and potentially yields inaccurate results. This study employed infrared thermal images (IRTIs), a noninvasive technique, to classify user emotional experiences while interacting with business-to-consumer (B2C) websites. By manipulating the usability and aesthetics of B2C websites, the facial thermal images of 24 participants were captured as they engaged with the different websites. Machine learning techniques were leveraged to classify their emotional experiences, with participants' self-assessments serving as the ground truth. The findings revealed significant fluctuations in emotional valence, while the participants' arousal levels remained consistent, enabling the categorization of emotional experiences into positive and negative states. The support vector machine (SVM) model performed well in distinguishing between baseline and emotional experiences. Furthermore, this study identified key regions of interest (ROIs) and effective classification features in machine learning. These findings not only established a significant connection between user emotional experiences and IRTIs but also broadened the research perspective on the utility of IRTIs in the field of emotion analysis.
Collapse
Affiliation(s)
- Lanxin Li
- School of Mechanical Engineering, Southeast University, 2 Southeast University Road, Nanjing 211189, China; (L.L.); (W.T.)
| | - Wenzhe Tang
- School of Mechanical Engineering, Southeast University, 2 Southeast University Road, Nanjing 211189, China; (L.L.); (W.T.)
| | - Han Yang
- School of Instrument Science and Engineering, Southeast University, 2 Southeast University Road, Nanjing 211189, China;
| | - Chengqi Xue
- School of Mechanical Engineering, Southeast University, 2 Southeast University Road, Nanjing 211189, China; (L.L.); (W.T.)
| |
Collapse
|
17
|
Fu B, Gu C, Fu M, Xia Y, Liu Y. A novel feature fusion network for multimodal emotion recognition from EEG and eye movement signals. Front Neurosci 2023; 17:1234162. [PMID: 37600016 PMCID: PMC10436100 DOI: 10.3389/fnins.2023.1234162] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2023] [Accepted: 07/20/2023] [Indexed: 08/22/2023] Open
Abstract
Emotion recognition is a challenging task, and the use of multimodal fusion methods for emotion recognition has become a trend. Fusion vectors can provide a more comprehensive representation of changes in the subject's emotional state, leading to more accurate emotion recognition results. Different fusion inputs or feature fusion methods have varying effects on the final fusion outcome. In this paper, we propose a novel Multimodal Feature Fusion Neural Network model (MFFNN) that effectively extracts complementary information from eye movement signals and performs feature fusion with EEG signals. We construct a dual-branch feature extraction module to extract features from both modalities while ensuring temporal alignment. A multi-scale feature fusion module is introduced, which utilizes cross-channel soft attention to adaptively select information from different spatial scales, enabling the acquisition of features at different spatial scales for effective fusion. We conduct experiments on the publicly available SEED-IV dataset, and our model achieves an accuracy of 87.32% in recognizing four emotions (happiness, sadness, fear, and neutrality). The results demonstrate that the proposed model can better explore complementary information from EEG and eye movement signals, thereby improving accuracy, and stability in emotion recognition.
Collapse
Affiliation(s)
- Baole Fu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Chunrui Gu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Ming Fu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Yuxiao Xia
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
| | - Yinhua Liu
- School of Automation, Qingdao University, Qingdao, China
- Institute for Future, Qingdao University, Qingdao, China
- Shandong Key Laboratory of Industrial Control Technology, Qingdao, China
| |
Collapse
|
18
|
Rahman MM, Albeverio S, Kagawa T, Kawasaki S, Okai T, Oya H, Yahagi Y, Yoshida MW. High accuracy distinction of shockable and non-shockable arrhythmias in abnormal classes through wavelet transform with pseudo differential like operators. Sci Rep 2023; 13:9513. [PMID: 37308508 DOI: 10.1038/s41598-023-36463-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Accepted: 06/04/2023] [Indexed: 06/14/2023] Open
Abstract
Arrhythmia is an abnormal rhythm of the heart which leads to sudden death. Among these arrhythmias, some are shockable, and some are non-shockable arrhythmias with external defibrillation. The automated external defibrillator (AED) is used as the automated arrhythmia diagnosis system and requires an accurate and rapid decision to increase the survival rate. Therefore, a precise and quick decision by the AED has become essential in improving the survival rate. This paper presents an arrhythmia diagnosis system for the AED by engineering methods and generalized function theories. In the arrhythmia diagnosis system, the proposed wavelet transform with pseudo-differential like operators-based method effectively generates a distinguishable scalogram for the shockable and non-shockable arrhythmia in the abnormal class signals, which leads to the decision algorithm getting the best distinction. Then, a new quality parameter is introduced to get more details by quantizing the statistical features on the scalogram. Finally, design a simple AED shock and non-shock advice method by following this information to improve the precision and rapid decision. Here, an adequate topology (metric function) is adopted to the space of the scatter plot, where we can give different scales to select the best area of the scatter plot for the test sample. As a consequence, the proposed decision method gives the highest accuracy and rapid decision between shockable and non-shockable arrhythmias. The proposed arrhythmia diagnosis system increases the accuracy to 97.98%, with a gain of 11.75% compared to the conventional approach in the abnormal class signals. Therefore, the proposed method contributes an additional 11.75% possibility for increasing the survival rate. The proposed arrhythmia diagnosis system is general and could be applied to distinguish different arrhythmia-based applications. Also, each contribution could be used independently in various applications.
Collapse
Affiliation(s)
- Md Masudur Rahman
- Graduate School of Engineering, Kanagawa University, Yokohama, Japan.
| | - Sergio Albeverio
- Inst. Angewandte Mathematik, and HCM, University of Bonn, Bonn, Germany
| | - Toshinao Kagawa
- School of General Education and Management Studies, Suwa University of Science, Nagano, Japan
| | - Shuji Kawasaki
- Faculty of Science and Engineering, Iwate University, Iwate, Japan
| | - Takayuki Okai
- Faculty of Information Engineering, Tokyo City University, Tokyo, Japan
| | - Hidetoshi Oya
- Faculty of Information Engineering, Tokyo City University, Tokyo, Japan
| | - Yumi Yahagi
- Department of Information Systems, Tokyo City University, Tokyo, Japan
| | - Minoru W Yoshida
- Graduate School of Engineering, Kanagawa University, Yokohama, Japan
| |
Collapse
|
19
|
Hakim A, Golan I, Yefet S, Levy DJ. DeePay: deep learning decodes EEG to predict consumer's willingness to pay for neuromarketing. Front Hum Neurosci 2023; 17:1153413. [PMID: 37342823 PMCID: PMC10277553 DOI: 10.3389/fnhum.2023.1153413] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2023] [Accepted: 05/16/2023] [Indexed: 06/23/2023] Open
Abstract
There is an increasing demand within consumer-neuroscience (or neuromarketing) for objective neural measures to quantify consumers' subjective valuations and predict responses to marketing campaigns. However, the properties of EEG raise difficulties for these aims: small datasets, high dimensionality, elaborate manual feature extraction, intrinsic noise, and between-subject variations. We aimed to overcome these limitations by combining unique techniques of Deep Learning Networks (DLNs), while providing interpretable results for neuroscientific and decision-making insight. In this study, we developed a DLN to predict subjects' willingness to pay (WTP) based on their EEG data. In each trial, 213 subjects observed a product's image, from 72 possible products, and then reported their WTP for the product. The DLN employed EEG recordings from product observation to predict the corresponding reported WTP values. Our results showed 0.276 test root-mean-square-error and 75.09% test accuracy in predicting high vs. low WTP, surpassing other models and a manual feature extraction approach. Network visualizations provided the predictive frequencies of neural activity, their scalp distributions, and critical timepoints, shedding light on the neural mechanisms involved with evaluation. In conclusion, we show that DLNs may be the superior method to perform EEG-based predictions, to the benefit of decision-making researchers and marketing practitioners alike.
Collapse
Affiliation(s)
- Adam Hakim
- Neuroeconomics and Neuromarketing Lab, Sagol School of Neuroscience, Tel Aviv University, Tel Aviv-Yafo, Israel
| | - Itamar Golan
- Amir Globerson Research Group, Blavatnik School of Computer Science, Tel Aviv-Yafo, Israel
| | - Sharon Yefet
- Neuroeconomics and Neuromarketing Lab, Coller School of Management, Tel Aviv University, Tel Aviv-Yafo, Israel
| | - Dino J. Levy
- Neuroeconomics and Neuromarketing Lab, Sagol School of Neuroscience, Tel Aviv University, Tel Aviv-Yafo, Israel
- Neuroeconomics and Neuromarketing Lab, Coller School of Management, Tel Aviv University, Tel Aviv-Yafo, Israel
| |
Collapse
|
20
|
Goshvarpour A, Goshvarpour A. Emotion Recognition Using a Novel Granger Causality Quantifier and Combined Electrodes of EEG. Brain Sci 2023; 13:brainsci13050759. [PMID: 37239231 DOI: 10.3390/brainsci13050759] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2023] [Revised: 04/30/2023] [Accepted: 05/02/2023] [Indexed: 05/28/2023] Open
Abstract
Electroencephalogram (EEG) connectivity patterns can reflect neural correlates of emotion. However, the necessity of evaluating bulky data for multi-channel measurements increases the computational cost of the EEG network. To date, several approaches have been presented to pick the optimal cerebral channels, mainly depending on available data. Consequently, the risk of low data stability and reliability has increased by reducing the number of channels. Alternatively, this study suggests an electrode combination approach in which the brain is divided into six areas. After extracting EEG frequency bands, an innovative Granger causality-based measure was introduced to quantify brain connectivity patterns. The feature was subsequently subjected to a classification module to recognize valence-arousal dimensional emotions. A Database for Emotion Analysis Using Physiological Signals (DEAP) was used as a benchmark database to evaluate the scheme. The experimental results revealed a maximum accuracy of 89.55%. Additionally, EEG-based connectivity in the beta-frequency band was able to effectively classify dimensional emotions. In sum, combined EEG electrodes can efficiently replicate 32-channel EEG information.
Collapse
Affiliation(s)
- Atefeh Goshvarpour
- Department of Biomedical Engineering, Faculty of Electrical Engineering, Sahand University of Technology, Tabriz 51335-1996, Iran
| | - Ateke Goshvarpour
- Department of Biomedical Engineering, Imam Reza International University, Mashhad 91388-3186, Iran
| |
Collapse
|
21
|
Adhikary S, Jain K, Saha B, Chowdhury D. Optimized EEG based mood detection with signal processing and deep neural networks for brain-computer interface. Biomed Phys Eng Express 2023; 9. [PMID: 36745911 DOI: 10.1088/2057-1976/acb942] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Accepted: 02/06/2023] [Indexed: 02/08/2023]
Abstract
Electroencephalogram (EEG) is a very promising and widely implemented procedure to study brain signals and activities by amplifying and measuring the post-synaptical potential arising from electrical impulses produced by neurons and detected by specialized electrodes attached to specific points in the scalp. It can be studied for detecting brain abnormalities, headaches, and other conditions. However, there are limited studies performed to establish a smart decision-making model to identify EEG's relation with the mood of the subject. In this experiment, EEG signals of 28 healthy human subjects have been observed with consent and attempts have been made to study and recognise moods. Savitzky-Golay band-pass filtering and Independent Component Analysis have been used for data filtration.Different neural network algorithms have been implemented to analyze and classify the EEG data based on the mood of the subject. The model is further optimised by the usage of Blackman window-based Fourier Transformation and extracting the most significant frequencies for each electrode. Using these techniques, up to 96.01% detection accuracy has been obtained.
Collapse
Affiliation(s)
- Subhrangshu Adhikary
- Department of Research & Development, Spiraldevs Automation Industries Pvt. Ltd, Raignaj, Uttar Dinajpur, West Bengal-733123, India
| | - Kushal Jain
- Resident Doctor, Vardhman Mahaveer Medical College and Safdarjung Hospital, New Delhi-110029, India
| | - Biswajit Saha
- Department of Computer Science and Engineering, Dr B.C. Roy Engineering College, Durgapur, West Bengal-713206, India
| | - Deepraj Chowdhury
- Department of Electronics and Communication Engineering, International Institute of Information Technology Naya Raipur, Naya Raipur, India
| |
Collapse
|
22
|
Kumar GV, Bellary MI, Reddy TB. Prostate cancer classification with MRI using Taylor-Bird Squirrel Optimization based Deep Recurrent Neural Network. THE IMAGING SCIENCE JOURNAL 2023. [DOI: 10.1080/13682199.2023.2165242] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/11/2023]
Affiliation(s)
- Goddumarri Vijay Kumar
- Dept. of Computer Science and Technology, Sri Krishnadevaraya University, Ananthapuram, A.P., India
| | - Mohammed Ismail Bellary
- Department of Artificial Intelligence & Machine Learning, P.A. College of Engineering, Managalore, Affiliated to Visvesvaraya Technological University, Belagavi, K.A., India
| | - Thota Bhaskara Reddy
- Dept. of Computer Science and Technology, Sri Krishnadevaraya University, Ananthapuram, A.P., India
| |
Collapse
|
23
|
Khare SK, Bajaj V, Acharya UR. SchizoNET: a robust and accurate Margenau-Hill time-frequency distribution based deep neural network model for schizophrenia detection using EEG signals. Physiol Meas 2023; 44. [PMID: 36787641 DOI: 10.1088/1361-6579/acbc06] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2022] [Accepted: 02/14/2023] [Indexed: 02/16/2023]
Abstract
Objective.Schizophrenia (SZ) is a severe chronic illness characterized by delusions, cognitive dysfunctions, and hallucinations that impact feelings, behaviour, and thinking. Timely detection and treatment of SZ are necessary to avoid long-term consequences. Electroencephalogram (EEG) signals are one form of a biomarker that can reveal hidden changes in the brain during SZ. However, the EEG signals are non-stationary in nature with low amplitude. Therefore, extracting the hidden information from the EEG signals is challenging.Approach.The time-frequency domain is crucial for the automatic detection of SZ. Therefore, this paper presents the SchizoNET model combining the Margenau-Hill time-frequency distribution (MH-TFD) and convolutional neural network (CNN). The instantaneous information of EEG signals is captured in the time-frequency domain using MH-TFD. The time-frequency amplitude is converted to two-dimensional plots and fed to the developed CNN model.Results.The SchizoNET model is developed using three different validation techniques, including holdout, five-fold cross-validation, and ten-fold cross-validation techniques using three separate public SZ datasets (Dataset 1, 2, and 3). The proposed model achieved an accuracy of 97.4%, 99.74%, and 96.35% on Dataset 1 (adolescents: 45 SZ and 39 HC subjects), Dataset 2 (adults: 14 SZ and 14 HC subjects), and Dataset 3 (adults: 49 SZ and 32 HC subjects), respectively. We have also evaluated six performance parameters and the area under the curve to evaluate the performance of our developed model.Significance.The SchizoNET is robust, effective, and accurate, as it performed better than the state-of-the-art techniques. To the best of our knowledge, this is the first work to explore three publicly available EEG datasets for the automated detection of SZ. Our SchizoNET model can help neurologists detect the SZ in various scenarios.
Collapse
Affiliation(s)
- Smith K Khare
- Electrical and Computer Engineering Department, Aarhus University, Denmark
| | - Varun Bajaj
- Discipline of Electronics and Communication Engineering, Indian Institute of Information Technology, Design, and Manufacturing (IIITDM) Jabalpur, India
| | - U Rajendra Acharya
- School of Mathematics, Physics, and Computing, University of Southern Queensland, Springfield, Australia.,Department of Biomedical Engineering, School of Science and Technology, University of Social Sciences, Singapore.,Department of Biomedical Informatics and Medical Engineering, Asia University, Taiwan.,Distinguished Professor, Kumamoto University, Japan.,Adjunct Professor, University of Malaya, Malaysia
| |
Collapse
|
24
|
Cai Y, Li X, Li J. Emotion Recognition Using Different Sensors, Emotion Models, Methods and Datasets: A Comprehensive Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23052455. [PMID: 36904659 PMCID: PMC10007272 DOI: 10.3390/s23052455] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/31/2023] [Revised: 02/18/2023] [Accepted: 02/21/2023] [Indexed: 06/12/2023]
Abstract
In recent years, the rapid development of sensors and information technology has made it possible for machines to recognize and analyze human emotions. Emotion recognition is an important research direction in various fields. Human emotions have many manifestations. Therefore, emotion recognition can be realized by analyzing facial expressions, speech, behavior, or physiological signals. These signals are collected by different sensors. Correct recognition of human emotions can promote the development of affective computing. Most existing emotion recognition surveys only focus on a single sensor. Therefore, it is more important to compare different sensors or unimodality and multimodality. In this survey, we collect and review more than 200 papers on emotion recognition by literature research methods. We categorize these papers according to different innovations. These articles mainly focus on the methods and datasets used for emotion recognition with different sensors. This survey also provides application examples and developments in emotion recognition. Furthermore, this survey compares the advantages and disadvantages of different sensors for emotion recognition. The proposed survey can help researchers gain a better understanding of existing emotion recognition systems, thus facilitating the selection of suitable sensors, algorithms, and datasets.
Collapse
|
25
|
Wang Y, Liu R, Lin D, Chen D, Li P, Hu Q, Chen CLP. Coarse-to-Fine: Progressive Knowledge Transfer-Based Multitask Convolutional Neural Network for Intelligent Large-Scale Fault Diagnosis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:761-774. [PMID: 34370676 DOI: 10.1109/tnnls.2021.3100928] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
In modern industry, large-scale fault diagnosis of complex systems is emerging and becoming increasingly important. Most deep learning-based methods perform well on small number of fault diagnosis, but cannot converge to satisfactory results when handling large-scale fault diagnosis because the huge number of fault types will lead to the problems of intra/inter-class distance unbalance and poor local minima in neural networks. To address the above problems, a progressive knowledge transfer-based multitask convolutional neural network (PKT-MCNN) is proposed. First, to construct the coarse-to-fine knowledge structure intelligently, a structure learning algorithm is proposed via clustering fault types in different coarse-grained nodes. Thus, the intra/inter-class distance unbalance problem can be mitigated by spreading similar tasks into different nodes. Then, an MCNN architecture is designed to learn the coarse and fine-grained task simultaneously and extract more general fault information, thereby pushing the algorithm away from poor local minima. Last but not least, a PKT algorithm is proposed, which can not only transfer the coarse-grained knowledge to the fine-grained task and further alleviate the intra/inter-class distance unbalance in feature space, but also regulate different learning stages by adjusting the attention weight to each task progressively. To verify the effectiveness of the proposed method, a dataset of a nuclear power system with 66 fault types was collected and analyzed. The results demonstrate that the proposed method can be a promising tool for large-scale fault diagnosis.
Collapse
|
26
|
Shioiri K, Saho K. Exploration of Effective Time-Velocity Distribution for Doppler-Radar-Based Personal Gait Identification Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2023; 23:604. [PMID: 36679401 PMCID: PMC9864811 DOI: 10.3390/s23020604] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/22/2022] [Revised: 12/23/2022] [Accepted: 12/27/2022] [Indexed: 06/17/2023]
Abstract
Personal identification based on radar gait measurement is an important application of biometric technology because it enables remote and continuous identification of people, irrespective of the lighting conditions and subjects' outfits. This study explores an effective time-velocity distribution and its relevant parameters for Doppler-radar-based personal gait identification using deep learning. Most conventional studies on radar-based gait identification used a short-time Fourier transform (STFT), which is a general method to obtain time-velocity distribution for motion recognition using Doppler radar. However, the length of the window function that controls the time and velocity resolutions of the time-velocity image was empirically selected, and several other methods for calculating high-resolution time-velocity distributions were not considered. In this study, we compared four types of representative time-velocity distributions calculated from the Doppler-radar-received signals: STFT, wavelet transform, Wigner-Ville distribution, and smoothed pseudo-Wigner-Ville distribution. In addition, the identification accuracies of various parameter settings were also investigated. We observed that the optimally tuned STFT outperformed other high-resolution distributions, and a short length of the window function in the STFT process led to a reasonable accuracy; the best identification accuracy was 99% for the identification of twenty-five test subjects. These results indicate that STFT is the optimal time-velocity distribution for gait-based personal identification using the Doppler radar, although the time and velocity resolutions of the other methods were better than those of the STFT.
Collapse
Affiliation(s)
| | - Kenshi Saho
- Department of Intelligent Robotics, Toyama Prefectural University, Imizu 939-0398, Japan
| |
Collapse
|
27
|
Li R, Gao R, Suganthan P. A Decomposition-Based Hybrid Ensemble CNN Framework for Driver Fatigue Recognition. Inf Sci (N Y) 2023. [DOI: 10.1016/j.ins.2022.12.088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
|
28
|
Accurate Emotion Recognition Utilizing Extracted EEG Sources as Graph Neural Network Nodes. Cognit Comput 2022. [DOI: 10.1007/s12559-022-10077-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
|
29
|
Harimi A, Majd Y, Gharahbagh AA, Hajihashemi V, Esmaileyan Z, Machado JJM, Tavares JMRS. Classification of Heart Sounds Using Chaogram Transform and Deep Convolutional Neural Network Transfer Learning. SENSORS (BASEL, SWITZERLAND) 2022; 22:9569. [PMID: 36559937 PMCID: PMC9782852 DOI: 10.3390/s22249569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/12/2022] [Revised: 12/04/2022] [Accepted: 12/04/2022] [Indexed: 06/17/2023]
Abstract
Heart sounds convey important information regarding potential heart diseases. Currently, heart sound classification attracts many researchers from the fields of telemedicine, digital signal processing, and machine learning-among others-mainly to identify cardiac pathology as quickly as possible. This article proposes chaogram as a new transform to convert heart sound signals to colour images. In the proposed approach, the output image is, therefore, the projection of the reconstructed phase space representation of the phonocardiogram (PCG) signal on three coordinate planes. This has two major benefits: (1) it makes possible to apply deep convolutional neural networks to heart sounds and (2) it is also possible to employ a transfer learning scheme by converting a heart sound signal to an image. The performance of the proposed approach was verified on the PhysioNet dataset. Due to the imbalanced data on this dataset, it is common to assess the results quality using the average of sensitivity and specificity, which is known as score, instead of accuracy. In this study, the best results were achieved using the InceptionV3 model, which achieved a score of 88.06%.
Collapse
Affiliation(s)
- Ali Harimi
- Department of Electrical Engineering, Shahrood Branch, Islamic Azad University, Shahrood 43189-36199, Iran
| | - Yahya Majd
- School of Surveying and Built Environment, Toowoomba Campus, University of Southern Queensland (USQ), Darling Heights, QLD 4350, Australia
| | | | - Vahid Hajihashemi
- Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
| | - Zeynab Esmaileyan
- Department of Electrical Engineering, Shahrood Branch, Islamic Azad University, Shahrood 43189-36199, Iran
| | - José J. M. Machado
- Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
| | - João Manuel R. S. Tavares
- Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465 Porto, Portugal
| |
Collapse
|
30
|
Awan AW, Usman SM, Khalid S, Anwar A, Alroobaea R, Hussain S, Almotiri J, Ullah SS, Akram MU. An Ensemble Learning Method for Emotion Charting Using Multimodal Physiological Signals. SENSORS (BASEL, SWITZERLAND) 2022; 22:9480. [PMID: 36502183 PMCID: PMC9739519 DOI: 10.3390/s22239480] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/28/2022] [Revised: 11/24/2022] [Accepted: 11/28/2022] [Indexed: 06/17/2023]
Abstract
Emotion charting using multimodal signals has gained great demand for stroke-affected patients, for psychiatrists while examining patients, and for neuromarketing applications. Multimodal signals for emotion charting include electrocardiogram (ECG) signals, electroencephalogram (EEG) signals, and galvanic skin response (GSR) signals. EEG, ECG, and GSR are also known as physiological signals, which can be used for identification of human emotions. Due to the unbiased nature of physiological signals, this field has become a great motivation in recent research as physiological signals are generated autonomously from human central nervous system. Researchers have developed multiple methods for the classification of these signals for emotion detection. However, due to the non-linear nature of these signals and the inclusion of noise, while recording, accurate classification of physiological signals is a challenge for emotion charting. Valence and arousal are two important states for emotion detection; therefore, this paper presents a novel ensemble learning method based on deep learning for the classification of four different emotional states including high valence and high arousal (HVHA), low valence and low arousal (LVLA), high valence and low arousal (HVLA) and low valence high arousal (LVHA). In the proposed method, multimodal signals (EEG, ECG, and GSR) are preprocessed using bandpass filtering and independent components analysis (ICA) for noise removal in EEG signals followed by discrete wavelet transform for time domain to frequency domain conversion. Discrete wavelet transform results in spectrograms of the physiological signal and then features are extracted using stacked autoencoders from those spectrograms. A feature vector is obtained from the bottleneck layer of the autoencoder and is fed to three classifiers SVM (support vector machine), RF (random forest), and LSTM (long short-term memory) followed by majority voting as ensemble classification. The proposed system is trained and tested on the AMIGOS dataset with k-fold cross-validation. The proposed system obtained the highest accuracy of 94.5% and shows improved results of the proposed method compared with other state-of-the-art methods.
Collapse
Affiliation(s)
- Amna Waheed Awan
- Department of Computer Engineering, Bahria University, Islamabad 44000, Pakistan
| | - Syed Muhammad Usman
- Department of Creative Technologies, Faculty of Computing and AI, Air University, Islamabad 44000, Pakistan
| | - Shehzad Khalid
- Department of Computer Engineering, Bahria University, Islamabad 44000, Pakistan
| | - Aamir Anwar
- School of Computing and Engineering, The University of West London, London W5 5RF, UK
| | - Roobaea Alroobaea
- Department of Computer Science, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
| | - Saddam Hussain
- School of Digital Science, Universiti Brunei Darussalam, Jalan Tungku Link, Gadong BE1410, Brunei
| | - Jasem Almotiri
- Department of Computer Science, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
| | - Syed Sajid Ullah
- Department of Information and Communication Technology, University of Agder (UiA), N-4898 Grimstad, Norway
- Department of Electrical and Computer Engineering, Villanova University, Villanova, PA 19085, USA
| | - Muhammad Usman Akram
- College of Eletrical and Mechanical Engineering (E & ME), National University of Science and Technology (NUST), Islamabad 44000, Pakistan
| |
Collapse
|
31
|
Deniz E, Sobahi N, Omar N, Sengur A, Acharya UR. Automated robust human emotion classification system using hybrid EEG features with ICBrainDB dataset. Health Inf Sci Syst 2022; 10:31. [PMID: 36387749 PMCID: PMC9649575 DOI: 10.1007/s13755-022-00201-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2022] [Accepted: 10/23/2022] [Indexed: 11/11/2022] Open
Abstract
Emotion identification is an essential task for human-computer interaction systems. Electroencephalogram (EEG) signals have been widely used in emotion recognition. So far, there have been several EEG-based emotion recognition datasets that the researchers have used to validate their developed models. Hence, we have used a new ICBrainDB EEG dataset to classify angry, neutral, happy, and sad emotions in this work. Signal processing-based wavelet transform (WT), tunable Q-factor wavelet transform (TQWT), and image processing-based histogram of oriented gradients (HOG), local binary pattern (LBP), and convolutional neural network (CNN) features have been used extracted from the EEG signals. The WT is used to extract the rhythms from each channel of the EEG signal. The instantaneous frequency and spectral entropy are computed from each EEG rhythm signal. The average, and standard deviation of instantaneous frequency, and spectral entropy of each rhythm of the signal are the final feature vectors. The spectral entropy in each channel of the EEG signal after performing the TQWT is used to create the feature vectors in the second signal side method. Each EEG channel is transformed into time-frequency plots using the synchrosqueezed wavelet transform. Then, the feature vectors are constructed individually using windowed HOG and LBP features. Also, each channel of the EEG data is fed to a pretrained CNN to extract the features. In the feature selection process, the ReliefF feature selector is employed. Various feature classification algorithms namely, k-nearest neighbor (KNN), support vector machines, and neural networks are used for the automated classification of angry, neutral, happy, and sad emotions. Our developed model obtained an average accuracy of 90.7% using HOG features and a KNN classifier with a tenfold cross-validation strategy.
Collapse
Affiliation(s)
- Erkan Deniz
- Electrical and Electronics Engineering Department, Technology Faculty, Firat University, Elazig, Turkey
| | - Nebras Sobahi
- Department of Electrical and Computer Engineering, King Abdulaziz University, Jeddah, Saudi Arabia
| | - Naaman Omar
- Department of Information Technology Management, Administration Technical College, Duhok Polytechnic University, Duhok, Iraq
| | - Abdulkadir Sengur
- Electrical and Electronics Engineering Department, Technology Faculty, Firat University, Elazig, Turkey
| | - U. Rajendra Acharya
- Department of Electronics and Computer Engineering, Ngee Ann Polytechnic, Singapore, 599489 Singapore
- Biomedical Engineering, School of Science and Technology, SUSS University, Singapore, Singapore
- Biomedical Informatics and Medical Engineering, Asia University, Taichung, Taiwan
| |
Collapse
|
32
|
Wang Q, Guo Y, Shen Y, Tong S, Guo H. Multi-Layer Graph Attention Network for Sleep Stage Classification Based on EEG. SENSORS (BASEL, SWITZERLAND) 2022; 22:9272. [PMID: 36501974 PMCID: PMC9735886 DOI: 10.3390/s22239272] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/03/2022] [Revised: 11/24/2022] [Accepted: 11/24/2022] [Indexed: 06/17/2023]
Abstract
Graph neural networks have been successfully applied to sleep stage classification, but there are still challenges: (1) How to effectively utilize epoch information of EEG-adjacent channels owing to their different interaction effects. (2) How to extract the most representative features according to confused transitional information in confused stages. (3) How to improve classification accuracy of sleep stages compared with existing models. To address these shortcomings, we propose a multi-layer graph attention network (MGANet). Node-level attention prompts the graph attention convolution and GRU to focus on and differentiate the interaction between channels in the time-frequency domain and the spatial domain, respectively. The multi-head spatial-temporal mechanism balances the channel weights and dynamically adjusts channel features, and a multi-layer graph attention network accurately expresses the spatial sleep information. Moreover, stage-level attention is applied to easily confused sleep stages, which effectively improves the limitations of a graph convolutional network in large-scale graph sleep stages. The experimental results demonstrated classification accuracy; MF1 and Kappa reached 0.825, 0.814, and 0.775 and 0.873, 0.801, and 0.827 for the ISRUC and SHHS datasets, respectively, which showed that MGANet outperformed the state-of-the-art baselines.
Collapse
Affiliation(s)
| | - Yecai Guo
- Correspondence: ; Tel.: +86-151-8980-2968
| | | | | | | |
Collapse
|
33
|
Yang K, Tong L, Zeng Y, Lu R, Zhang R, Gao Y, Yan B. Exploration of effective electroencephalography features for the recognition of different valence emotions. Front Neurosci 2022; 16:1010951. [PMID: 36325479 PMCID: PMC9620477 DOI: 10.3389/fnins.2022.1010951] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Accepted: 09/05/2022] [Indexed: 11/22/2022] Open
Abstract
Recent studies have shown that the recognition and monitoring of different valence emotions can effectively avoid the occurrence of human errors due to the decline in cognitive ability. The quality of features directly affects emotion recognition results, so this manuscript explores the effective electroencephalography (EEG) features for the recognition of different valence emotions. First, 110 EEG features were extracted from the time domain, frequency domain, time-frequency domain, spatial domain, and brain network, including all the current mainly used features. Then, the classification performance, computing time, and important electrodes of each feature were systematically compared and analyzed on the self-built dataset involving 40 subjects and the public dataset DEAP. The experimental results show that the first-order difference, second-order difference, high-frequency power, and high-frequency differential entropy features perform better in the recognition of different valence emotions. Also, the time-domain features, especially the first-order difference features and second-order difference features, have less computing time, so they are suitable for real-time emotion recognition applications. Besides, the features extracted from the frontal, temporal, and occipital lobes are more effective than others for the recognition of different valence emotions. Especially, when the number of electrodes is reduced by 3/4, the classification accuracy of using features from 16 electrodes located in these brain regions is 91.8%, which is only about 2% lower than that of using all electrodes. The study results can provide an important reference for feature extraction and selection in emotion recognition based on EEG.
Collapse
|
34
|
Emotion recognition in EEG signals using the continuous wavelet transform and CNNs. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07843-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
35
|
Wei P. Emotional Cognitive Expression in Lacquer Colors Based on Prior Knowledge. JOURNAL OF ENVIRONMENTAL AND PUBLIC HEALTH 2022; 2022:1151676. [PMID: 36081422 PMCID: PMC9448576 DOI: 10.1155/2022/1151676] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/12/2022] [Revised: 07/20/2022] [Accepted: 07/22/2022] [Indexed: 11/18/2022]
Abstract
Since lacquer painting first appeared in the world of art, research into it has grown steadily. People have developed a keen interest in modern lacquer painting as a result of the extensive study of lacquer culture in both domestic and international academic circles. Many artists and art enthusiasts have contributed significantly to the study and research of lacquer painting and have made helpful attempts at modern lacquer painting. But it is challenging to describe the emotion that a lacquer painting's color conveys. This paper presents a decision-making framework for emotional cognitive learning based on the theory of emotional cognitive evaluation because there are relatively few researchers who have specifically studied the relationship between the creation of lacquer paintings and emotions and because there are also few research materials and documents for reference. The assessment of an emotional state is the central component of this framework. The observation module in the model framework is used to gather the emotional data that the lacquer painting expresses. The issue of emotional expression in lacquer painting is resolved by the emotional evaluation system, which combines the preprocessed information with prior knowledge to evaluate. The importance of affective cognitive expression in lacquer painting and the necessity of affective computing in the fields of machine learning and decision control is obtained on the basis of discussing the research status and content of affective cognition and affective computing. The efficiency of expression was increased by 1.3 percent as a result.
Collapse
Affiliation(s)
- Ping Wei
- Shanghai University, Shanghai Academy of Fine Arts, Shanghai 200444, China
| |
Collapse
|
36
|
Yuan Q. A Classroom Emotion Recognition Model Based on a Convolutional Neural Network Speech Emotion Algorithm. Occup Ther Int 2022; 2022:9563877. [PMID: 35912313 PMCID: PMC9282988 DOI: 10.1155/2022/9563877] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2022] [Revised: 06/13/2022] [Accepted: 06/15/2022] [Indexed: 11/23/2022] Open
Abstract
In this paper, we construct a model of convolutional neural network speech emotion algorithm, analyze the classroom identified by the neural network with a certain degree of confidence together with the school used in the dataset, find the characteristics and rules of teachers' control of classroom emotion nowadays using big data, find the parts of classroom emotion, and design a classroom emotion recognition model based on convolutional neural network speech emotion algorithm according to these characteristics. This paper will investigate the factors and patterns of teachers' emotional control in the classroom. In this paper, the existing neural network is adapted and improved, and some preprocessing is performed on the current dataset to train the network. The network used in this paper is a combination of convolutional neural network (CNN) and recurrent neural network (RNN), which takes advantage of both CNN for feature extraction and RNN for memory capability in the sequence model. This network has a good effect on both object labeling and speech recognition. For the problem of extracting emotion features of whole-sentence speech, we propose an attention mechanism-based emotion recognition algorithm for variable-length speech and design a spatiotemporal attention module for the speech emotion algorithm and a convolutional channel attention module for the CNN network to reduce the contribution of the spatiotemporal data of the speech emotion algorithm and the unimportant parts of the CNN convolutional channel feature data in the subsequent recognition by the attention mechanism. In turn, the weight of core key data and features is increased to improve the model recognition accuracy.
Collapse
Affiliation(s)
- Qinying Yuan
- Xi'an Jiaotong University, Xi'an, Shaanxi 710049, China
| |
Collapse
|
37
|
Accurate emotion recognition using Bayesian model based EEG sources as dynamic graph convolutional neural network nodes. Sci Rep 2022; 12:10282. [PMID: 35717542 PMCID: PMC9206685 DOI: 10.1038/s41598-022-14217-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2021] [Accepted: 06/02/2022] [Indexed: 11/18/2022] Open
Abstract
Due to the effect of emotions on interactions, interpretations, and decisions, automatic detection and analysis of human emotions based on EEG signals has an important role in the treatment of psychiatric diseases. However, the low spatial resolution of EEG recorders poses a challenge. In order to overcome this problem, in this paper we model each emotion by mapping from scalp sensors to brain sources using Bernoulli–Laplace-based Bayesian model. The standard low-resolution electromagnetic tomography (sLORETA) method is used to initialize the source signals in this algorithm. Finally, a dynamic graph convolutional neural network (DGCNN) is used to classify emotional EEG in which the sources of the proposed localization model are considered as the underlying graph nodes. In the proposed method, the relationships between the EEG source signals are encoded in the DGCNN adjacency matrix. Experiments on our EEG dataset recorded at the Brain-Computer Interface Research Laboratory, University of Tabriz as well as publicly available SEED and DEAP datasets show that brain source modeling by the proposed algorithm significantly improves the accuracy of emotion recognition, such that it achieve a classification accuracy of 99.25% during the classification of the two classes of positive and negative emotions. These results represent an absolute 1–2% improvement in terms of classification accuracy over subject-dependent and subject-independent scenarios over the existing approaches.
Collapse
|
38
|
|
39
|
Car engine sounds recognition based on deformable feature map residual network. Sci Rep 2022; 12:2744. [PMID: 35177780 PMCID: PMC8854583 DOI: 10.1038/s41598-022-06818-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2021] [Accepted: 02/04/2022] [Indexed: 11/17/2022] Open
Abstract
Aiming at the difficulty in extracting the features of time–frequency images for the recognition of car engine sounds, we propose a method to recognize them based on a deformable feature map residual network. A deformable feature map residual block includes offset and convolutional layers. The offset layers shift the pixels of the input feature map. The shifted feature map is superimposed on the feature map extracted by the convolutional layers through shortcut connections to concentrate the network to the sampling in the region of interest, and to transmit the information of the offset feature map to the lower network. Then, a deformable convolution residual network is designed, and the features extracted through this network are fused with the Mel frequency cepstral coefficients of car engine sounds. After recalibration by the squeeze and excitation block, the fused results are fed into the fully connected layer for classification. Experiments on a car engine sound dataset show that the accuracy of the proposed method is 84.28%. Compared with the existing state-of-the-art methods, in terms of the accuracy of recognizing car engine sounds under various operating conditions, the proposed method represents an improvement over the method based on dictionary learning and a convolutional neural network.
Collapse
|
40
|
Bahador N, Kortelainen J. Deep learning-based classification of multichannel bio-signals using directedness transfer learning. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2021.103300] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
41
|
Emotional Analysis Model for Social Hot Topics of Professional Migrant Workers. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:3812055. [PMID: 35140770 PMCID: PMC8820853 DOI: 10.1155/2022/3812055] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/23/2021] [Revised: 01/06/2022] [Accepted: 01/07/2022] [Indexed: 11/17/2022]
Abstract
Text makes up a large portion of network data because it is the vehicle for people's direct expression of emotions and opinions. How to analyze and mine these emotional text data has become a hot topic of concern in academia and industry in recent years. The online LDA (Latent Dirichlet Allocation) model is used in this paper to train the social hot topic data of professional migrant workers on the same time slice, and the subtopic evolution and intensity are obtained. The topic development is divided into four categories, and the classification model is created using SVM (Support Vector Machine). Instead of decision makers, a virtual human with sensibility and rationality is built using a hierarchical emotional cognitive model to solve multiobjective optimization problems interactively. It analyzes human body structure and emotional signals, and then combines them with visual and physiological signals to create multimodal emotional data. An example is used to demonstrate the effectiveness of the proposed model.
Collapse
|
42
|
A Dataset for Emotion Recognition Using Virtual Reality and EEG (DER-VREEG): Emotional State Classification Using Low-Cost Wearable VR-EEG Headsets. BIG DATA AND COGNITIVE COMPUTING 2022. [DOI: 10.3390/bdcc6010016] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Emotions are viewed as an important aspect of human interactions and conversations, and allow effective and logical decision making. Emotion recognition uses low-cost wearable electroencephalography (EEG) headsets to collect brainwave signals and interpret these signals to provide information on the mental state of a person, with the implementation of a virtual reality environment in different applications; the gap between human and computer interaction, as well as the understanding process, would shorten, providing an immediate response to an individual’s mental health. This study aims to use a virtual reality (VR) headset to induce four classes of emotions (happy, scared, calm, and bored), to collect brainwave samples using a low-cost wearable EEG headset, and to run popular classifiers to compare the most feasible ones that can be used for this particular setup. Firstly, we attempt to build an immersive VR database that is accessible to the public and that can potentially assist with emotion recognition studies using virtual reality stimuli. Secondly, we use a low-cost wearable EEG headset that is both compact and small, and can be attached to the scalp without any hindrance, allowing freedom of movement for participants to view their surroundings inside the immersive VR stimulus. Finally, we evaluate the emotion recognition system by using popular machine learning algorithms and compare them for both intra-subject and inter-subject classification. The results obtained here show that the prediction model for the four-class emotion classification performed well, including the more challenging inter-subject classification, with the support vector machine (SVM Class Weight kernel) obtaining 85.01% classification accuracy. This shows that using less electrode channels but with proper parameter tuning and selection features affects the performance of the classifications.
Collapse
|
43
|
López-Dorado A, Ortiz M, Satue M, Rodrigo MJ, Barea R, Sánchez-Morla EM, Cavaliere C, Rodríguez-Ascariz JM, Orduna-Hospital E, Boquete L, Garcia-Martin E. Early Diagnosis of Multiple Sclerosis Using Swept-Source Optical Coherence Tomography and Convolutional Neural Networks Trained with Data Augmentation. SENSORS (BASEL, SWITZERLAND) 2021; 22:167. [PMID: 35009710 PMCID: PMC8747672 DOI: 10.3390/s22010167] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/03/2021] [Revised: 12/21/2021] [Accepted: 12/22/2021] [Indexed: 05/07/2023]
Abstract
BACKGROUND The aim of this paper is to implement a system to facilitate the diagnosis of multiple sclerosis (MS) in its initial stages. It does so using a convolutional neural network (CNN) to classify images captured with swept-source optical coherence tomography (SS-OCT). METHODS SS-OCT images from 48 control subjects and 48 recently diagnosed MS patients have been used. These images show the thicknesses (45 × 60 points) of the following structures: complete retina, retinal nerve fiber layer, two ganglion cell layers (GCL+, GCL++) and choroid. The Cohen distance is used to identify the structures and the regions within them with greatest discriminant capacity. The original database of OCT images is augmented by a deep convolutional generative adversarial network to expand the CNN's training set. RESULTS The retinal structures with greatest discriminant capacity are the GCL++ (44.99% of image points), complete retina (26.71%) and GCL+ (22.93%). Thresholding these images and using them as inputs to a CNN comprising two convolution modules and one classification module obtains sensitivity = specificity = 1.0. CONCLUSIONS Feature pre-selection and the use of a convolutional neural network may be a promising, nonharmful, low-cost, easy-to-perform and effective means of assisting the early diagnosis of MS based on SS-OCT thickness data.
Collapse
Affiliation(s)
- Almudena López-Dorado
- Biomedical Engineering Group, Department of Electronics, University of Alcalá, 28801 Alcalá de Henares, Spain; (A.L.-D.); (R.B.); (C.C.); (J.M.R.-A.)
| | - Miguel Ortiz
- Computer Vision, Imaging and Machine Intelligence Research Group, Interdisciplinary Center for Security, Reliability and Trust (SnT), University of Luxembourg, 4365 Luxembourg, Luxembourg;
| | - María Satue
- Miguel Servet Ophthalmology Innovation and Research Group (GIMSO), Department of Ophthalmology, Aragon Institute for Health Research (IIS Aragon), Miguel Servet University Hospital, University of Zaragoza, 50018 Zaragoza, Spain; (M.S.); (M.J.R.); (E.O.-H.)
| | - María J. Rodrigo
- Miguel Servet Ophthalmology Innovation and Research Group (GIMSO), Department of Ophthalmology, Aragon Institute for Health Research (IIS Aragon), Miguel Servet University Hospital, University of Zaragoza, 50018 Zaragoza, Spain; (M.S.); (M.J.R.); (E.O.-H.)
| | - Rafael Barea
- Biomedical Engineering Group, Department of Electronics, University of Alcalá, 28801 Alcalá de Henares, Spain; (A.L.-D.); (R.B.); (C.C.); (J.M.R.-A.)
| | - Eva M. Sánchez-Morla
- Department of Psychiatry, Hospital 12 de Octubre Research Institute (i+12), 28041 Madrid, Spain;
- Faculty of Medicine, Complutense University of Madrid, 28040 Madrid, Spain
- Biomedical Research Networking Centre in Mental Health (CIBERSAM), 28029 Madrid, Spain
| | - Carlo Cavaliere
- Biomedical Engineering Group, Department of Electronics, University of Alcalá, 28801 Alcalá de Henares, Spain; (A.L.-D.); (R.B.); (C.C.); (J.M.R.-A.)
| | - José M. Rodríguez-Ascariz
- Biomedical Engineering Group, Department of Electronics, University of Alcalá, 28801 Alcalá de Henares, Spain; (A.L.-D.); (R.B.); (C.C.); (J.M.R.-A.)
| | - Elvira Orduna-Hospital
- Miguel Servet Ophthalmology Innovation and Research Group (GIMSO), Department of Ophthalmology, Aragon Institute for Health Research (IIS Aragon), Miguel Servet University Hospital, University of Zaragoza, 50018 Zaragoza, Spain; (M.S.); (M.J.R.); (E.O.-H.)
| | - Luciano Boquete
- Biomedical Engineering Group, Department of Electronics, University of Alcalá, 28801 Alcalá de Henares, Spain; (A.L.-D.); (R.B.); (C.C.); (J.M.R.-A.)
| | - Elena Garcia-Martin
- Miguel Servet Ophthalmology Innovation and Research Group (GIMSO), Department of Ophthalmology, Aragon Institute for Health Research (IIS Aragon), Miguel Servet University Hospital, University of Zaragoza, 50018 Zaragoza, Spain; (M.S.); (M.J.R.); (E.O.-H.)
| |
Collapse
|
44
|
Gonzalez H, George R, Muzaffar S, Acevedo J, Hoppner S, Mayr C, Yoo J, Fitzek F, Elfadel I. Hardware Acceleration of EEG-Based Emotion Classification Systems: A Comprehensive Survey. IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS 2021; 15:412-442. [PMID: 34125683 DOI: 10.1109/tbcas.2021.3089132] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Recent years have witnessed a growing interest in EEG-based wearable classifiers of emotions, which could enable the real-time monitoring of patients suffering from neurological disorders such as Amyotrophic Lateral Sclerosis (ALS), Autism Spectrum Disorder (ASD), or Alzheimer's. The hope is that such wearable emotion classifiers would facilitate the patients' social integration and lead to improved healthcare outcomes for them and their loved ones. Yet in spite of their direct relevance to neuro-medicine, the hardware platforms for emotion classification have yet to fill up some important gaps in their various approaches to emotion classification in a healthcare context. In this paper, we present the first hardware-focused critical review of EEG-based wearable classifiers of emotions and survey their implementation perspectives, their algorithmic foundations, and their feature extraction methodologies. We further provide a neuroscience-based analysis of current hardware accelerators of emotion classifiers and use it to map out several research opportunities, including multi-modal hardware platforms, accelerators with tightly-coupled cores operating robustly in the near/supra-threshold region, and pre-processing libraries for universal EEG-based datasets.
Collapse
|
45
|
Siva Teja G, Saurav Vara Prasad C, Venkatesham B, Sri Rama Murty K. Identification of sloshing noises using convolutional neural network. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2021; 149:3027. [PMID: 34241106 DOI: 10.1121/10.0004829] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/12/2020] [Accepted: 04/09/2021] [Indexed: 06/13/2023]
Abstract
Sloshing in fuel tanks has become a new source of noise in hybrid and high-end vehicles in the wake of reduced noise from major sources like the engine. It occurs due to the interactions of fluid inside the tank under various driving conditions of the vehicle. Interactions of fluid with the tank walls cause hit noise, and the fluid-fluid interactions cause splash noise. As the generation mechanism is different, the hit and splash noises demand different noise controlling strategies. Thus, identifying these noises during the design stage is important for implementing effective solutions in designing a quieter fuel tank. This paper presents a convolutional neural network (CNN) based methodology for the identification of sloshing noises under different conditions of fill level, excitation, baffle configuration, etc. Data for training and testing the network are collected using a reciprocating test setup, which facilitates the generation of hit and splash noises in a rectangular tank. The identification accuracy of the features learned by CNN is compared with the hand-crafted features using support vector machines. The applicability of the proposed CNN model is tested for practical scenarios like vehicle braking, where different types of sloshing noises occur in quick succession.
Collapse
Affiliation(s)
- Golla Siva Teja
- Department of Mechanical and Aerospace Engineering, Indian Institute of Technology Hyderabad, Sangareddy, Telangana 502285, India
| | - Chennuri Saurav Vara Prasad
- Department of Engineering Science, Indian Institute of Technology Hyderabad, Sangareddy, Telangana 502285, India
| | - B Venkatesham
- Department of Mechanical and Aerospace Engineering, Indian Institute of Technology Hyderabad, Sangareddy, Telangana 502285, India
| | - K Sri Rama Murty
- Department of Electrical Engineering, Indian Institute of Technology Hyderabad, Sangareddy, Telangana 502285, India
| |
Collapse
|
46
|
Nour M, Cömert Z, Polat K. A Novel Medical Diagnosis model for COVID-19 infection detection based on Deep Features and Bayesian Optimization. Appl Soft Comput 2020; 97:106580. [PMID: 32837453 PMCID: PMC7385069 DOI: 10.1016/j.asoc.2020.106580] [Citation(s) in RCA: 150] [Impact Index Per Article: 37.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2020] [Revised: 07/21/2020] [Accepted: 07/22/2020] [Indexed: 12/24/2022]
Abstract
A pneumonia of unknown causes, which was detected in Wuhan, China, and spread rapidly throughout the world, was declared as Coronavirus disease 2019 (COVID-19). Thousands of people have lost their lives to this disease. Its negative effects on public health are ongoing. In this study, an intelligence computer-aided model that can automatically detect positive COVID-19 cases is proposed to support daily clinical applications. The proposed model is based on the convolution neural network (CNN) architecture and can automatically reveal discriminative features on chest X-ray images through its convolution with rich filter families, abstraction, and weight-sharing characteristics. Contrary to the generally used transfer learning approach, the proposed deep CNN model was trained from scratch. Instead of the pre-trained CNNs, a novel serial network consisting of five convolution layers was designed. This CNN model was utilized as a deep feature extractor. The extracted deep discriminative features were used to feed the machine learning algorithms, which were k-nearest neighbor, support vector machine (SVM), and decision tree. The hyperparameters of the machine learning models were optimized using the Bayesian optimization algorithm. The experiments were conducted on a public COVID-19 radiology database. The database was divided into two parts as training and test sets with 70% and 30% rates, respectively. As a result, the most efficient results were ensured by the SVM classifier with an accuracy of 98.97%, a sensitivity of 89.39%, a specificity of 99.75%, and an F-score of 96.72%. Consequently, a cheap, fast, and reliable intelligence tool has been provided for COVID-19 infection detection. The developed model can be used to assist field specialists, physicians, and radiologists in the decision-making process. Thanks to the proposed tool, the misdiagnosis rates can be reduced, and the proposed model can be used as a retrospective evaluation tool to validate positive COVID-19 infection cases.
Collapse
Affiliation(s)
- Majid Nour
- Department of Electrical and Computer Engineering, Faculty of Engineering, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| | - Zafer Cömert
- Department of Software Engineering, Engineering Faculty, Samsun University, Samsun, Turkey
| | - Kemal Polat
- Department of Electrical and Electronics Engineering, Faculty of Engineering, Bolu Abant Izzet Baysal University, 14280, Bolu, Turkey
| |
Collapse
|
47
|
Murugappan M, Alshuaib W, Bourisly AK, Khare SK, Sruthi S, Bajaj V. Tunable Q wavelet transform based emotion classification in Parkinson's disease using Electroencephalography. PLoS One 2020; 15:e0242014. [PMID: 33211717 PMCID: PMC7676721 DOI: 10.1371/journal.pone.0242014] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2020] [Accepted: 10/25/2020] [Indexed: 12/20/2022] Open
Abstract
Parkinson's disease (PD) is a severe incurable neurological disorder. It is mostly characterized by non-motor symptoms like fatigue, dementia, anxiety, speech and communication problems, depression, and so on. Electroencephalography (EEG) play a key role in the detection of the true emotional state of a person. Various studies have been proposed for the detection of emotional impairment in PD using filtering, Fourier transforms, wavelet transforms, and non-linear methods. However, these methods require a selection of basis and are confined in terms of accuracy. In this paper, tunable Q wavelet transform (TQWT) is proposed for the classification of emotions in PD and normal controls (NC). EEG signals of six emotional states namely happiness, sadness, fear, anger, surprise, and disgust are studied. Power, entropy, and statistical moments based features are elicited from the highpass and lowpass sub-bands of TQWT. Six features selected by statistical analysis are classified with a k-nearest neighbor, probabilistic neural network, random forest, decision tree, and extreme learning machine. Three performance measures are obtained, maximum mean accuracy, sensitivity, and specificity of 96.16%, 97.59%, and 88.51% for NC and 93.88%, 96.33%, and 81.67% for PD are achieved with a probabilistic neural network. The proposed method proved to be very effective such that it classifies emotions in PD and could be used as a potential tool for diagnosing emotional impairment in hospitals.
Collapse
Affiliation(s)
- Murugappan Murugappan
- Intelligent Signal Processing Research Lab, Department of Electronics and Communication Engineering, Kuwait College of Science and Technology (A Private University), Doha, Kuwait
| | - Waleed Alshuaib
- Department of Physiology, Faculty of Medicine, Kuwait University, Kuwait City, Kuwait
| | - Ali K. Bourisly
- Department of Physiology, Faculty of Medicine, Kuwait University, Kuwait City, Kuwait
| | - Smith K. Khare
- Department of Electronics and Communication, PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur, India
| | - Sai Sruthi
- Intelligent Signal Processing Research Lab, Department of Electronics and Communication Engineering, Kuwait College of Science and Technology (A Private University), Doha, Kuwait
| | - Varun Bajaj
- Department of Electronics and Communication, PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur, India
| |
Collapse
|
48
|
Ismael AM, Alçin ÖF, Abdalla KH, Şengür A. Two-stepped majority voting for efficient EEG-based emotion classification. Brain Inform 2020; 7:9. [PMID: 32940803 PMCID: PMC7498529 DOI: 10.1186/s40708-020-00111-3] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2020] [Accepted: 09/08/2020] [Indexed: 12/24/2022] Open
Abstract
In this paper, a novel approach that is based on two-stepped majority voting is proposed for efficient EEG-based emotion classification. Emotion recognition is important for human–machine interactions. Facial features- and body gestures-based approaches have been generally proposed for emotion recognition. Recently, EEG-based approaches become more popular in emotion recognition. In the proposed approach, the raw EEG signals are initially low-pass filtered for noise removal and band-pass filters are used for rhythms extraction. For each rhythm, the best performed EEG channels are determined based on wavelet-based entropy features and fractal dimension-based features. The k-nearest neighbor (KNN) classifier is used in classification. The best five EEG channels are used in majority voting for getting the final predictions for each EEG rhythm. In the second majority voting step, the predictions from all rhythms are used to get a final prediction. The DEAP dataset is used in experiments and classification accuracy, sensitivity and specificity are used for performance evaluation metrics. The experiments are carried out to classify the emotions into two binary classes such as high valence (HV) vs low valence (LV) and high arousal (HA) vs low arousal (LA). The experiments show that 86.3% HV vs LV discrimination accuracy and 85.0% HA vs LA discrimination accuracy is obtained. The obtained results are also compared with some of the existing methods. The comparisons show that the proposed method has potential in the use of EEG-based emotion classification.
Collapse
Affiliation(s)
- Aras M Ismael
- Sulaimani Polytechnic University, Sulaymaniyah, Iraq.
| | - Ömer F Alçin
- Electrical Engineering Department, Engineering and Natural Sciences Faculty, Malatya Turgut Ozal University, 44210, Malatya, Turkey
| | | | - Abdulkadir Şengür
- Electrical-Electronics Engineering Department, Technology Faculty, Firat University, Elazig, Turkey
| |
Collapse
|
49
|
Shanmuga Priya K, Vasanthi S. Emotion classification using EEG signal for women safety application based on deep learning. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2014. [DOI: 10.3233/jifs-221825] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/03/2023]
Abstract
An emotion is a conscious logical response that varies for different situations in women’s life. These mental responses are caused by physiological, cognitive, and behavioral changes. Gender-based violence undermines the participation of women in decision-making, resulting in a decline in their quality of life. More accurate and automatic classification of women’s emotions can enhance human-computer interfaces and security in real time. There are some wearable technologies and mobile applications that claim to ensure the safety of women. However, they rely on limited social action and are ineffective at ensuring women’s safety when and where it is needed. In this work, a novel CDB-LSTM network has been proposed to accurately classify the emotions of women in seven different classes. The electroencephalogram (EEG) offers non-radioactive methods of identifying emotions. Initially, the EEG signals are preprocessed and they are converted into images via Time-Frequency Representation (TPR). A smoothed pseudo-Wigner-Ville distribution (SPWVD) is employed to convert the EEG time-domain signals into input images. Consequently, these converted images are given as input to the Convolutional Deep Belief Network (CDBN) for extracting the most relevant features. Finally, Bi-directional LSTM is used for classifying the emotions of women into seven classes namely: happy, relax, sad, fear, anxiety, anger, and stress. The proposed CDB-LSTM network preserves the high accuracy range of 97.27% in the validation phase. The proposed CDB-LSTM network improves the overall accuracy by 6.20% 32.98% 6.85% and 3.30% better than CNN-LSTM, Multi-domain feature fusion model, GCNN-LSTM and CNN with SVM and DT respectively.
Collapse
|