1
|
Zhang J, Sato W, Kawamura N, Shimokawa K, Tang B, Nakamura Y. Sensing emotional valence and arousal dynamics through automated facial action unit analysis. Sci Rep 2024; 14:19563. [PMID: 39174675 PMCID: PMC11341571 DOI: 10.1038/s41598-024-70563-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2024] [Accepted: 08/19/2024] [Indexed: 08/24/2024] Open
Abstract
Information about the concordance between dynamic emotional experiences and objective signals is practically useful. Previous studies have shown that valence dynamics can be estimated by recording electrical activity from the muscles in the brows and cheeks. However, whether facial actions based on video data and analyzed without electrodes can be used for sensing emotion dynamics remains unknown. We investigated this issue by recording video of participants' faces and obtaining dynamic valence and arousal ratings while they observed emotional films. Action units (AUs) 04 (i.e., brow lowering) and 12 (i.e., lip-corner pulling), detected through an automated analysis of the video data, were negatively and positively correlated with dynamic ratings of subjective valence, respectively. Several other AUs were also correlated with dynamic valence or arousal ratings. Random forest regression modeling, interpreted using the SHapley Additive exPlanation tool, revealed non-linear associations between the AUs and dynamic ratings of valence or arousal. These results suggest that an automated analysis of facial expression video data can be used to estimate dynamic emotional states, which could be applied in various fields including mental health diagnosis, security monitoring, and education.
Collapse
Affiliation(s)
- Junyao Zhang
- Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| | - Wataru Sato
- Psychological Process Research Team, Guardian Robot Project, RIKEN, 2-2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto, 619-0288, Japan.
| | - Naoya Kawamura
- Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| | - Koh Shimokawa
- Psychological Process Research Team, Guardian Robot Project, RIKEN, 2-2-2 Hikaridai, Seika-cho, Soraku-gun, Kyoto, 619-0288, Japan
| | - Budu Tang
- Graduate School of Informatics, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| | - Yuichi Nakamura
- Academic Center for Computing and Media Studies, Kyoto University, Yoshida-Honmachi, Sakyo, Kyoto, 606-8507, Japan
| |
Collapse
|
2
|
Medani M, Alsubai S, Min H, Dutta AK, Anjum M. Discriminant Input Processing Scheme for Self-Assisted Intelligent Healthcare Systems. Bioengineering (Basel) 2024; 11:715. [PMID: 39061797 PMCID: PMC11274065 DOI: 10.3390/bioengineering11070715] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2024] [Revised: 07/06/2024] [Accepted: 07/09/2024] [Indexed: 07/28/2024] Open
Abstract
Modern technology and analysis of emotions play a crucial role in enabling intelligent healthcare systems to provide diagnostics and self-assistance services based on observation. However, precise data predictions and computational models are critical for these systems to perform their jobs effectively. Traditionally, healthcare monitoring has been the primary emphasis. However, there were a couple of negatives, including the pattern feature generating the method's scalability and reliability, which was tested with different data sources. This paper delves into the Discriminant Input Processing Scheme (DIPS), a crucial instrument for resolving challenges. Data-segmentation-based complex processing techniques allow DIPS to merge many emotion analysis streams. The DIPS recommendation engine uses segmented data characteristics to sift through inputs from the emotion stream for patterns. The recommendation is more accurate and flexible since DIPS uses transfer learning to identify similar data across different streams. With transfer learning, this study can be sure that the previous recommendations and data properties will be available in future data streams, making the most of them. Data utilization ratio, approximation, accuracy, and false rate are some of the metrics used to assess the effectiveness of the advised approach. Self-assisted intelligent healthcare systems that use emotion-based analysis and state-of-the-art technology are crucial when managing healthcare. This study improves healthcare management's accuracy and efficiency using computational models like DIPS to guarantee accurate data forecasts and recommendations.
Collapse
Affiliation(s)
- Mohamed Medani
- Applied College of Mahail Aseer, King Khalid University, Abha 62529, Saudi Arabia
| | - Shtwai Alsubai
- Department of Computer Science, College of Computer Engineering and Sciences in Al-Kharj, Prince Sattam bin Abdulaziz University, P.O. Box 151, Al-Kharj 16278, Saudi Arabia
| | - Hong Min
- School of Computing, Gachon University, Seongnam 13120, Republic of Korea
| | - Ashit Kumar Dutta
- Department of Computer Science and Information Systems, College of Applied Sciences, AlMaarefa University, Ad Diriyah, Riyadh 13713, Saudi Arabia
| | - Mohd Anjum
- Department of Computer Engineering, Aligarh Muslim University, Aligarh 202002, India;
| |
Collapse
|
3
|
Aina J, Akinniyi O, Rahman MM, Odero-Marah V, Khalifa F. A Hybrid Learning-Architecture for Mental Disorder Detection Using Emotion Recognition. IEEE ACCESS : PRACTICAL INNOVATIONS, OPEN SOLUTIONS 2024; 12:91410-91425. [PMID: 39054996 PMCID: PMC11270886 DOI: 10.1109/access.2024.3421376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 07/27/2024]
Abstract
Mental illness has grown to become a prevalent and global health concern that affects individuals across various demographics. Timely detection and accurate diagnosis of mental disorders are crucial for effective treatment and support as late diagnosis could result in suicidal, harmful behaviors and ultimately death. To this end, the present study introduces a novel pipeline for the analysis of facial expressions, leveraging both the AffectNet and 2013 Facial Emotion Recognition (FER) datasets. Consequently, this research goes beyond traditional diagnostic methods by contributing a system capable of generating a comprehensive mental disorder dataset and concurrently predicting mental disorders based on facial emotional cues. Particularly, we introduce a hybrid architecture for mental disorder detection leveraging the state-of-the-art object detection algorithm, YOLOv8 to detect and classify visual cues associated with specific mental disorders. To achieve accurate predictions, an integrated learning architecture based on the fusion of Convolution Neural Networks (CNNs) and Visual Transformer (ViT) models is developed to form an ensemble classifier that predicts the presence of mental illness (e.g., depression, anxiety, and other mental disorder). The overall accuracy is improved to about 81% using the proposed ensemble technique. To ensure transparency and interpretability, we integrate techniques such as Gradient-weighted Class Activation Mapping (Grad-CAM) and saliency maps to highlight the regions in the input image that significantly contribute to the model's predictions thus providing healthcare professionals with a clear understanding of the features influencing the system's decisions thereby enhancing trust and more informed diagnostic process.
Collapse
Affiliation(s)
- Joseph Aina
- Electrical and Computer Engineering Department, School of Engineering, Morgan State University, Baltimore, MD 21251, USA
| | - Oluwatunmise Akinniyi
- Electrical and Computer Engineering Department, School of Engineering, Morgan State University, Baltimore, MD 21251, USA
| | - Md Mahmudur Rahman
- Department of Computer Science, School of Computer, Mathematical and Natural Sciences, Morgan State University, Baltimore, MD 21251, USA
| | - Valerie Odero-Marah
- Center for Urban Health Disparities Research and Innovation, Department of Biology, Morgan State University, Baltimore, MD 21251, USA
| | - Fahmi Khalifa
- Electrical and Computer Engineering Department, School of Engineering, Morgan State University, Baltimore, MD 21251, USA
- Electronics and Communications Engineering Department, Mansoura University, Mansoura 35516, Egypt
| |
Collapse
|
4
|
Jeong DK, Kim HG, Kim JY. Emotion Recognition Using Hierarchical Spatiotemporal Electroencephalogram Information from Local to Global Brain Regions. Bioengineering (Basel) 2023; 10:1040. [PMID: 37760143 PMCID: PMC10525488 DOI: 10.3390/bioengineering10091040] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2023] [Revised: 08/26/2023] [Accepted: 08/28/2023] [Indexed: 09/29/2023] Open
Abstract
To understand human emotional states, local activities in various regions of the cerebral cortex and the interactions among different brain regions must be considered. This paper proposes a hierarchical emotional context feature learning model that improves multichannel electroencephalography (EEG)-based emotion recognition by learning spatiotemporal EEG features from a local brain region to a global brain region. The proposed method comprises a regional brain-level encoding module, a global brain-level encoding module, and a classifier. First, multichannel EEG signals grouped into nine regions based on the functional role of the brain are input into a regional brain-level encoding module to learn local spatiotemporal information. Subsequently, the global brain-level encoding module improved emotional classification performance by integrating local spatiotemporal information from various brain regions to learn the global context features of brain regions related to emotions. Next, we applied a two-layer bidirectional gated recurrent unit (BGRU) with self-attention to the regional brain-level module and a one-layer BGRU with self-attention to the global brain-level module. Experiments were conducted using three datasets to evaluate the EEG-based emotion recognition performance of the proposed method. The results proved that the proposed method achieves superior performance by reflecting the characteristics of multichannel EEG signals better than state-of-the-art methods.
Collapse
Affiliation(s)
- Dong-Ki Jeong
- Department of Electronic Convergence Engineering, Kwangwoon University, 20 Gwangun-ro, Nowon-gu, Seoul 01897, Republic of Korea;
| | - Hyoung-Gook Kim
- Department of Electronic Convergence Engineering, Kwangwoon University, 20 Gwangun-ro, Nowon-gu, Seoul 01897, Republic of Korea;
| | - Jin-Young Kim
- Department of ICT Convergence System Engineering, Chonnam National University, 77 Yongbong-ro, Buk-gu, Gwangju 61186, Republic of Korea;
| |
Collapse
|
5
|
Davodabadi A, Daneshian B, Saati S, Razavyan S. Mathematical model and artificial intelligence for diagnosis of Alzheimer's disease. EUROPEAN PHYSICAL JOURNAL PLUS 2023; 138:474. [PMID: 37274456 PMCID: PMC10226030 DOI: 10.1140/epjp/s13360-023-04128-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/31/2023] [Accepted: 05/22/2023] [Indexed: 06/06/2023]
Abstract
Degeneration of the neurological system linked to cognitive deficits, daily living exercise clutters, and behavioral disturbing impacts may define Alzheimer's disease. Alzheimer's disease research conducted later in life focuses on describing ways for early detection of dementia, a kind of mental disorder. To tailor our care to each patient, we utilized visual cues to determine how they were feeling. We did this by outlining two approaches to diagnosing a person's mental health. Support vector machine is the first technique. Image characteristics are extracted using a fractal model for classification in this method. With this technique, the histogram of a picture is modeled after a Gaussian distribution. Classification was performed with several support vector machines kernels, and the outcomes were compared. Step two proposes using a deep convolutional neural network architecture to identify Alzheimer's-related mental disorders. According to the findings, the support vector machines approach accurately recognized over 93% of the photos tested. The deep convolutional neural network approach was one hundred percent accurate during model training, whereas the support vector machines approach achieved just 93 percent accuracy. In contrast to support vector machines accuracy of 89.3%, the deep convolutional neural network model test findings were accurate 98.8% of the time. Based on the findings reported here, the proposed deep convolutional neural network architecture may be used for diagnostic purposes involving the patient's mental state.
Collapse
Affiliation(s)
- Afsaneh Davodabadi
- Department of Mathematics, Central Tehran Branch, Islamic Azad University, Tehran, Iran
| | - Behrooz Daneshian
- Department of Mathematics, Central Tehran Branch, Islamic Azad University, Tehran, Iran
| | - Saber Saati
- Department of Mathematics, North Tehran Branch, Islamic Azad University, Tehran, Iran
| | - Shabnam Razavyan
- Department of Mathematics, South Tehran Branch, Islamic Azad University, Tehran, Iran
| |
Collapse
|
6
|
Di Luzio F, Rosato A, Panella M. A randomized deep neural network for emotion recognition with landmarks detection. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104418] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
|
7
|
Podder T, Bhattacharya D, Majumder P, Balas VE. A feature boosted deep learning method for automatic facial expression recognition. PeerJ Comput Sci 2023; 9:e1216. [PMID: 37346544 PMCID: PMC10280470 DOI: 10.7717/peerj-cs.1216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2022] [Accepted: 12/28/2022] [Indexed: 06/23/2023]
Abstract
Automatic facial expression recognition (FER) plays a crucial role in human-computer based applications such as psychiatric treatment, classroom assessment, surveillance systems, and many others. However, automatic FER is challenging in real-time environment. The traditional methods used handcrafted methods for FER but mostly failed to produce superior results in the wild environment. In this regard, a deep learning-based FER approach with minimal parameters is proposed, which gives better results for lab-controlled and wild datasets. The method uses features boosting module with skip connections which help to focus on expression-specific features. The proposed approach is applied to FER-2013 (wild dataset), JAFFE (lab-controlled), and CK+ (lab-controlled) datasets which achieve accuracy of 70.21%, 96.16%, and 96.52%. The observed experimental results demonstrate that the proposed method outperforms the other related research concerning accuracy and time.
Collapse
Affiliation(s)
- Tanusree Podder
- Department of Computer Science and Engineering, National Institute of Technology Agartala, Agartala, Tripura, India
| | - Diptendu Bhattacharya
- Department of Computer Science and Engineering, National Institute of Technology Agartala, Agartala, Tripura, India
| | - Priyanka Majumder
- Department of Basic Science and Humanities, Techno College of Engineering Agartala, Agartala, Tripura, India
| | - Valentina Emilia Balas
- Department of Automation and Applied Informatics, Aurel Vlaicu University of Arad, Arad, Romania
| |
Collapse
|
8
|
Ryumina E, Dresvyanskiy D, Karpov A. In search of a robust facial expressions recognition model: A large-scale visual cross-corpus study. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.10.013] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
9
|
CH S, Dubey SR, Ghorai M. UFKT: Unimportant filters knowledge transfer for CNN pruning. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.09.150] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
10
|
Li C, Yang M, Zhang Y, Lai KW. An Intelligent Mental Health Identification Method for College Students: A Mixed-Method Study. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:14976. [PMID: 36429697 PMCID: PMC9690277 DOI: 10.3390/ijerph192214976] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/24/2022] [Revised: 11/10/2022] [Accepted: 11/11/2022] [Indexed: 06/16/2023]
Abstract
PURPOSE Mental health assessments that combine patients' facial expressions and behaviors have been proven effective, but screening large-scale student populations for mental health problems is time-consuming and labor-intensive. This study aims to provide an efficient and accurate intelligent method for further psychological diagnosis and treatment, which combines artificial intelligence technologies to assist in evaluating the mental health problems of college students. MATERIALS AND METHODS We propose a mixed-method study of mental health assessment that combines psychological questionnaires with facial emotion analysis to comprehensively evaluate the mental health of students on a large scale. The Depression Anxiety and Stress Scale-21(DASS-21) is used for the psychological questionnaire. The facial emotion recognition model is implemented by transfer learning based on neural networks, and the model is pre-trained using FER2013 and CFEE datasets. Among them, the FER2013 dataset consists of 48 × 48-pixel face gray images, a total of 35,887 face images. The CFEE dataset contains 950,000 facial images with annotated action units (au). Using a random sampling strategy, we sent online questionnaires to 400 college students and received 374 responses, and the response rate was 93.5%. After pre-processing, 350 results were available, including 187 male and 153 female students. First, the facial emotion data of students were collected in an online questionnaire test. Then, a pre-trained model was used for emotion recognition. Finally, the online psychological questionnaire scores and the facial emotion recognition model scores were collated to give a comprehensive psychological evaluation score. RESULTS The experimental results of the facial emotion recognition model proposed to show that its classification results are broadly consistent with the mental health survey results. This model can be used to improve efficiency. In particular, the accuracy of the facial emotion recognition model proposed in this paper is higher than that of the general mental health model, which only uses the traditional single questionnaire. Furthermore, the absolute errors of this study in the three symptoms of depression, anxiety, and stress are lower than other mental health survey results and are only 0.8%, 8.1%, 3.5%, and 1.8%, respectively. CONCLUSION The mixed method combining intelligent methods and scales for mental health assessment has high recognition accuracy. Therefore, it can support efficient large-scale screening of students' psychological problems.
Collapse
Affiliation(s)
- Chong Li
- Graduate School, Xuzhou Medical University, Xuzhou 221004, China
| | - Mingzhao Yang
- Institute of Medical Information Security, Xuzhou Medical University, Xuzhou 221004, China
| | - Yongting Zhang
- Institute of Medical Information Security, Xuzhou Medical University, Xuzhou 221004, China
- Department of Biomedical Engineering, Faculty of Engineering, Universiti of Malaya, Kuala Lumpur 50603, Malaysia
| | - Khin Wee Lai
- Department of Biomedical Engineering, Faculty of Engineering, Universiti of Malaya, Kuala Lumpur 50603, Malaysia
| |
Collapse
|
11
|
Lang J, Sun X, Li J, Wang M. Multi-stage and multi-branch network with similar expressions label distribution learning for facial expression recognition. Pattern Recognit Lett 2022. [DOI: 10.1016/j.patrec.2022.09.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/31/2022]
|
12
|
Lu Z. Analysis model of college students' mental health based on online community topic mining and emotion analysis in novel coronavirus epidemic situation. Front Public Health 2022; 10:1000313. [PMID: 36187685 PMCID: PMC9516716 DOI: 10.3389/fpubh.2022.1000313] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Accepted: 08/22/2022] [Indexed: 01/26/2023] Open
Abstract
Under the epidemic situation of COVID-19, university students have different levels of anxiety, depression, and other psychological problems, and these differing levels present different challenges. Therefore, universities and relevant departments should carry out accurate psychological health education for university students. Through research, this paper found that students' psychological problems during the COVID-19 epidemic were mainly reflected in four aspects: depression, interpersonal relationship, sleep and eating disorders, and compulsive behavior. Through the discussion of family of origin, self-awareness and motivation attribution, and social pressure, this paper analyzed the causes of psychological problems. The information resources of the network are usually unstructured data, and the text information, as the most typical unstructured data, occupies a large proportion. Moreover, this text information often contains users' emotional response to major events. In this paper, a data preprocessing system is designed, and three data preprocessing rules are defined: expression data conversion rules, data deduplication rules and invalid data cleaning rules. The characteristics of online community text data are analyzed, and the text feature extraction method is selected according to its characteristics. The results of this study show that the proportion of university students with psychological problems is about 23%, which is slightly higher than the research results during the non-epidemic period. This paper suggests that college students should master methods of self-regulation, improve their levels of physical exercise, improve their physical fitness, and establish and improve their defense mechanisms to alleviate psychological conflicts and pressures.
Collapse
Affiliation(s)
- Zuqin Lu
- Department of Special Education, School of Educational Sciences, Lingnan Normal University, Zhanjiang, China,Guangdong Provincial Key Laboratory of Development and Education for Special Needs Children, Lingnan Normal University, Zhanjiang, China,*Correspondence: Zuqin Lu
| |
Collapse
|
13
|
An Assessment and Analysis Model of Psychological Health of College Students Based on Convolutional Neural Networks. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:7586918. [PMID: 35785078 PMCID: PMC9242777 DOI: 10.1155/2022/7586918] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/24/2022] [Revised: 05/23/2022] [Accepted: 06/01/2022] [Indexed: 12/31/2022]
Abstract
Psychological health assessment and psychological problem identification essentially belong to problems of pattern recognition or nonlinear classification; its system contains complex nonlinear interactions among various factors, having basic characteristics of multivariable, multilevel, and strong coupling. An important problem in the field of artificial intelligence solved by convolutional neural networks (CNN) is to simplify complex problems, minimize the number of parameters, and thus greatly improve the algorithm's performance. Therefore, CNN has outstanding advantages in establishing the assessment and analysis model of college students' psychological health. This study determined the psychological health standards of college students, selected measurement tools for college students' psychological state, elaborated the principles of psychological assessment based on text information, performed the sample set data establishment and data processing of the assessment and analysis model of psychological health, conducted network establishment, training, and simulation, carried out a case experiment and its result analysis, explored the cause analysis of college students' psychological health problems, and finally discussed the prevention and intervention of college students' psychological problems. The study results show that the input and output of the CNN-based assessment and analysis model of college students' psychological health are their evaluation data and assessment results, respectively, and the optimal hyperparameters of the model are determined through fold cross-validation analysis to improve the model's over-fitting problem. After the training is completed, the model can predict the changes in college students' psychological state in the future through the psychological test data. The CNN uses supervised machine learning method to construct an assessment and analysis model of college students' psychological health, and establishes the mapping relationship between college students' personal background and their psychological health. The network error continuously adjusts network connection weight according to gradient descent algorithm to minimize its error, so that the convolutional layer and the pooling layer can learn the optimized feature expression of the input data.
Collapse
|
14
|
Sukhavasi SB, Sukhavasi SB, Elleithy K, El-Sayed A, Elleithy A. A Hybrid Model for Driver Emotion Detection Using Feature Fusion Approach. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2022; 19:3085. [PMID: 35270777 PMCID: PMC8909976 DOI: 10.3390/ijerph19053085] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/31/2021] [Revised: 03/02/2022] [Accepted: 03/03/2022] [Indexed: 12/10/2022]
Abstract
Machine and deep learning techniques are two branches of artificial intelligence that have proven very efficient in solving advanced human problems. The automotive industry is currently using this technology to support drivers with advanced driver assistance systems. These systems can assist various functions for proper driving and estimate drivers' capability of stable driving behavior and road safety. Many studies have proved that the driver's emotions are the significant factors that manage the driver's behavior, leading to severe vehicle collisions. Therefore, continuous monitoring of drivers' emotions can help predict their behavior to avoid accidents. A novel hybrid network architecture using a deep neural network and support vector machine has been developed to predict between six and seven driver's emotions in different poses, occlusions, and illumination conditions to achieve this goal. To determine the emotions, a fusion of Gabor and LBP features has been utilized to find the features and been classified using a support vector machine classifier combined with a convolutional neural network. Our proposed model achieved better performance accuracy of 84.41%, 95.05%, 98.57%, and 98.64% for FER 2013, CK+, KDEF, and KMU-FED datasets, respectively.
Collapse
Affiliation(s)
- Suparshya Babu Sukhavasi
- Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT 06604, USA; (S.B.S.); (S.B.S.); (A.E.-S.)
| | - Susrutha Babu Sukhavasi
- Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT 06604, USA; (S.B.S.); (S.B.S.); (A.E.-S.)
| | - Khaled Elleithy
- Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT 06604, USA; (S.B.S.); (S.B.S.); (A.E.-S.)
| | - Ahmed El-Sayed
- Department of Computer Science and Engineering, University of Bridgeport, Bridgeport, CT 06604, USA; (S.B.S.); (S.B.S.); (A.E.-S.)
| | - Abdelrahman Elleithy
- Department of Computer Science, William Paterson University, Wayne, NJ 07470, USA;
| |
Collapse
|
15
|
Saurav S, Saini AK, Saini R, Singh S. Deep learning inspired intelligent embedded system for haptic rendering of facial emotions to the blind. Neural Comput Appl 2022. [DOI: 10.1007/s00521-021-06613-3] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
|
16
|
Kumari N, Bhatia R. Efficient facial emotion recognition model using deep convolutional neural network and modified joint trilateral filter. Soft comput 2022. [DOI: 10.1007/s00500-022-06804-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
17
|
A Novel deep neural network-based emotion analysis system for automatic detection of mild cognitive impairment in the elderly. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.10.038] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
18
|
Liu Y, Dai W, Fang F, Chen Y, Huang R, Wang R, Wan B. Dynamic multi-channel metric network for joint pose-aware and identity-invariant facial expression recognition. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2021.07.034] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
19
|
|
20
|
EmNet: a deep integrated convolutional neural network for facial emotion recognition in the wild. APPL INTELL 2021. [DOI: 10.1007/s10489-020-02125-0] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
21
|
Marcolin F, Vezzetti E, Monaci M. Face perception foundations for pattern recognition algorithms. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.02.074] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
|
22
|
Li H, Wang N, Yu Y, Yang X, Gao X. LBAN-IL: A novel method of high discriminative representation for facial expression recognition. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2020.12.076] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
23
|
|
24
|
New framework for person-independent facial expression recognition combining textural and shape analysis through new feature extraction approach. Inf Sci (N Y) 2021. [DOI: 10.1016/j.ins.2020.10.065] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022]
|
25
|
Abrami A, Gunzler S, Kilbane C, Ostrand R, Ho B, Cecchi G. Automated Computer Vision Assessment of Hypomimia in Parkinson Disease: Proof-of-Principle Pilot Study. J Med Internet Res 2021; 23:e21037. [PMID: 33616535 PMCID: PMC7939934 DOI: 10.2196/21037] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2020] [Revised: 07/30/2020] [Accepted: 12/18/2020] [Indexed: 01/30/2023] Open
Abstract
BACKGROUND Facial expressions require the complex coordination of 43 different facial muscles. Parkinson disease (PD) affects facial musculature leading to "hypomimia" or "masked facies." OBJECTIVE We aimed to determine whether modern computer vision techniques can be applied to detect masked facies and quantify drug states in PD. METHODS We trained a convolutional neural network on images extracted from videos of 107 self-identified people with PD, along with 1595 videos of controls, in order to detect PD hypomimia cues. This trained model was applied to clinical interviews of 35 PD patients in their on and off drug motor states, and seven journalist interviews of the actor Alan Alda obtained before and after he was diagnosed with PD. RESULTS The algorithm achieved a test set area under the receiver operating characteristic curve of 0.71 on 54 subjects to detect PD hypomimia, compared to a value of 0.75 for trained neurologists using the United Parkinson Disease Rating Scale-III Facial Expression score. Additionally, the model accuracy to classify the on and off drug states in the clinical samples was 63% (22/35), in contrast to an accuracy of 46% (16/35) when using clinical rater scores. Finally, each of Alan Alda's seven interviews were successfully classified as occurring before (versus after) his diagnosis, with 100% accuracy (7/7). CONCLUSIONS This proof-of-principle pilot study demonstrated that computer vision holds promise as a valuable tool for PD hypomimia and for monitoring a patient's motor state in an objective and noninvasive way, particularly given the increasing importance of telemedicine.
Collapse
Affiliation(s)
- Avner Abrami
- IBM Research - Computational Biology Center, Yorktown Heights, NY, United States
| | - Steven Gunzler
- Parkinson's and Movement Disorders Center, Neurological Institute, University Hospitals Cleveland Medical Center, Cleveland, OH, United States
| | - Camilla Kilbane
- Parkinson's and Movement Disorders Center, Neurological Institute, University Hospitals Cleveland Medical Center, Cleveland, OH, United States
| | - Rachel Ostrand
- IBM Research - Computational Biology Center, Yorktown Heights, NY, United States
| | - Bryan Ho
- Department of Neurology, Tufts Medical Center, Boston, MA, United States
| | - Guillermo Cecchi
- IBM Research - Computational Biology Center, Yorktown Heights, NY, United States
| |
Collapse
|
26
|
Turkoglu M. COVIDetectioNet: COVID-19 diagnosis system based on X-ray images using features selected from pre-learned deep features ensemble. APPL INTELL 2020; 51:1213-1226. [PMID: 34764550 PMCID: PMC7498308 DOI: 10.1007/s10489-020-01888-w] [Citation(s) in RCA: 68] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
Abstract
The recent novel coronavirus (also known as COVID-19) has rapidly spread worldwide, causing an infectious respiratory disease that has killed hundreds of thousands and infected millions. While test kits are used for diagnosis of the disease, the process takes time and the test kits are limited in their availability. However, the COVID-19 disease is also diagnosable using radiological images taken through lung X-rays. This process is known to be both faster and more reliable as a form of identification and diagnosis. In this regard, the current study proposes an expert-designed system called COVIDetectioNet model, which utilizes features selected from combination of deep features for diagnosis of COVID-19. For this purpose, a pretrained Convolutional Neural Network (CNN)-based AlexNet architecture that employed the transfer learning approach, was used. The effective features that were selected using the Relief feature selection algorithm from all layers of the architecture were then classified using the Support Vector Machine (SVM) method. To verify the validity of the model proposed, a total of 6092 X-ray images, classified as Normal (healthy), COVID-19, and Pneumonia, were obtained from a combination of public datasets. In the experimental results, an accuracy of 99.18% was achieved using the model proposed. The results demonstrate that the proposed COVIDetectioNet model achieved a superior level of success when compared to previous studies.
Collapse
Affiliation(s)
- Muammer Turkoglu
- Computer Engineering Department, Engineering Faculty, Bingol University, 12000 Bingol, Turkey
| |
Collapse
|
27
|
Jaramillo-Quintanar D, Cruz-Albarran IA, Guzman-Sandoval VM, Morales-Hernandez LA. Smart Sensor Based on Biofeedback to Measure Child Relaxation in Out-of-Home Care. SENSORS 2020; 20:s20154194. [PMID: 32731523 PMCID: PMC7435878 DOI: 10.3390/s20154194] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/02/2020] [Revised: 07/19/2020] [Accepted: 07/20/2020] [Indexed: 12/28/2022]
Abstract
Children from out-of-home care are a vulnerable population that faces high stress and anxiety levels due to stressful experiences, such as being abused, being raped, and violence. This problem could have negative effects on their bio-psycho-social well-being if they are not provided with comprehensive psychological treatment. Numerous methods have been developed to help them relax, but there are no current approaches for assessing the relaxation level they reach. Based on this, a novel smart sensor that can evaluate the level of relaxation a child experiences is developed in this paper. It evaluates changes in thermal biomarkers (forehead, right and left cheek, chin, and maxillary) and heart rate (HR). Then, through a k-nearest neighbors (K-NN) intelligent classifier, four possible levels of relaxation can be obtained: no-relax, low-relax, relax, and very-relax. Additionally, an application (called i-CARE) for anxiety management, which is based on biofeedback diaphragmatic breathing, guided imagery, and video games, is evaluated. After testing the developed smart sensor, an 89.7% accuracy is obtained. The smart sensor used provides a reliable measurement of relaxation levels and the i-CARE application is effective for anxiety management, both of which are focused on children exposed to out-of-home care conditions.
Collapse
Affiliation(s)
- Daniel Jaramillo-Quintanar
- Mechatronics, Engineering Faculty, Campus San Juan del Rio, University Autonomous of Queretaro, San Juan del Rio, Queretaro 76803, Mexico; (D.J.-Q.); (I.A.C.-A.)
| | - Irving A. Cruz-Albarran
- Mechatronics, Engineering Faculty, Campus San Juan del Rio, University Autonomous of Queretaro, San Juan del Rio, Queretaro 76803, Mexico; (D.J.-Q.); (I.A.C.-A.)
| | | | - Luis A. Morales-Hernandez
- Mechatronics, Engineering Faculty, Campus San Juan del Rio, University Autonomous of Queretaro, San Juan del Rio, Queretaro 76803, Mexico; (D.J.-Q.); (I.A.C.-A.)
- Correspondence:
| |
Collapse
|