1
|
Yang C, Chengzhen L, Daiyu Y, Hao T, Liang G, Jian L, Xiaoqing L, Dong W. Evaluation of comorbid psychological disorders in functional gastrointestinal disorders patients by vibraimage technology: protocol of a prospective, single-center trial. Front Med (Lausanne) 2024; 11:1452187. [PMID: 39281819 PMCID: PMC11392798 DOI: 10.3389/fmed.2024.1452187] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2024] [Accepted: 08/19/2024] [Indexed: 09/18/2024] Open
Abstract
Introduction Functional gastrointestinal disorders (FGIDs) affect over 40% of individuals globally, and impact the quality of life. A significant portion of FGIDs patients comorbids with anxiety and depression. Traditional screening tools for psychological disorders may lack comprehensiveness. Vibraimage technology currently enables non-contact, objective analysis of psychological indicators through high-frame-rate cameras and computer analysis of micro-movements. Therefore, this study aims to (1) explore the use of vibraimage technology as a non-contact objective method to assess the psychological status of FGIDs patients, comparing this technology with the Hospital Anxiety and Depression Scale (HADS) to evaluate its screening efficacy, and (2) observe the therapeutic outcomes of FGIDs patients with or without comorbid psychological disorders after the same conventional treatment. Methods This is a prospective, single-center observational trial. 276 FGIDs outpatients who visit Peking Union Medical College Hospital will be evaluated simultaneously by HADS and vibraimage technology, then to evaluate the screen efficacy of this technology. The patients will be allocated into two groups (those with or without psychological disorders). The primary endpoint is the overall rate of improvement, specifically referring to the proportion of patients who achieved Likert scores greater than or equal to 4. The secondary endpoints encompass evaluating whether there is a reduction of more than 50% in symptom evaluation scores such as IBS-SSS. Additionally, the study will assess changes in health status and quality of life using SF-36 questionnaires and the patients' satisfaction with treatment. Furthermore, psychological status will be reassessed by vibraimage technology and HADS after treatment to evaluate the effect of combined psychological factors on FGIDs treatment.
Collapse
Affiliation(s)
- Chen Yang
- State Key Laboratory of Complex Severe and Rare Diseases, Department of Gastroenterology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Lyu Chengzhen
- State Key Laboratory of Complex Severe and Rare Diseases, Department of Gastroenterology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Yang Daiyu
- Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Tang Hao
- State Key Laboratory of Complex Severe and Rare Diseases, Department of Gastroenterology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Gong Liang
- State Key Laboratory of Complex Severe and Rare Diseases, Department of Internal Medicine, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Li Jian
- Beijing Sino Voice Technology Co., Ltd., Beijing, China
| | - Li Xiaoqing
- State Key Laboratory of Complex Severe and Rare Diseases, Department of Gastroenterology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Wu Dong
- State Key Laboratory of Complex Severe and Rare Diseases, Department of Gastroenterology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
- Department of Gastroenterology, The People's Hospital of Tibetan Autonomous Region, Lhasa, China
| |
Collapse
|
2
|
Tian H, Gong W, Li W, Qian Y. PASTFNet: a paralleled attention spatio-temporal fusion network for micro-expression recognition. Med Biol Eng Comput 2024; 62:1911-1924. [PMID: 38413518 DOI: 10.1007/s11517-024-03041-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Accepted: 02/06/2024] [Indexed: 02/29/2024]
Abstract
Micro-expressions (MEs) play such an important role in predicting a person's genuine emotions, as to make micro-expression recognition such an important resea rch focus in recent years. Most recent researchers have made efforts to recognize MEs with spatial and temporal information of video clips. However, because of their short duration and subtle intensity, capturing spatio-temporal features of micro-expressions remains challenging. To effectively promote the recognition performance, this paper presents a novel paralleled dual-branch attention-based spatio-temporal fusion network (PASTFNet). We jointly extract short- and long-range spatial relationships in spatial branch. Inspired by the composite architecture of the convolutional neural network (CNN) and long short-term memory (LSTM) for temporal modeling, we propose a novel attention-based multi-scale feature fusion network (AMFNet) to encode features of sequential frames, which can learn more expressive facial-detailed features for it implements the integrated use of attention and multi-scale feature fusion, then design an aggregation block to aggregate and acquire temporal features. At last, the features learned by the above two branches are fused to accomplish expression recognition with outstanding effect. Experiments on two MER datasets (CASMEII and SAMM) show that the PASTFNet model achieves promising ME recognition performance compared with other methods.
Collapse
Affiliation(s)
- Haichen Tian
- School of Information Science and Engineering, Xinjiang University, Urumqi, China
| | - Weijun Gong
- School of Information Science and Engineering, Xinjiang University, Urumqi, China
| | - Wei Li
- School of Software, Xinjiang University, Urumqi, China
| | - Yurong Qian
- School of Information Science and Engineering, Xinjiang University, Urumqi, China.
- School of Software, Xinjiang University, Urumqi, China.
- Key Laboratory of Signal Detection and Processing in Xinjiang Uygur Autonomous Region, Urumqi, China.
| |
Collapse
|
3
|
Ahmad A, Li Z, Iqbal S, Aurangzeb M, Tariq I, Flah A, Blazek V, Prokop L. A comprehensive bibliometric survey of micro-expression recognition system based on deep learning. Heliyon 2024; 10:e27392. [PMID: 38495163 PMCID: PMC10943397 DOI: 10.1016/j.heliyon.2024.e27392] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Revised: 02/21/2024] [Accepted: 02/28/2024] [Indexed: 03/19/2024] Open
Abstract
Micro-expressions (ME) are rapidly occurring expressions that reveal the true emotions that a human being is trying to hide, cover, or suppress. These expressions, which reveal a person's actual feelings, have a broad spectrum of applications in public safety and clinical diagnosis. This study provides a comprehensive review of the area of ME recognition. A bibliometric and network analysis techniques is used to compile all the available literature related to ME recognition. A total of 735 publications from the Web of Science (WOS) and Scopus databases were evaluated from December 2012 to December 2022 using all relevant keywords. The first round of data screening produced some basic information, which was further extracted for citation, coupling, co-authorship, co-occurrence, bibliographic, and co-citation analysis. Additionally, a thematic and descriptive analysis was executed to investigate the content of prior research findings, and research techniques used in the literature. The year wise publications indicated that the published literature between 2012 and 2017 was relatively low but however by 2021, a nearly 24-fold increment made it to 154 publications. The three topmost productive journals and conferences included IEEE Transactions on Affective Computing (n = 20 publications) followed by Neurocomputing (n = 17) and Multimedia tools and applications (n = 15). Zhao G was the most proficient author with 48 publications and the top influential country was China (620 publications). Publications by citations showed that each of the authors acquired citations ranging from 100 to 1225. While publications by organizations indicated that the University of Oulu had the most published papers (n = 51). Deep learning, facial expression recognition, and emotion recognition were among the most frequently used terms. It has been discovered that ME research was primarily classified in the discipline of engineering, with more contribution from China and Malaysia comparatively.
Collapse
Affiliation(s)
- Adnan Ahmad
- Key Laboratory of Underwater Acoustic Signal Processing of Ministry of Education, School of Information Science and Engineering, Southeast University, Nanjing, 210096, China
| | - Zhao Li
- Key Laboratory of Underwater Acoustic Signal Processing of Ministry of Education, School of Information Science and Engineering, Southeast University, Nanjing, 210096, China
| | - Sheeraz Iqbal
- Department of Electrical Engineering, University of Azad Jammu and Kashmir, Muzaffarabad, 13100, AJK, Pakistan
| | - Muhammad Aurangzeb
- School of Electrical Engineering, Southeast University, Nanjing, 210096, China
| | - Irfan Tariq
- Key Laboratory of Underwater Acoustic Signal Processing of Ministry of Education, School of Information Science and Engineering, Southeast University, Nanjing, 210096, China
| | - Ayman Flah
- College of Engineering, University of Business and Technology (UBT), Jeddah, 21448, Saudi Arabia
- MEU Research Unit, Middle East University, Amman, Jordan
- The Private Higher School of Applied Sciences and Technology of Gabes, University of Gabes, Gabes, Tunisia
- National Engineering School of Gabes, University of Gabes, Gabes, 6029, Tunisia
| | - Vojtech Blazek
- ENET Centre, VSB—Technical University of Ostrava, Ostrava, Czech Republic
| | - Lukas Prokop
- ENET Centre, VSB—Technical University of Ostrava, Ostrava, Czech Republic
| |
Collapse
|
4
|
Yang H, Xie L, Pan H, Li C, Wang Z, Zhong J. Multimodal Attention Dynamic Fusion Network for Facial Micro-Expression Recognition. ENTROPY (BASEL, SWITZERLAND) 2023; 25:1246. [PMID: 37761545 PMCID: PMC10528512 DOI: 10.3390/e25091246] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/29/2023] [Revised: 08/07/2023] [Accepted: 08/15/2023] [Indexed: 09/29/2023]
Abstract
The emotional changes in facial micro-expressions are combinations of action units. The researchers have revealed that action units can be used as additional auxiliary data to improve facial micro-expression recognition. Most of the researchers attempt to fuse image features and action unit information. However, these works ignore the impact of action units on the facial image feature extraction process. Therefore, this paper proposes a local detail feature enhancement model based on a multimodal dynamic attention fusion network (MADFN) method for micro-expression recognition. This method uses a masked autoencoder based on learnable class tokens to remove local areas with low emotional expression ability in micro-expression images. Then, we utilize the action unit dynamic fusion module to fuse action unit representation to improve the potential representation ability of image features. The state-of-the-art performance of our proposed model is evaluated and verified on SMIC, CASME II, SAMM, and their combined 3DB-Combined datasets. The experimental results demonstrated that the proposed model achieved competitive performance with accuracy rates of 81.71%, 82.11%, and 77.21% on SMIC, CASME II, and SAMM datasets, respectively, that show the MADFN model can help to improve the discrimination of facial image emotional features.
Collapse
Affiliation(s)
- Hongling Yang
- Department of Computer Science, Changzhi University, Changzhi 046011, China;
| | - Lun Xie
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (L.X.); (C.L.); (Z.W.)
| | - Hang Pan
- Department of Computer Science, Changzhi University, Changzhi 046011, China;
| | - Chiqin Li
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (L.X.); (C.L.); (Z.W.)
| | - Zhiliang Wang
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; (L.X.); (C.L.); (Z.W.)
| | - Jialiang Zhong
- School of Mathematics and Computer Sciences, Nanchang University, Nanchang 330031, China;
| |
Collapse
|
5
|
Tang X. Application of Intelligent Lie Recognition Technology in Laws and Regulations Based on Occupational Mental Health Protection. Psychol Res Behav Manag 2023; 16:2943-2959. [PMID: 37554305 PMCID: PMC10404594 DOI: 10.2147/prbm.s409723] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Accepted: 07/06/2023] [Indexed: 08/10/2023] Open
Abstract
INTRODUCTION Since the reform and opening up, the social economy has developed rapidly. The competition in the employer market is fierce, which leads leaders to have strict requirements for workers, and workplace stress increases. The blind pursuit of corporate economic benefits has led to the neglect of workers' mental health. Employee retaliation against the corporate occurs frequently. The perfection of the legal system for occupational mental health protection is imminent. METHODS Based on the above questions, this study first introduces the research background, significance, and purpose in the introduction. Second, in the literature review, the current status of research is sorted out, the problems in the existing research are summarized, and the innovation points of this study are highlighted. Then, in the method section, the algorithms and models used here are introduced, including convolutional neural networks, long short-term memory networks, and the design of interview processes. Finally, the results of the questionnaire survey and the experimental test are analyzed. RESULTS (1) There is further room for optimization of intelligent lie recognition technology. (2) The employee assistance program system can effectively solve the mental health problems of employees. (3) There is a need to expand the legislative mechanism for workers' mental health protection at the legal level. DISCUSSION This study mainly explores the loopholes of occupational mental health protection under the formulation of laws and regulations. Intelligent lie recognition technology reduces workers' adverse physical and mental health risks due to work. It is dedicated to protecting workers' legitimate rights and interests from the formulation of laws and regulations.
Collapse
Affiliation(s)
- Xin Tang
- School of Law, Chongqing University, Chongqing, 400044, People’s Republic of China
| |
Collapse
|
6
|
Pan H, Yang H, Xie L, Wang Z. Multi-scale fusion visual attention network for facial micro-expression recognition. Front Neurosci 2023; 17:1216181. [PMID: 37575295 PMCID: PMC10412924 DOI: 10.3389/fnins.2023.1216181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Accepted: 06/26/2023] [Indexed: 08/15/2023] Open
Abstract
Introduction Micro-expressions are facial muscle movements that hide genuine emotions. In response to the challenge of micro-expression low-intensity, recent studies have attempted to locate localized areas of facial muscle movement. However, this ignores the feature redundancy caused by the inaccurate locating of the regions of interest. Methods This paper proposes a novel multi-scale fusion visual attention network (MFVAN), which learns multi-scale local attention weights to mask regions of redundancy features. Specifically, this model extracts the multi-scale features of the apex frame in the micro-expression video clips by convolutional neural networks. The attention mechanism focuses on the weights of local region features in the multi-scale feature maps. Then, we mask operate redundancy regions in multi-scale features and fuse local features with high attention weights for micro-expression recognition. The self-supervision and transfer learning reduce the influence of individual identity attributes and increase the robustness of multi-scale feature maps. Finally, the multi-scale classification loss, mask loss, and removing individual identity attributes loss joint to optimize the model. Results The proposed MFVAN method is evaluated on SMIC, CASME II, SAMM, and 3DB-Combined datasets that achieve state-of-the-art performance. The experimental results show that focusing on local at the multi-scale contributes to micro-expression recognition. Discussion This paper proposed MFVAN model is the first to combine image generation with visual attention mechanisms to solve the combination challenge problem of individual identity attribute interference and low-intensity facial muscle movements. Meanwhile, the MFVAN model reveal the impact of individual attributes on the localization of local ROIs. The experimental results show that a multi-scale fusion visual attention network contributes to micro-expression recognition.
Collapse
Affiliation(s)
- Hang Pan
- Department of Computer Science, Changzhi University, Changzhi, China
| | - Hongling Yang
- Department of Computer Science, Changzhi University, Changzhi, China
| | - Lun Xie
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, China
| | - Zhiliang Wang
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing, China
| |
Collapse
|
7
|
Fu C, Yang W, Chen D, Wei F. AM3F-FlowNet: Attention-Based Multi-Scale Multi-Branch Flow Network. ENTROPY (BASEL, SWITZERLAND) 2023; 25:1064. [PMID: 37510012 PMCID: PMC10378207 DOI: 10.3390/e25071064] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/27/2023] [Revised: 07/02/2023] [Accepted: 07/12/2023] [Indexed: 07/30/2023]
Abstract
Micro-expressions are the small, brief facial expression changes that humans momentarily show during emotional experiences, and their data annotation is complicated, which leads to the scarcity of micro-expression data. To extract salient and distinguishing features from a limited dataset, we propose an attention-based multi-scale, multi-modal, multi-branch flow network to thoroughly learn the motion information of micro-expressions by exploiting the attention mechanism and the complementary properties between different optical flow information. First, we extract optical flow information (horizontal optical flow, vertical optical flow, and optical strain) based on the onset and apex frames of micro-expression videos, and each branch learns one kind of optical flow information separately. Second, we propose a multi-scale fusion module to extract more prosperous and more stable feature expressions using spatial attention to focus on locally important information at each scale. Then, we design a multi-optical flow feature reweighting module to adaptively select features for each optical flow separately by channel attention. Finally, to better integrate the information of the three branches and to alleviate the problem of uneven distribution of micro-expression samples, we introduce a logarithmically adjusted prior knowledge weighting loss. This loss function weights the prediction scores of samples from different categories to mitigate the negative impact of category imbalance during the classification process. The effectiveness of the proposed model is demonstrated through extensive experiments and feature visualization on three benchmark datasets (CASMEII, SAMM, and SMIC), and its performance is comparable to that of state-of-the-art methods.
Collapse
Affiliation(s)
- Chenghao Fu
- School of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
| | - Wenzhong Yang
- School of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
- Xinjiang Key Laboratory of Multilingual Information Technology, Xinjiang University, Urumqi 830017, China
| | - Danny Chen
- School of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
| | - Fuyuan Wei
- School of Information Science and Engineering, Xinjiang University, Urumqi 830017, China
| |
Collapse
|
8
|
Li Z, Zhang Y, Xing H, Chan KL. Facial Micro-Expression Recognition Using Double-Stream 3D Convolutional Neural Network with Domain Adaptation. SENSORS (BASEL, SWITZERLAND) 2023; 23:3577. [PMID: 37050637 PMCID: PMC10098639 DOI: 10.3390/s23073577] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/14/2023] [Revised: 03/17/2023] [Accepted: 03/27/2023] [Indexed: 06/19/2023]
Abstract
Humans show micro-expressions (MEs) under some circumstances. MEs are a display of emotions that a human wants to conceal. The recognition of MEs has been applied in various fields. However, automatic ME recognition remains a challenging problem due to two major obstacles. As MEs are typically of short duration and low intensity, it is hard to extract discriminative features from ME videos. Moreover, it is tedious to collect ME data. Existing ME datasets usually contain insufficient video samples. In this paper, we propose a deep learning model, double-stream 3D convolutional neural network (DS-3DCNN), for recognizing MEs captured in video. The recognition framework contains two streams of 3D-CNN. The first extracts spatiotemporal features from the raw ME videos. The second extracts variations of the facial motions within the spatiotemporal domain. To facilitate feature extraction, the subtle motion embedded in a ME is amplified. To address the insufficient ME data, a macro-expression dataset is employed to expand the training sample size. Supervised domain adaptation is adopted in model training in order to bridge the difference between ME and macro-expression datasets. The DS-3DCNN model is evaluated on two publicly available ME datasets. The results show that the model outperforms various state-of-the-art models; in particular, the model outperformed the best model presented in MEGC2019 by more than 6%.
Collapse
Affiliation(s)
- Zhengdao Li
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong, China; (Z.L.); (H.X.)
| | - Yupei Zhang
- Centre for Intelligent Multidimensional Data Analysis Limited, Hong Kong, China;
| | - Hanwen Xing
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong, China; (Z.L.); (H.X.)
| | - Kwok-Leung Chan
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong, China; (Z.L.); (H.X.)
| |
Collapse
|
9
|
A Survey of Micro-expression Recognition Methods Based on LBP, Optical Flow and Deep Learning. Neural Process Lett 2023. [DOI: 10.1007/s11063-022-11123-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
|
10
|
Thuseethan S, Rajasegarar S, Yearwood J. Deep3DCANN: A Deep 3DCNN-ANN Framework for Spontaneous Micro-expression Recognition. Inf Sci (N Y) 2022. [DOI: 10.1016/j.ins.2022.11.113] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2022]
|
11
|
Fan X, Shahid AR, Yan H. Edge-aware motion based facial micro-expression generation with attention mechanism. Pattern Recognit Lett 2022. [DOI: 10.1016/j.patrec.2022.09.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/14/2022]
|
12
|
Zhao S, Tang H, Liu S, Zhang Y, Wang H, Xu T, Chen E, Guan C. ME-PLAN: A deep prototypical learning with local attention network for dynamic micro-expression recognition. Neural Netw 2022; 153:427-443. [DOI: 10.1016/j.neunet.2022.06.024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2021] [Revised: 05/09/2022] [Accepted: 06/20/2022] [Indexed: 10/17/2022]
|
13
|
The design of error-correcting output codes based deep forest for the micro-expression recognition. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03590-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
14
|
Deep learning-based microexpression recognition: a survey. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07157-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
15
|
Learning two groups of discriminative features for micro-expression recognition. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.12.088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|
16
|
Cai L, Li H, Dong W, Fang H. Micro-expression recognition using 3D DenseNet fused Squeeze-and-Excitation Networks. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.108594] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
17
|
Cen S, Yu Y, Yan G, Yu M, Kong Y. Micro-expression recognition based on facial action learning with muscle movement constraints. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2021. [DOI: 10.3233/jifs-202962] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
As a spontaneous facial expression, micro-expression reveals the psychological responses of human beings. However, micro-expression recognition (MER) is highly susceptible to noise interference due to the short existing time and low-intensity of facial actions. Research on facial action coding systems explores the correlation between emotional states and facial actions, which provides more discriminative features. Therefore, based on the exploration of correlation information, the goal of our work is to propose a spatiotemporal network that is robust to low-intensity muscle movements for the MER task. Firstly, a multi-scale weighted module is proposed to encode the spatial global context, which is obtained by merging features of different resolutions preserved from the backbone network. Secondly, we propose a multi-task-based facial action learning module using the constraints of the correlation between muscle movement and micro-expressions to encode local action features. Besides, a clustering constraint term is introduced to restrict the feature distribution of similar actions to improve categories’ separability in feature space. Finally, the global context and local action features are stacked as high-quality spatial descriptions to predict micro-expressions by passing through the Convolutional Long Short-Term Memory (ConvLSTM) network. The proposed method is proved to outperform other mainstream methods through comparative experiments on the SMIC, CASME-I, and CASME-II datasets.
Collapse
Affiliation(s)
- Shixin Cen
- School of Electronic and Information Engineering, Hebei University of Technology, Tianjin, P.R. China
| | - Yang Yu
- School of Artificial Intelligence, Hebei University of Technology, Tianjin, P.R. China
| | - Gang Yan
- School of Artificial Intelligence, Hebei University of Technology, Tianjin, P.R. China
| | - Ming Yu
- School of Electronic and Information Engineering, Hebei University of Technology, Tianjin, P.R. China
- School of Artificial Intelligence, Hebei University of Technology, Tianjin, P.R. China
| | - Yanlei Kong
- School of Artificial Intelligence, Hebei University of Technology, Tianjin, P.R. China
| |
Collapse
|
18
|
Review of Automatic Microexpression Recognition in the Past Decade. MACHINE LEARNING AND KNOWLEDGE EXTRACTION 2021. [DOI: 10.3390/make3020021] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Facial expressions provide important information concerning one’s emotional state. Unlike regular facial expressions, microexpressions are particular kinds of small quick facial movements, which generally last only 0.05 to 0.2 s. They reflect individuals’ subjective emotions and real psychological states more accurately than regular expressions which can be acted. However, the small range and short duration of facial movements when microexpressions happen make them challenging to recognize both by humans and machines alike. In the past decade, automatic microexpression recognition has attracted the attention of researchers in psychology, computer science, and security, amongst others. In addition, a number of specialized microexpression databases have been collected and made publicly available. The purpose of this article is to provide a comprehensive overview of the current state of the art automatic facial microexpression recognition work. To be specific, the features and learning methods used in automatic microexpression recognition, the existing microexpression data sets, the major outstanding challenges, and possible future development directions are all discussed.
Collapse
|