1
|
Zhang Y, Gao W, Yu H, Dong J, Xia Y. Artificial Intelligence-Based Facial Palsy Evaluation: A Survey. IEEE Trans Neural Syst Rehabil Eng 2024; 32:3116-3134. [PMID: 39172615 DOI: 10.1109/tnsre.2024.3447881] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/24/2024]
Abstract
Facial palsy evaluation (FPE) aims to assess facial palsy severity of patients, which plays a vital role in facial functional treatment and rehabilitation. The traditional manners of FPE are based on subjective judgment by clinicians, which may ultimately depend on individual experience. Compared with subjective and manual evaluation, objective and automated evaluation using artificial intelligence (AI) has shown great promise in improving traditional manners and recently received significant attention. The motivation of this survey paper is mainly to provide a systemic review that would guide researchers in conducting their future research work and thus make automatic FPE applicable in real-life situations. In this survey, we comprehensively review the state-of-the-art development of AI-based FPE. First, we summarize the general pipeline of FPE systems with the related background introduction. Following this pipeline, we introduce the existing public databases and give the widely used objective evaluation metrics of FPE. In addition, the preprocessing methods in FPE are described. Then, we provide an overview of selected key publications from 2008 and summarize the state-of-the-art methods of FPE that are designed based on AI techniques. Finally, we extensively discuss the current research challenges faced by FPE and provide insights about potential future directions for advancing state-of-the-art research in this field.
Collapse
|
2
|
Feng S, Zhang R, Zhang W, Yang Y, Song A, Chen J, Wang F, Xu J, Liang C, Liang X, Chen R, Liang Z. Predicting Acute Exacerbation Phenotype in Chronic Obstructive Pulmonary Disease Patients Using VGG-16 Deep Learning. Respiration 2024:1-14. [PMID: 39047695 DOI: 10.1159/000540383] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2024] [Accepted: 07/11/2024] [Indexed: 07/27/2024] Open
Abstract
INTRODUCTION Exacerbations of chronic obstructive pulmonary disease (COPD) have a significant impact on hospitalizations, morbidity, and mortality of patients. This study aimed to develop a model for predicting acute exacerbation in COPD patients (AECOPD) based on deep-learning (DL) features. METHODS We performed a retrospective study on 219 patients with COPD who underwent inspiratory and expiratory HRCT scans. By recording the acute respiratory events of the previous year, these patients were further divided into non-AECOPD group and AECOPD group according to the presence of acute exacerbation events. Sixty-nine quantitative CT (QCT) parameters of emphysema and airway were calculated by NeuLungCARE software, and 2,000 DL features were extracted by VGG-16 method. The logistic regression method was employed to identify AECOPD patients, and 29 patients of external validation cohort were used to access the robustness of the results. RESULTS The model 3-B achieved an area under the receiver operating characteristic curve (AUC) of 0.933 and 0.865 in the testing cohort and external validation cohort, respectively. Model 3-I obtained AUC of 0.895 in the testing cohort and AUC of 0.774 in the external validation cohort. Model 7-B combined clinical characteristics, QCT parameters, and DL features achieved the best performance with an AUC of 0.979 in the testing cohort and demonstrating robust predictability with an AUC of 0.932 in the external validation cohort. Likewise, model 7-I achieved an AUC of 0.938 and 0.872 in the testing cohort and external validation cohort, respectively. CONCLUSIONS DL features extracted from HRCT scans can effectively predict acute exacerbation phenotype in COPD patients.
Collapse
Affiliation(s)
- Shengchuan Feng
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, National Center for Respiratory Medicine, Department of Respiratory Medicine, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China,
| | - Ran Zhang
- Neusoft Medical Systems Co., Ltd., Shenyang, China
| | - Wenxiu Zhang
- Institute of Research and Clinical Innovations, Neusoft Medical Systems Co., Ltd., Shanghai, China
| | - Yuqiong Yang
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, National Center for Respiratory Medicine, Department of Respiratory Medicine, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Aiqi Song
- Nanshan School, Guangzhou Medical University, Guangzhou, China
| | - Jiawei Chen
- First Clinical School, Guangzhou Medical University, Guangzhou, China
| | - Fengyan Wang
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, National Center for Respiratory Medicine, Department of Respiratory Medicine, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Jiaxuan Xu
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, National Center for Respiratory Medicine, Department of Respiratory Medicine, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Cuixia Liang
- Neusoft Medical Systems Co., Ltd., Shenyang, China
| | - Xiaoyun Liang
- Institute of Research and Clinical Innovations, Neusoft Medical Systems Co., Ltd., Shanghai, China
| | - Rongchang Chen
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, National Center for Respiratory Medicine, Department of Respiratory Medicine, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| | - Zhenyu Liang
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, National Center for Respiratory Medicine, Department of Respiratory Medicine, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China
| |
Collapse
|
3
|
Lindroth H, Nalaie K, Raghu R, Ayala IN, Busch C, Bhattacharyya A, Moreno Franco P, Diedrich DA, Pickering BW, Herasevich V. Applied Artificial Intelligence in Healthcare: A Review of Computer Vision Technology Application in Hospital Settings. J Imaging 2024; 10:81. [PMID: 38667979 PMCID: PMC11050909 DOI: 10.3390/jimaging10040081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 03/08/2024] [Accepted: 03/11/2024] [Indexed: 04/28/2024] Open
Abstract
Computer vision (CV), a type of artificial intelligence (AI) that uses digital videos or a sequence of images to recognize content, has been used extensively across industries in recent years. However, in the healthcare industry, its applications are limited by factors like privacy, safety, and ethical concerns. Despite this, CV has the potential to improve patient monitoring, and system efficiencies, while reducing workload. In contrast to previous reviews, we focus on the end-user applications of CV. First, we briefly review and categorize CV applications in other industries (job enhancement, surveillance and monitoring, automation, and augmented reality). We then review the developments of CV in the hospital setting, outpatient, and community settings. The recent advances in monitoring delirium, pain and sedation, patient deterioration, mechanical ventilation, mobility, patient safety, surgical applications, quantification of workload in the hospital, and monitoring for patient events outside the hospital are highlighted. To identify opportunities for future applications, we also completed journey mapping at different system levels. Lastly, we discuss the privacy, safety, and ethical considerations associated with CV and outline processes in algorithm development and testing that limit CV expansion in healthcare. This comprehensive review highlights CV applications and ideas for its expanded use in healthcare.
Collapse
Affiliation(s)
- Heidi Lindroth
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- Center for Aging Research, Regenstrief Institute, School of Medicine, Indiana University, Indianapolis, IN 46202, USA
- Center for Health Innovation and Implementation Science, School of Medicine, Indiana University, Indianapolis, IN 46202, USA
| | - Keivan Nalaie
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Roshini Raghu
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
| | - Ivan N. Ayala
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
| | - Charles Busch
- Division of Nursing Research, Department of Nursing, Mayo Clinic, Rochester, MN 55905, USA; (K.N.); (R.R.); (I.N.A.); (C.B.)
- College of Engineering, University of Wisconsin-Madison, Madison, WI 53705, USA
| | | | - Pablo Moreno Franco
- Department of Transplantation Medicine, Mayo Clinic, Jacksonville, FL 32224, USA
| | - Daniel A. Diedrich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Brian W. Pickering
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, Rochester, MN 55905, USA; (D.A.D.); (B.W.P.); (V.H.)
| |
Collapse
|
4
|
Rytivaara R, Näpänkangas R, Kainulainen T, Sipola A, Kallio-Pulkkinen S, Raustia A, Thevenot J. Thermographic findings related to facial pain - a survey of 40 subjects. Cranio 2024; 42:69-76. [PMID: 33689590 DOI: 10.1080/08869634.2021.1894859] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
OBJECTIVE The purpose of this study was to assess how thermography findings relate painful symptoms and signs of temporomandibular disorders (TMD). METHODS Thermography, combined with chewing of paraffin wax, was performed on 40 subjects. The results were analyzed according to gender and pain-related TMD symptoms and clinical signs. RESULTS The overall temperatures after chewing were higher in TMD patients than in controls. For females, the most significant findings were the thermal increase between the relaxed state and subjects' state after chewing in temporal and temporomandibular joint (TMJ) regions. For males, all calculated parameters demonstrated a poor ability to discriminate TMD from controls. CONCLUSION Thermography could be a potential tool in diagnostics of female TMD patients. The results suggest that the thermal information assessed in specific facial areas could help to discriminate TMD patients from non-TMD patients and could be used to quantify the pain associated with TMD.
Collapse
Affiliation(s)
- Riina Rytivaara
- Department of Dental Imaging, Oulu University Hospital, Oulu, Finland
- Medical Research Center, Oulu University Hospital and University of Oulu, Oulu, Finland
| | - Ritva Näpänkangas
- Medical Research Center, Oulu University Hospital and University of Oulu, Oulu, Finland
- Research Unit of Oral Health Sciences, Faculty of Medicine, University of Oulu, Oulu, Finland
| | - Tiina Kainulainen
- Research Unit of Oral Health Sciences, Faculty of Medicine, University of Oulu, Oulu, Finland
| | - Annina Sipola
- Department of Dental Imaging, Oulu University Hospital, Oulu, Finland
- Medical Research Center, Oulu University Hospital and University of Oulu, Oulu, Finland
- Research Unit of Oral Health Sciences, Faculty of Medicine, University of Oulu, Oulu, Finland
| | - Soili Kallio-Pulkkinen
- Department of Dental Imaging, Oulu University Hospital, Oulu, Finland
- Medical Research Center, Oulu University Hospital and University of Oulu, Oulu, Finland
| | - Aune Raustia
- Medical Research Center, Oulu University Hospital and University of Oulu, Oulu, Finland
- Research Unit of Oral Health Sciences, Faculty of Medicine, University of Oulu, Oulu, Finland
| | - Jérôme Thevenot
- Medical Imaging, Physics and Technology Research Unit, University of Oulu, Oulu, Finland
| |
Collapse
|
5
|
Zhang N, Jiang Z, Li M, Zhang D. A novel multi-feature learning model for disease diagnosis using face skin images. Comput Biol Med 2024; 168:107837. [PMID: 38086142 DOI: 10.1016/j.compbiomed.2023.107837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Revised: 11/15/2023] [Accepted: 12/07/2023] [Indexed: 01/10/2024]
Abstract
BACKGROUND Facial skin characteristics can provide valuable information about a patient's underlying health conditions. OBJECTIVE In practice, there are often samples with divergent characteristics (commonly known as divergent samples) that can be attributed to environmental factors, living conditions, or genetic elements. These divergent samples significantly degrade the accuracy of diagnoses. METHODOLOGY To tackle this problem, we propose a novel multi-feature learning method called Multi-Feature Learning with Centroid Matrix (MFLCM), which aims to mitigate the influence of divergent samples on the accurate classification of samples located on the boundary. In this approach, we introduce a novel discriminator that incorporates a centroid matrix strategy and simultaneously adapt it to a classifier in a unified model. We effectively apply the centroid matrix to the embedding feature spaces, which are transformed from the multi-feature observation space, by calculating a relaxed Hamming distance. The purpose of the centroid vectors for each category is to act as anchors, ensuring that samples from the same class are positioned close to their corresponding centroid vector while being pushed further away from the remaining centroids. RESULTS Validation of the proposed method with clinical facial skin dataset showed that the proposed method achieved F1 scores of 92.59%, 83.35%, 82.84% and 85.46%, respectively for the detection the Healthy, Diabetes Mellitus (DM), Fatty Liver (FL) and Chronic Renal Failure (CRF). CONCLUSION Experimental results demonstrate the superiority of the proposed method compared with typical classifiers single-view-based and state-of-the-art multi-feature approaches. To the best of our knowledge, this study represents the first to demonstrate concept of multi-feature learning using only facial skin images as an effective non-invasive approach for simultaneously identifying DM, FL and CRF in Han Chinese, the largest ethnic group in the world.
Collapse
Affiliation(s)
- Nannan Zhang
- The Chinese University of Hong Kong (Shenzhen), Shenzhen, China; Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen, China; Shenzhen Research Institute of Big Data, Shenzhen, China.
| | - Zhixing Jiang
- The Chinese University of Hong Kong (Shenzhen), Shenzhen, China; Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen, China; Shenzhen Research Institute of Big Data, Shenzhen, China.
| | - Mu Li
- Harbin Institute of Technology at Shenzhen, Shenzhen, China.
| | - David Zhang
- The Chinese University of Hong Kong (Shenzhen), Shenzhen, China; Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen, China; Shenzhen Research Institute of Big Data, Shenzhen, China.
| |
Collapse
|
6
|
Bannister JJ, Wilms M, Aponte JD, Katz DC, Klein OD, Bernier FP, Spritz RA, Hallgrímsson B, Forkert ND. Comparing 2D and 3D representations for face-based genetic syndrome diagnosis. Eur J Hum Genet 2023; 31:1010-1016. [PMID: 36750664 PMCID: PMC10474012 DOI: 10.1038/s41431-023-01308-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Revised: 12/01/2022] [Accepted: 01/27/2023] [Indexed: 02/09/2023] Open
Abstract
Human genetic syndromes are often challenging to diagnose clinically. Facial phenotype is a key diagnostic indicator for hundreds of genetic syndromes and computer-assisted facial phenotyping is a promising approach to assist diagnosis. Most previous approaches to automated face-based syndrome diagnosis have analyzed different datasets of either 2D images or surface mesh-based 3D facial representations, making direct comparisons of performance challenging. In this work, we developed a set of subject-matched 2D and 3D facial representations, which we then analyzed with the aim of comparing the performance of 2D and 3D image-based approaches to computer-assisted syndrome diagnosis. This work represents the most comprehensive subject-matched analyses to date on this topic. In our analyses of 1907 subject faces representing 43 different genetic syndromes, 3D surface-based syndrome classification models significantly outperformed 2D image-based models trained and evaluated on the same subject faces. These results suggest that the clinical adoption of 3D facial scanning technology and continued collection of syndromic 3D facial scan data may substantially improve face-based syndrome diagnosis.
Collapse
Affiliation(s)
- Jordan J Bannister
- Department of Biomedical Engineering, University of Calgary, Calgary, AB, Canada.
| | - Matthias Wilms
- Department of Pediatrics, Department of Community Health Sciences, Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada
| | - J David Aponte
- Department of Cell Biology and Anatomy, University of Calgary, Calgary, AB, Canada
| | - David C Katz
- Department of Cell Biology and Anatomy, University of Calgary, Calgary, AB, Canada
| | - Ophir D Klein
- Program in Craniofacial Biology, Department of Orofacial Sciences, University of California, San Francisco, CA, USA
| | - Francois P Bernier
- Department of Medical Genetics and the Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada
| | - Richard A Spritz
- Department of Pediatrics, University of Colorado School of Medicine, Aurora, CO, USA
| | | | - Nils D Forkert
- Department of Radiology, Alberta Children's Hospital Research Institute, Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada
| |
Collapse
|
7
|
Application of a novel deep learning technique using CT images for COVID-19 diagnosis on embedded systems. ALEXANDRIA ENGINEERING JOURNAL 2023; 74:345-358. [PMCID: PMC10183629 DOI: 10.1016/j.aej.2023.05.036] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/05/2023] [Revised: 04/24/2023] [Accepted: 05/08/2023] [Indexed: 11/04/2023]
Abstract
Problem A novel coronavirus (COVID-19) has created a worldwide pneumonia epidemic, and it's important to make a computer-aided way for doctors to use computed tomography (CT) images to find people with COVID-19 as soon as possible. Aim: A fully automated, novel deep-learning method for diagnosis and prognostic analysis of COVID-19 on the embedded system is presented. Methods In this study, CT scans are utilized to identify individuals with COVID-19, pneumonia, or normal class. To achieve classification two pre-trained CNN models, namely ResNet50 and MobileNetv2, which are commonly used for image classification tasks. Additionally, a novel CNN architecture called CovidxNet-CT is introduced specifically designed for COVID-19 diagnosis using three classes of CT scans. To evaluate the effectiveness of the proposed method, k-fold cross-validation is employed, which is a common approach to estimate the performance of deep learning. The study is also evaluated the proposed method on two embedded system platforms, Jetson Nano and Tx2, to demonstrate its feasibility for deployment in resource-constrained environments. Results With an average accuracy of %98.83 and an AUC of 0.988, the system is trained and verified using a 4 fold cross-validation approach. Conclusion The optimistic outcomes from the investigation propose that CovidxNet-CT has the capacity to support radiologists and contribute towards the efforts to combat COVID-19. This study proposes a fully automated, deep-learning-based method for COVID-19 diagnosis and prognostic analysis that is specifically designed for use on embedded systems.
Collapse
|
8
|
Soto I, Zamorano-Illanes R, Becerra R, Palacios Játiva P, Azurdia-Meza CA, Alavia W, García V, Ijaz M, Zabala-Blanco D. A New COVID-19 Detection Method Based on CSK/QAM Visible Light Communication and Machine Learning. SENSORS (BASEL, SWITZERLAND) 2023; 23:1533. [PMID: 36772574 DOI: 10.3390/s23031533] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/24/2022] [Revised: 01/23/2023] [Accepted: 01/26/2023] [Indexed: 06/18/2023]
Abstract
This article proposes a novel method for detecting coronavirus disease 2019 (COVID-19) in an underground channel using visible light communication (VLC) and machine learning (ML). We present mathematical models of COVID-19 Deoxyribose Nucleic Acid (DNA) gene transfer in regular square constellations using a CSK/QAM-based VLC system. ML algorithms are used to classify the bands present in each electrophoresis sample according to whether the band corresponds to a positive, negative, or ladder sample during the search for the optimal model. Complexity studies reveal that the square constellation N=22i×22i,(i=3) yields a greater profit. Performance studies indicate that, for BER = 10-3, there are gains of -10 [dB], -3 [dB], 3 [dB], and 5 [dB] for N=22i×22i,(i=0,1,2,3), respectively. Based on a total of 630 COVID-19 samples, the best model is shown to be XGBoots, which demonstrated an accuracy of 96.03%, greater than that of the other models, and a recall of 99% for positive values.
Collapse
Affiliation(s)
- Ismael Soto
- CIMTT, Department of Electrical Engineering, Universidad de Santiago de Chile, Santiago 9170124, Chile
| | - Raul Zamorano-Illanes
- CIMTT, Department of Electrical Engineering, Universidad de Santiago de Chile, Santiago 9170124, Chile
| | - Raimundo Becerra
- Department of Electrical Engineering, Universidad de Chile, Santiago 8370451, Chile
| | - Pablo Palacios Játiva
- Department of Electrical Engineering, Universidad de Chile, Santiago 8370451, Chile
- Escuela de Informática y Telecomunicaciones, Universidad Diego Portales, Santiago 8370190, Chile
| | - Cesar A Azurdia-Meza
- Department of Electrical Engineering, Universidad de Chile, Santiago 8370451, Chile
| | - Wilson Alavia
- CIMTT, Department of Electrical Engineering, Universidad de Santiago de Chile, Santiago 9170124, Chile
| | - Verónica García
- Departamento en Ciencia y Tecnología de los Alimentos, de la Universidad de Santiago de Chile, Santiago 9170124, Chile
| | - Muhammad Ijaz
- Manchester Metropolitan University, Manchester M1 5GD, UK
| | - David Zabala-Blanco
- Department of Computer Science and Industry, Universidad Católica del Maule, Talca 3480112, Chile
| |
Collapse
|
9
|
Hennocq Q, Bongibault T, Bizière M, Delassus O, Douillet M, Cormier-Daire V, Amiel J, Lyonnet S, Marlin S, Rio M, Picard A, Khonsari RH, Garcelon N. An automatic facial landmarking for children with rare diseases. Am J Med Genet A 2023; 191:1210-1221. [PMID: 36714960 DOI: 10.1002/ajmg.a.63126] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Revised: 01/10/2023] [Accepted: 01/12/2023] [Indexed: 01/31/2023]
Abstract
Two to three thousand syndromes modify facial features: their screening requires the eye of an expert in dysmorphology. A widely used tool in shape characterization is geometric morphometrics based on landmarks, which are precise and reproducible anatomical points. Landmark positioning is user dependent and time consuming. Many automatic landmarking tools are currently available but do not work for children, because they have mainly been trained using photographic databases of healthy adults. Here, we developed a method for building an automatic landmarking pipeline for frontal and lateral facial photographs as well as photographs of external ears. We evaluated the algorithm on patients diagnosed with Treacher Collins (TC) syndrome as it is the most frequent mandibulofacial dysostosis in humans and is clinically recognizable although highly variable in severity. We extracted photographs from the photographic database of the maxillofacial surgery and plastic surgery department of Hôpital Necker-Enfants Malades in Paris, France with the diagnosis of TC syndrome. The control group was built from children admitted for craniofacial trauma or skin lesions. After testing two methods of object detection by bounding boxes, a Haar Cascade-based tool and a Faster Region-based Convolutional Neural Network (Faster R-CNN)-based tool, we evaluated three different automatic annotation algorithms: the patch-based active appearance model (AAM), the holistic AAM, and the constrained local model (CLM). The final error corresponding to the distance between the points placed by automatic annotation and those placed by manual annotation was reported. We included, respectively, 1664, 2044, and 1375 manually annotated frontal, profile, and ear photographs. Object recognition was optimized with the Faster R-CNN-based detector. The best annotation model was the patch-based AAM (p < 0.001 for frontal faces, p = 0.082 for profile faces and p < 0.001 for ears). This automatic annotation model resulted in the same classification performance as manually annotated data. Pretraining on public photographs did not improve the performance of the model. We defined a pipeline to create automatic annotation models adapted to faces with congenital anomalies, an essential prerequisite for research in dysmorphology.
Collapse
Affiliation(s)
- Quentin Hennocq
- Imagine Institute, INSERM UMR 1163, Paris, France.,Département de chirurgie maxillo-faciale et chirurgie plastique pédiatrique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Centre de Référence des Malformations Rares de la Face et de la Cavité Buccale MAFACE, Filière Maladies Rares TeteCou, Faculté de Médecine, Université de Paris Cité, Paris, France
| | | | | | | | | | - Valérie Cormier-Daire
- Fédération de médecine génomique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Faculté de Médecine, Université de Paris Cité, Paris, France
| | - Jeanne Amiel
- Fédération de médecine génomique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Faculté de Médecine, Université de Paris Cité, Paris, France
| | - Stanislas Lyonnet
- Fédération de médecine génomique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Faculté de Médecine, Université de Paris Cité, Paris, France
| | - Sandrine Marlin
- Fédération de médecine génomique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Faculté de Médecine, Université de Paris Cité, Paris, France
| | - Marlène Rio
- Fédération de médecine génomique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Faculté de Médecine, Université de Paris Cité, Paris, France
| | - Arnaud Picard
- Département de chirurgie maxillo-faciale et chirurgie plastique pédiatrique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Centre de Référence des Malformations Rares de la Face et de la Cavité Buccale MAFACE, Filière Maladies Rares TeteCou, Faculté de Médecine, Université de Paris Cité, Paris, France
| | - Roman Hossein Khonsari
- Département de chirurgie maxillo-faciale et chirurgie plastique pédiatrique, Hôpital Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Centre de Référence des Malformations Rares de la Face et de la Cavité Buccale MAFACE, Filière Maladies Rares TeteCou, Faculté de Médecine, Université de Paris Cité, Paris, France
| | | |
Collapse
|
10
|
Young P, Ebadi N, Das A, Bethany M, Desai K, Najafirad P. Can Hierarchical Transformers Learn Facial Geometry? SENSORS (BASEL, SWITZERLAND) 2023; 23:929. [PMID: 36679725 PMCID: PMC9862876 DOI: 10.3390/s23020929] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/01/2022] [Revised: 01/10/2023] [Accepted: 01/11/2023] [Indexed: 06/17/2023]
Abstract
Human faces are a core part of our identity and expression, and thus, understanding facial geometry is key to capturing this information. Automated systems that seek to make use of this information must have a way of modeling facial features in a way that makes them accessible. Hierarchical, multi-level architectures have the capability of capturing the different resolutions of representation involved. In this work, we propose using a hierarchical transformer architecture as a means of capturing a robust representation of facial geometry. We further demonstrate the versatility of our approach by using this transformer as a backbone to support three facial representation problems: face anti-spoofing, facial expression representation, and deepfake detection. The combination of effective fine-grained details alongside global attention representations makes this architecture an excellent candidate for these facial representation problems. We conduct numerous experiments first showcasing the ability of our approach to address common issues in facial modeling (pose, occlusions, and background variation) and capture facial symmetry, then demonstrating its effectiveness on three supplemental tasks.
Collapse
Affiliation(s)
- Paul Young
- Department of Computer Science, University of Texas at San Antonio, San Antonio, TX 78249, USA
| | - Nima Ebadi
- Department of Electrical and Computer Engineering, University of Texas at San Antonio, San Antonio, TX 78249, USA
| | - Arun Das
- Department of Electrical and Computer Engineering, University of Texas at San Antonio, San Antonio, TX 78249, USA
- Department of Medicine, University of Pittsburgh, Pittsburgh, PA 15260, USA
| | - Mazal Bethany
- Department of Information Systems, University of Texas at San Antonio, San Antonio, TX 78249, USA
| | - Kevin Desai
- Department of Computer Science, University of Texas at San Antonio, San Antonio, TX 78249, USA
| | - Peyman Najafirad
- Department of Computer Science, University of Texas at San Antonio, San Antonio, TX 78249, USA
- Department of Information Systems, University of Texas at San Antonio, San Antonio, TX 78249, USA
| |
Collapse
|
11
|
Bhatele KR, Jha A, Tiwari D, Bhatele M, Sharma S, Mithora MR, Singhal S. COVID-19 Detection: A Systematic Review of Machine and Deep Learning-Based Approaches Utilizing Chest X-Rays and CT Scans. Cognit Comput 2022; 16:1-38. [PMID: 36593991 PMCID: PMC9797382 DOI: 10.1007/s12559-022-10076-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2022] [Accepted: 11/15/2022] [Indexed: 12/30/2022]
Abstract
This review study presents the state-of-the-art machine and deep learning-based COVID-19 detection approaches utilizing the chest X-rays or computed tomography (CT) scans. This study aims to systematically scrutinize as well as to discourse challenges and limitations of the existing state-of-the-art research published in this domain from March 2020 to August 2021. This study also presents a comparative analysis of the performance of four majorly used deep transfer learning (DTL) models like VGG16, VGG19, ResNet50, and DenseNet over the COVID-19 local CT scans dataset and global chest X-ray dataset. A brief illustration of the majorly used chest X-ray and CT scan datasets of COVID-19 patients utilized in state-of-the-art COVID-19 detection approaches are also presented for future research. The research databases like IEEE Xplore, PubMed, and Web of Science are searched exhaustively for carrying out this survey. For the comparison analysis, four deep transfer learning models like VGG16, VGG19, ResNet50, and DenseNet are initially fine-tuned and trained using the augmented local CT scans and global chest X-ray dataset in order to observe their performance. This review study summarizes major findings like AI technique employed, type of classification performed, used datasets, results in terms of accuracy, specificity, sensitivity, F1 score, etc., along with the limitations, and future work for COVID-19 detection in tabular manner for conciseness. The performance analysis of the four majorly used deep transfer learning models affirms that Visual Geometry Group 19 (VGG19) model delivered the best performance over both COVID-19 local CT scans dataset and global chest X-ray dataset.
Collapse
Affiliation(s)
| | - Anand Jha
- RJIT BSF Academy, Tekanpur, Gwalior India
| | | | | | | | | | | |
Collapse
|
12
|
Bannister JJ, Wilms M, Aponte JD, Katz DC, Klein OD, Bernier FPJ, Spritz RA, Hallgrímsson B, Forkert ND. Detecting 3D syndromic faces as outliers using unsupervised normalizing flow models. Artif Intell Med 2022; 134:102425. [PMID: 36462895 PMCID: PMC10949379 DOI: 10.1016/j.artmed.2022.102425] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2022] [Revised: 10/08/2022] [Accepted: 10/09/2022] [Indexed: 12/13/2022]
Abstract
Many genetic syndromes are associated with distinctive facial features. Several computer-assisted methods have been proposed that make use of facial features for syndrome diagnosis. Training supervised classifiers, the most common approach for this purpose, requires large, comprehensive, and difficult to collect databases of syndromic facial images. In this work, we use unsupervised, normalizing flow-based manifold and density estimation models trained entirely on unaffected subjects to detect syndromic 3D faces as statistical outliers. Furthermore, we demonstrate a general, user-friendly, gradient-based interpretability mechanism that enables clinicians and patients to understand model inferences. 3D facial surface scans of 2471 unaffected subjects and 1629 syndromic subjects representing 262 different genetic syndromes were used to train and evaluate the models. The flow-based models outperformed unsupervised comparison methods, with the best model achieving an ROC-AUC of 86.3% on a challenging, age and sex diverse data set. In addition to highlighting the viability of outlier-based syndrome screening tools, our methods generalize and extend previously proposed outlier scores for 3D face-based syndrome detection, resulting in improved performance for unsupervised syndrome detection.
Collapse
Affiliation(s)
- Jordan J Bannister
- Biomedical Engineering Graduate Program, University of Calgary, Calgary, AB, Canada.
| | - Matthias Wilms
- Department of Pediatrics, Department of Community Health Sciences, Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada
| | - J David Aponte
- Department of Cell Biology and Anatomy, University of Calgary, 2500 University Dr NW, Calgary, AB, Canada
| | - David C Katz
- Department of Cell Biology and Anatomy, University of Calgary, 2500 University Dr NW, Calgary, AB, Canada
| | - Ophir D Klein
- Program in Craniofacial Biology, Department of Orofacial Sciences, University of California, San Francisco, CA, USA
| | - Francois P J Bernier
- Department of Medical Genetics, Alberta Children's Hospital Research Institute, University of Calgary, Calgary, AB, Canada
| | - Richard A Spritz
- Department of Pediatrics, University of Colorado School of Medicine, Aurora, CO, USA
| | - Benedikt Hallgrímsson
- Department of Cell Biology and Anatomy, University of Calgary, 2500 University Dr NW, Calgary, AB, Canada
| | - Nils D Forkert
- Department of Radiology, Alberta Children's Hospital Research Institute, Hotchkiss Brain Institute, University of Calgary, Calgary, AB, Canada
| |
Collapse
|
13
|
Fuzi J, Meller C, Ch'ng S, Hadlock TM, Dusseldorp J. Voluntary and Spontaneous Smile Quantification in Facial Palsy Patients: Validation of a Novel Mobile Application. Facial Plast Surg Aesthet Med 2022. [DOI: 10.1089/fpsam.2022.0104] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022] Open
Affiliation(s)
- Jordan Fuzi
- Prince of Wales Hospital, Randwick, Australia
- University of Sydney, Camperdown, Australia
| | | | - Sydney Ch'ng
- University of Sydney, Camperdown, Australia
- Chris O'Brien Lifehouse, Camperdown, Australia
| | | | - Joseph Dusseldorp
- University of Sydney, Camperdown, Australia
- Chris O'Brien Lifehouse, Camperdown, Australia
- Concord Repatriation General Hospital, Concord, Australia
| |
Collapse
|
14
|
Zlatintsi A, Filntisis PP, Garoufis C, Efthymiou N, Maragos P, Menychtas A, Maglogiannis I, Tsanakas P, Sounapoglou T, Kalisperakis E, Karantinos T, Lazaridi M, Garyfalli V, Mantas A, Mantonakis L, Smyrnis N. E-Prevention: Advanced Support System for Monitoring and Relapse Prevention in Patients with Psychotic Disorders Analyzing Long-Term Multimodal Data from Wearables and Video Captures. SENSORS (BASEL, SWITZERLAND) 2022; 22:7544. [PMID: 36236643 PMCID: PMC9572170 DOI: 10.3390/s22197544] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/24/2022] [Revised: 09/23/2022] [Accepted: 09/26/2022] [Indexed: 06/16/2023]
Abstract
Wearable technologies and digital phenotyping foster unique opportunities for designing novel intelligent electronic services that can address various well-being issues in patients with mental disorders (i.e., schizophrenia and bipolar disorder), thus having the potential to revolutionize psychiatry and its clinical practice. In this paper, we present e-Prevention, an innovative integrated system for medical support that facilitates effective monitoring and relapse prevention in patients with mental disorders. The technologies offered through e-Prevention include: (i) long-term continuous recording of biometric and behavioral indices through a smartwatch; (ii) video recordings of patients while being interviewed by a clinician, using a tablet; (iii) automatic and systematic storage of these data in a dedicated Cloud server and; (iv) the ability of relapse detection and prediction. This paper focuses on the description of the e-Prevention system and the methodologies developed for the identification of feature representations that correlate with and can predict psychopathology and relapses in patients with mental disorders. Specifically, we tackle the problem of relapse detection and prediction using Machine and Deep Learning techniques on all collected data. The results are promising, indicating that such predictions could be made and leading eventually to the prediction of psychopathology and the prevention of relapses.
Collapse
Affiliation(s)
- Athanasia Zlatintsi
- School of ECE, National Technical University of Athens, 157 73 Athens, Greece
| | | | - Christos Garoufis
- School of ECE, National Technical University of Athens, 157 73 Athens, Greece
| | - Niki Efthymiou
- School of ECE, National Technical University of Athens, 157 73 Athens, Greece
| | - Petros Maragos
- School of ECE, National Technical University of Athens, 157 73 Athens, Greece
| | - Andreas Menychtas
- Department of Digital Systems, University of Piraeus, 185 34 Pireas, Greece
| | - Ilias Maglogiannis
- Department of Digital Systems, University of Piraeus, 185 34 Pireas, Greece
| | - Panayiotis Tsanakas
- School of ECE, National Technical University of Athens, 157 73 Athens, Greece
| | | | - Emmanouil Kalisperakis
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
- 1st Department of Psychiatry, Eginition Hospital, Medical School, National and Kapodistrian University of Athens, 115 28 Athens, Greece
| | - Thomas Karantinos
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
| | - Marina Lazaridi
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
- 1st Department of Psychiatry, Eginition Hospital, Medical School, National and Kapodistrian University of Athens, 115 28 Athens, Greece
| | - Vasiliki Garyfalli
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
- 1st Department of Psychiatry, Eginition Hospital, Medical School, National and Kapodistrian University of Athens, 115 28 Athens, Greece
| | - Asimakis Mantas
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
| | - Leonidas Mantonakis
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
- 1st Department of Psychiatry, Eginition Hospital, Medical School, National and Kapodistrian University of Athens, 115 28 Athens, Greece
| | - Nikolaos Smyrnis
- Laboratory of Cognitive Neuroscience and Sensorimotor Control, University Mental Health, Neurosciences and Precision Medicine Research Institute “COSTAS STEFANIS”, 115 27 Athens, Greece
- 2nd Department of Psychiatry, University General Hospital “ATTIKON”, Medical School, National and Kapodistrian University of Athens, 124 62 Athens, Greece
| |
Collapse
|
15
|
Birhane A. The unseen Black faces of AI algorithms. Nature 2022; 610:451-452. [PMID: 36261566 DOI: 10.1038/d41586-022-03050-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
16
|
Breaking CAPTCHA with Capsule Networks. Neural Netw 2022; 154:246-254. [PMID: 35908374 DOI: 10.1016/j.neunet.2022.06.041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2021] [Revised: 05/17/2022] [Accepted: 06/30/2022] [Indexed: 11/22/2022]
Abstract
Convolutional Neural Networks have achieved state-of-the-art performance in image classification. Their lack of ability to recognise the spatial relationship between features, however, leads to misclassification of the variants of the same image. Capsule Networks were introduced to address this issue by incorporating the spatial information of image features into neural networks. In this paper, we are interested in showcasing the digit recognition task on CAPTCHA images, widely considered a challenge for computers in relation to human capabilities. Our intention is to provide a rigorous empirical regime in which we can compare the competitive performance of Capsule Networks against the Convolutional Neural Networks. Indeed since CAPTCHA distorts images, by adjusting the spatial positioning of features, we aim to demonstrate the advantages and limitations of Capsule Networks architecture. We train the Capsule Networks with Dynamic Routing version and the convolutional-neural-network-based deep-CAPTCHA baseline model to predict the digit sequences on numerical CAPTCHAs, investigate the performance results and propose two improvements to the Capsule Networks model.
Collapse
|
17
|
Parra-Dominguez GS, Garcia-Capulin CH, Sanchez-Yanez RE. Automatic Facial Palsy Diagnosis as a Classification Problem Using Regional Information Extracted from a Photograph. Diagnostics (Basel) 2022; 12:diagnostics12071528. [PMID: 35885434 PMCID: PMC9317944 DOI: 10.3390/diagnostics12071528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2022] [Revised: 06/17/2022] [Accepted: 06/21/2022] [Indexed: 11/27/2022] Open
Abstract
The incapability to move the facial muscles is known as facial palsy, and it affects various abilities of the patient, for example, performing facial expressions. Recently, automatic approaches aiming to diagnose facial palsy using images and machine learning algorithms have emerged, focusing on providing an objective evaluation of the paralysis severity. This research proposes an approach to analyze and assess the lesion severity as a classification problem with three levels: healthy, slight, and strong palsy. The method explores the use of regional information, meaning that only certain areas of the face are of interest. Experiments carrying on multi-class classification tasks are performed using four different classifiers to validate a set of proposed hand-crafted features. After a set of experiments using this methodology on available image databases, great results are revealed (up to 95.61% of correct detection of palsy patients and 95.58% of correct assessment of the severity level). This perspective leads us to believe that the analysis of facial paralysis is possible with partial occlusions if face detection is accomplished and facial features are obtained adequately. The results also show that our methodology is suited to operate with other databases while attaining high performance, even though the image conditions are different and the participants do not perform equivalent facial expressions.
Collapse
|
18
|
Cai T, Ni H, Yu M, Huang X, Wong K, Volpi J, Wang JZ, Wong ST. DeepStroke: An Efficient Stroke Screening Framework for Emergency Rooms with Multimodal Adversarial Deep Learning. Med Image Anal 2022; 80:102522. [DOI: 10.1016/j.media.2022.102522] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2021] [Revised: 04/02/2022] [Accepted: 06/24/2022] [Indexed: 11/24/2022]
|
19
|
Smith CF, Nicholson BD. Creating space for gut feelings in the diagnosis of cancer in primary care. Br J Gen Pract 2022; 72:210-211. [PMID: 35483938 PMCID: PMC11189036 DOI: 10.3399/bjgp22x719249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/31/2022] Open
Affiliation(s)
| | - Brian D Nicholson
- National Institute for Health Research Academic Clinical Lecturer and GP, Nuffield Department of Primary Care Health Sciences, Oxford
| |
Collapse
|
20
|
Su Z, Bentley BL, McDonnell D, Ahmad J, He J, Shi F, Takeuchi K, Cheshmehzangi A, da Veiga CP. 6G and Artificial Intelligence Technologies for Dementia Care: Literature Review and Practical Analysis. J Med Internet Res 2022; 24:e30503. [PMID: 35475733 PMCID: PMC9096635 DOI: 10.2196/30503] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2021] [Revised: 11/23/2021] [Accepted: 01/31/2022] [Indexed: 01/20/2023] Open
Abstract
Background The dementia epidemic is progressing fast. As the world’s older population keeps skyrocketing, the traditional incompetent, time-consuming, and laborious interventions are becoming increasingly insufficient to address dementia patients’ health care needs. This is particularly true amid COVID-19. Instead, efficient, cost-effective, and technology-based strategies, such as sixth-generation communication solutions (6G) and artificial intelligence (AI)-empowered health solutions, might be the key to successfully managing the dementia epidemic until a cure becomes available. However, while 6G and AI technologies hold great promise, no research has examined how 6G and AI applications can effectively and efficiently address dementia patients’ health care needs and improve their quality of life. Objective This study aims to investigate ways in which 6G and AI technologies could elevate dementia care to address this study gap. Methods A literature review was conducted in databases such as PubMed, Scopus, and PsycINFO. The search focused on three themes: dementia, 6G, and AI technologies. The initial search was conducted on April 25, 2021, complemented by relevant articles identified via a follow-up search on November 11, 2021, and Google Scholar alerts. Results The findings of the study were analyzed in terms of the interplay between people with dementia’s unique health challenges and the promising capabilities of health technologies, with in-depth and comprehensive analyses of advanced technology-based solutions that could address key dementia care needs, ranging from impairments in memory (eg, Egocentric Live 4D Perception), speech (eg, Project Relate), motor (eg, Avatar Robot Café), cognitive (eg, Affectiva), to social interactions (eg, social robots). Conclusions To live is to grow old. Yet dementia is neither a proper way to live nor a natural aging process. By identifying advanced health solutions powered by 6G and AI opportunities, our study sheds light on the imperative of leveraging the potential of advanced technologies to elevate dementia patients’ will to live, enrich their daily activities, and help them engage in societies across shapes and forms.
Collapse
Affiliation(s)
- Zhaohui Su
- School of Public Health, Southeast University, Nanjing, China
| | - Barry L Bentley
- Cardiff School of Technologies, Cardiff Metropolitan University, Cardiff, United Kingdom
| | - Dean McDonnell
- Department of Humanities, Institute of Technology Carlow, Carlow, Ireland
| | - Junaid Ahmad
- Prime Institute of Public Health, Peshawar Medical College, Peshawar, Pakistan
| | - Jiguang He
- Centre for Wireless Communications, University of Oulu, Oulu, Finland
| | - Feng Shi
- Department of Research and Development, Shanghai United Imaging Intelligence, Shanghai, China
| | - Kazuaki Takeuchi
- Ory Laboratory Inc, Tokyo, Japan.,Kanagawa Institute of Technology, Kanagawa, Japan
| | - Ali Cheshmehzangi
- Department of Architecture and Built Environment, University of Nottingham Ningbo China, Ningbo, China.,Network for Education and Research on Peace and Sustainability, Hiroshima University, Hiroshima, Japan
| | | |
Collapse
|
21
|
Matthews H, de Jong G, Maal T, Claes P. Static and Motion Facial Analysis for Craniofacial Assessment and Diagnosing Diseases. Annu Rev Biomed Data Sci 2022; 5:19-42. [PMID: 35440145 DOI: 10.1146/annurev-biodatasci-122120-111413] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Deviation from a normal facial shape and symmetry can arise from numerous sources, including physical injury and congenital birth defects. Such abnormalities can have important aesthetic and functional consequences. Furthermore, in clinical genetics distinctive facial appearances are often associated with clinical or genetic diagnoses; the recognition of a characteristic facial appearance can substantially narrow the search space of potential diagnoses for the clinician. Unusual patterns of facial movement and expression can indicate disturbances to normal mechanical functioning or emotional affect. Computational analyses of static and moving 2D and 3D images can serve clinicians and researchers by detecting and describing facial structural, mechanical, and affective abnormalities objectively. In this review we survey traditional and emerging methods of facial analysis, including statistical shape modeling, syndrome classification, modeling clinical face phenotype spaces, and analysis of facial motion and affect. Expected final online publication date for the Annual Review of Biomedical Data Science, Volume 5 is August 2022. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.
Collapse
Affiliation(s)
- Harold Matthews
- Department of Human Genetics, KU Leuven, Leuven, Belgium; .,Medical Imaging Research Center, UZ Leuven, Leuven, Belgium.,Facial Sciences Research Group, Murdoch Children's Research Institute, Parkville, Australia
| | - Guido de Jong
- 3D Lab, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Thomas Maal
- 3D Lab, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Peter Claes
- Department of Human Genetics, KU Leuven, Leuven, Belgium; .,Medical Imaging Research Center, UZ Leuven, Leuven, Belgium.,Facial Sciences Research Group, Murdoch Children's Research Institute, Parkville, Australia.,Processing Speech and Images (PSI), Department of Electrical Engineering (ESAT), KU Leuven, Leuven, Belgium
| |
Collapse
|
22
|
Jiang Z, Luskus M, Seyedi S, Griner EL, Rad AB, Clifford GD, Boazak M, Cotes RO. Utilizing computer vision for facial behavior analysis in schizophrenia studies: A systematic review. PLoS One 2022; 17:e0266828. [PMID: 35395049 PMCID: PMC8992987 DOI: 10.1371/journal.pone.0266828] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2021] [Accepted: 03/28/2022] [Indexed: 12/22/2022] Open
Abstract
Background Schizophrenia is a severe psychiatric disorder that causes significant social and functional impairment. Currently, the diagnosis of schizophrenia is based on information gleaned from the patient’s self-report, what the clinician observes directly, and what the clinician gathers from collateral informants, but these elements are prone to subjectivity. Utilizing computer vision to measure facial expressions is a promising approach to adding more objectivity in the evaluation and diagnosis of schizophrenia. Method We conducted a systematic review using PubMed and Google Scholar. Relevant publications published before (including) December 2021 were identified and evaluated for inclusion. The objective was to conduct a systematic review of computer vision for facial behavior analysis in schizophrenia studies, the clinical findings, and the corresponding data processing and machine learning methods. Results Seventeen studies published between 2007 to 2021 were included, with an increasing trend in the number of publications over time. Only 14 articles used interviews to collect data, of which different combinations of passive to evoked, unstructured to structured interviews were used. Various types of hardware were adopted and different types of visual data were collected. Commercial, open-access, and in-house developed models were used to recognize facial behaviors, where frame-level and subject-level features were extracted. Statistical tests and evaluation metrics varied across studies. The number of subjects ranged from 2-120, with an average of 38. Overall, facial behaviors appear to have a role in estimating diagnosis of schizophrenia and psychotic symptoms. When studies were evaluated with a quality assessment checklist, most had a low reporting quality. Conclusion Despite the rapid development of computer vision techniques, there are relatively few studies that have applied this technology to schizophrenia research. There was considerable variation in the clinical paradigm and analytic techniques used. Further research is needed to identify and develop standardized practices, which will help to promote further advances in the field.
Collapse
Affiliation(s)
- Zifan Jiang
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA, United States of America
- Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, GA, United States of America
- * E-mail:
| | - Mark Luskus
- Department of Psychiatry and Behavioral Sciences, University of California, San Francisco, San Francisco, CA, United States of America
| | - Salman Seyedi
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA, United States of America
| | - Emily L. Griner
- Department of Psychiatry and Behavioral Sciences, Emory University School of Medicine, Atlanta, GA, United States of America
| | - Ali Bahrami Rad
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA, United States of America
| | - Gari D. Clifford
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA, United States of America
- Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, GA, United States of America
| | - Mina Boazak
- Animo Sano Psychiatry, PLLC, Durham, NC, United States of America
| | - Robert O. Cotes
- Department of Psychiatry and Behavioral Sciences, Emory University School of Medicine, Atlanta, GA, United States of America
| |
Collapse
|
23
|
Bannister JJ, Wilms M, Aponte JD, Katz DC, Klein OD, Bernier FPJ, Spritz RA, Hallgrimsson B, Forkert ND. A Deep Invertible 3D Facial Shape Model For Interpretable Genetic Syndrome Diagnosis. IEEE J Biomed Health Inform 2022; 26:3229-3239. [PMID: 35380975 DOI: 10.1109/jbhi.2022.3164848] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
One of the primary difficulties in treating patients with genetic syndromes is diagnosing their condition. Many syndromes are associated with characteristic facial features that can be imaged and utilized by computer-assisted diagnosis systems. In this work, we develop a novel 3D facial surface modeling approach with the objective of maximizing diagnostic model interpretability within a flexible deep learning framework. Therefore, an invertible normalizing flow architecture is introduced to enable both inferential and generative tasks in a unified and efficient manner. The proposed model can be used (1) to infer syndrome diagnosis and other demographic variables given a 3D facial surface scan and (2) to explain model inferences to non-technical users via multiple interpretability mechanisms. The model was trained and evaluated on more than 4700 facial surface scans from subjects with 47 different syndromes. For the challenging task of predicting syndrome diagnosis given a new 3D facial surface scan, age, and sex of a subject, the model achieves a competitive overall top-1 accuracy of 71%, and a mean sensitivity of 43% across all syndrome classes. We believe that invertible models such as the one presented in this work can achieve competitive inferential performance while greatly increasing model interpretability in the domain of medical diagnosis.
Collapse
|
24
|
Mačianskytė D, Adaškevičius R. Automatic Detection of Human Maxillofacial Tumors by Using Thermal Imaging: A Preliminary Study. SENSORS 2022; 22:s22051985. [PMID: 35271132 PMCID: PMC8914763 DOI: 10.3390/s22051985] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/28/2021] [Revised: 02/24/2022] [Accepted: 02/25/2022] [Indexed: 11/29/2022]
Abstract
Traditional computed tomography (CT) delivers a relatively high dose of radiation to the patient and cannot be used as a method for screening of pathologies. Instead, infrared thermography (IRT) might help in the detection of pathologies, but interpreting thermal imaging (TI) is difficult even for the expert. The main objective of this work is to present a new, automated IRT method capable to discern the absence or presence of tumor in the orofacial/maxillofacial region of patients. We evaluated the use of a special feature vector extracted from face and mouth cavity thermograms in classifying TIs against the absence/presence of tumor (n = 23 patients per group). Eight statistical features extracted from TI were used in a k-nearest neighbor (kNN) classifier. Classification accuracy of kNN was evaluated by CT, and by creating a vector with the true class labels for TIs. The presented algorithm, constructed from a training data set, gives good results of classification accuracy of kNN: sensitivity of 77.9%, specificity of 94.9%, and accuracy of 94.1%. The new algorithm exhibited almost the same accuracy in detecting the absence/presence of tumor as CT, and is a proof-of-principle that IRT could be useful as an additional reliable screening tool for detecting orofacial/maxillofacial tumors.
Collapse
Affiliation(s)
- Diana Mačianskytė
- The Clinic of Radiology, Lithuanian University of Health Sciences, LT-50009 Kaunas, Lithuania
| | - Rimas Adaškevičius
- Department of Electrical Power Systems, Kaunas University of Technology, LT-51368 Kaunas, Lithuania
| |
Collapse
|
25
|
Gat L, Gerston A, Shikun L, Inzelberg L, Hanein Y. Similarities and disparities between visual analysis and high-resolution electromyography of facial expressions. PLoS One 2022; 17:e0262286. [PMID: 35192638 PMCID: PMC8863227 DOI: 10.1371/journal.pone.0262286] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2021] [Accepted: 12/21/2021] [Indexed: 11/19/2022] Open
Abstract
Computer vision (CV) is widely used in the investigation of facial expressions. Applications range from psychological evaluation to neurology, to name just two examples. CV for identifying facial expressions may suffer from several shortcomings: CV provides indirect information about muscle activation, it is insensitive to activations that do not involve visible deformations, such as jaw clenching. Moreover, it relies on high-resolution and unobstructed visuals. High density surface electromyography (sEMG) recordings with soft electrode array is an alternative approach which provides direct information about muscle activation, even from freely behaving humans. In this investigation, we compare CV and sEMG analysis of facial muscle activation. We used independent component analysis (ICA) and multiple linear regression (MLR) to quantify the similarity and disparity between the two approaches for posed muscle activations. The comparison reveals similarity in event detection, but discrepancies and inconsistencies in source identification. Specifically, the correspondence between sEMG and action unit (AU)-based analyses, the most widely used basis of CV muscle activation prediction, appears to vary between participants and sessions. We also show a comparison between AU and sEMG data of spontaneous smiles, highlighting the differences between the two approaches. The data presented in this paper suggests that the use of AU-based analysis should consider its limited ability to reliably compare between different sessions and individuals and highlight the advantages of high-resolution sEMG for facial expression analysis.
Collapse
Affiliation(s)
- Liraz Gat
- School of Electrical Engineering, Tel Aviv University, Tel Aviv, Israel
- Tel Aviv University Center for Nanoscience and Nanotechnology, Tel Aviv University, Tel Aviv, Israel
| | - Aaron Gerston
- Tel Aviv University Center for Nanoscience and Nanotechnology, Tel Aviv University, Tel Aviv, Israel
- X-trodes, Herzelia, Israel
| | - Liu Shikun
- School of Electrical Engineering, Tel Aviv University, Tel Aviv, Israel
- Tel Aviv University Center for Nanoscience and Nanotechnology, Tel Aviv University, Tel Aviv, Israel
| | - Lilah Inzelberg
- Tel Aviv University Center for Nanoscience and Nanotechnology, Tel Aviv University, Tel Aviv, Israel
- Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, Israel
| | - Yael Hanein
- School of Electrical Engineering, Tel Aviv University, Tel Aviv, Israel
- Tel Aviv University Center for Nanoscience and Nanotechnology, Tel Aviv University, Tel Aviv, Israel
- Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, Israel
- X-trodes, Herzelia, Israel
- * E-mail:
| |
Collapse
|
26
|
Chen LY, Tsai TH, Ho A, Li CH, Ke LJ, Peng LN, Lin MH, Hsiao FY, Chen LK. Predicting neuropsychiatric symptoms of persons with dementia in a day care center using a facial expression recognition system. Aging (Albany NY) 2022; 14:1280-1291. [PMID: 35113806 PMCID: PMC8876896 DOI: 10.18632/aging.203869] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Accepted: 01/17/2022] [Indexed: 11/25/2022]
Abstract
BACKGROUND Behavioral and psychological symptoms of dementia (BPSD) affect 90% of persons with dementia (PwD), resulting in various adverse outcomes and aggravating care burdens among their caretakers. This study aimed to explore the potential of artificial intelligence-based facial expression recognition systems (FERS) in predicting BPSDs among PwD. METHODS A hybrid of human labeling and a preconstructed deep learning model was used to differentiate basic facial expressions of individuals to predict the results of Neuropsychiatric Inventory (NPI) assessments by stepwise linear regression (LR), random forest (RF) with importance ranking, and ensemble method (EM) of equal importance, while the accuracy was determined by mean absolute error (MAE) and root-mean-square error (RMSE) methods. RESULTS Twenty-three PwD from an adult day care center were enrolled with ≥ 11,500 FERS data series and 38 comparative NPI scores. The overall accuracy was 86% on facial expression recognition. Negative facial expressions and variance in emotional switches were important features of BPSDs. A strong positive correlation was identified in each model (EM: r = 0.834, LR: r = 0.821, RF: r = 0.798 by the patientwise method; EM: r = 0.891, LR: r = 0.870, RF: r = 0.886 by the MinimPy method), and EM exhibited the lowest MAE and RMSE. CONCLUSIONS FERS successfully predicted the BPSD of PwD by negative emotions and the variance in emotional switches. This finding enables early detection and management of BPSDs, thus improving the quality of dementia care.
Collapse
Affiliation(s)
- Liang-Yu Chen
- Aging and Health Research Center, Taipei, Taiwan
- Institute of Public Health, National Yang-Ming Chiao-Tung University, Taipei, Taiwan
- Center for Geriatrics and Gerontology, Taipei, Taiwan
- uAge Day Care Center, Taipei Veterans General Hospital, Taipei, Taiwan
| | | | - Andy Ho
- Value Lab, Acer Incorporated, New Taipei City, Taiwan
| | - Chun-Hsien Li
- Value Lab, Acer Incorporated, New Taipei City, Taiwan
| | - Li-Ju Ke
- uAge Day Care Center, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Li-Ning Peng
- Aging and Health Research Center, Taipei, Taiwan
- Center for Geriatrics and Gerontology, Taipei, Taiwan
| | - Ming-Hsien Lin
- Aging and Health Research Center, Taipei, Taiwan
- Center for Geriatrics and Gerontology, Taipei, Taiwan
| | - Fei-Yuan Hsiao
- Graduate Institute of Clinical Pharmacy, National Taiwan University, Taipei, Taiwan
- School of Pharmacy, National Taiwan University, Taipei, Taiwan
- Department of Pharmacy, National Taiwan University Hospital, Taipei, Taiwan
| | - Liang-Kung Chen
- Aging and Health Research Center, Taipei, Taiwan
- Center for Geriatrics and Gerontology, Taipei, Taiwan
- Taipei Municipal Gan-Dau Hospital, Taipei, Taiwan
| |
Collapse
|
27
|
Do People Trust in Robot-Assisted Surgery? Evidence from Europe. INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH 2021; 18:ijerph182312519. [PMID: 34886244 PMCID: PMC8657248 DOI: 10.3390/ijerph182312519] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Received: 11/08/2021] [Revised: 11/24/2021] [Accepted: 11/25/2021] [Indexed: 11/17/2022]
Abstract
(1) Background: The goal of the paper was to establish the factors that influence how people feel about having a medical operation performed on them by a robot. (2) Methods: Data were obtained from a 2017 Flash Eurobarometer (number 460) of the European Commission with 27,901 citizens aged 15 years and over in the 28 countries of the European Union. Logistic regression (odds ratios, OR) to model the predictors of trust in robot-assisted surgery was calculated through motivational factors, using experience and sociodemographic independent variables. (3) Results: The results obtained indicate that, as the experience of using robots increases, the predictive coefficients related to information, attitude, and perception of robots become more negative. Furthermore, sociodemographic variables played an important predictive role. The effect of experience on trust in robots for surgical interventions was greater among men, people between 40 and 54 years old, and those with higher educational levels. (4) Conclusions: The results show that trust in robots goes beyond rational decision-making, since the final decision about whether it should be a robot that performs a complex procedure like a surgical intervention depends almost exclusively on the patient’s wishes.
Collapse
|
28
|
Rathee N, Pahal S, Sheoran P. Pain detection from facial expressions using domain adaptation technique. Pattern Anal Appl 2021. [DOI: 10.1007/s10044-021-01025-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
29
|
Naeem H, Bin-Salem AA. A CNN-LSTM network with multi-level feature extraction-based approach for automated detection of coronavirus from CT scan and X-ray images. Appl Soft Comput 2021; 113:107918. [PMID: 34608379 PMCID: PMC8482540 DOI: 10.1016/j.asoc.2021.107918] [Citation(s) in RCA: 22] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2021] [Revised: 08/17/2021] [Accepted: 09/14/2021] [Indexed: 12/18/2022]
Abstract
Auto-detection of diseases has become a prime issue in medical sciences as population density is fast growing. An intelligent framework for disease detection helps physicians identify illnesses, give reliable and consistent results, and reduce death rates. Coronavirus (Covid-19) has recently been one of the most severe and acute diseases in the world. An automatic detection framework should therefore be introduced as the fastest diagnostic alternative to avoid Covid-19 spread. In this paper, an automatic Covid-19 identification in the CT scan and chest X-ray is obtained with the help of a combined deep learning and multi-level feature extraction methodology. In this method, the multi-level feature extraction approach comprises GIST, Scale Invariant Feature Transform (SIFT), and Convolutional Neural Network (CNN) extract features from CT scans and chest X-rays. The objective of multi-level feature extraction is to reduce the training complexity of CNN network, which significantly assists in accurate and robust Covid-19 identification. Finally, Long Short-Term Memory (LSTM) along the CNN network is used to detect the extracted Covid-19 features. The Kaggle SARS-CoV-2 CT scan dataset and the Italian SIRM Covid-19 CT scan and chest X-ray dataset were employed for testing purposes. Experimental outcomes show that proposed approach obtained 98.94% accuracy with the SARS-CoV-2 CT scan dataset and 83.03% accuracy with the SIRM Covid-19 CT scan and chest X-ray dataset. The proposed approach helps radiologists and practitioners to detect and treat Covid-19 cases effectively over the pandemic.
Collapse
Affiliation(s)
- Hamad Naeem
- School of Computer Science and Technology, Zhoukou Normal University, Zhoukou 466001, Henan, China
| | - Ali Abdulqader Bin-Salem
- School of Computer Science and Technology, Zhoukou Normal University, Zhoukou 466001, Henan, China
| |
Collapse
|
30
|
Lin JD, Lin HH, Dy J, Chen JC, Tanveer M, Razzak I, Hua KL. Lightweight Face Anti-Spoofing Network for Telehealth Applications. IEEE J Biomed Health Inform 2021; 26:1987-1996. [PMID: 34432642 DOI: 10.1109/jbhi.2021.3107735] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Online healthcare applications have grown more popular over the years. For instance,telehealth is an online healthcare application that allows patients and doctors to schedule consultations,prescribe medication,share medical documents,and monitor health conditions conveniently. Apart from this,telehealth can also be used to store a patients personal and medical information. Given the amount of sensitive data it stores,security measures are necessary. With its rise in usage due to COVID-19,its usefulness may be undermined if security issues are not addressed. A simple way of making these applications more secure is through user authentication. One of the most common and often used authentications is face recognition. It is convenient and easy to use. However,face recognition systems are not foolproof. They are prone to malicious attacks like printed photos,paper cutouts,re-played videos,and 3D masks. In order to counter this,multiple face anti-spoofing methods have been proposed. The goal of face anti-spoofing is to differentiate real users (live) from attackers (spoof). Although effective in terms of performance,existing methods use a significant amount of parameters,making them resource-heavy and unsuitable for handheld devices. Apart from this,they fail to generalize well to new environments like changes in lighting or background. This paper proposes a lightweight face anti-spoofing framework that does not compromise on performance. A lightweight model is critical for applications like telehealth that run on handheld devices. Our proposed method achieves good performance with the help of an ArcFace Classifier (AC). The AC encourages differentiation between spoof and live samples by making clear boundaries between them. With clear boundaries,classification becomes more accurate. We further demonstrate our models capabilities by comparing the number of parameters,FLOPS,and performance with other state-of-the-art methods.
Collapse
|
31
|
Hennocq Q, Khonsari RH, Benoît V, Rio M, Garcelon N. Computational diagnostic methods on 2D photographs: A review of the literature. JOURNAL OF STOMATOLOGY, ORAL AND MAXILLOFACIAL SURGERY 2021; 122:e71-e75. [PMID: 33848665 DOI: 10.1016/j.jormas.2021.04.003] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/11/2020] [Revised: 02/21/2021] [Accepted: 04/07/2021] [Indexed: 11/18/2022]
Abstract
Here we provide a literature review of all the methods reported to date for analyzing 2D pictures for diagnostic purposes. Pubmed was used to screen the MEDLINE database using MeSH (Medical Subject Heading) terms and keyworks. The different recognition steps and the main results were reported. All human studies involving 2D facial photographs used to diagnose one or several conditions in healthy populations or in patients were included. We included 1515 articles and 27 publications were finally retained. 67% of the articles aimed at diagnosing one particular syndrome versus healthy controls and 33% aimed at performing multi-class syndrome recognition. Data volume varied from 15 to 17,106 patient pictures. Manual or automatic landmarks were one of the most commonly used tools in order to extract morphological information from images, in 22/27 (81%) publications. Geometrical features were extracted from landmarks based on Procrustes superimposition in 4/27 (15%). Textural features were extracted in 19/27 (70%) publications. Features were then classified using machine learning methods in 89% of publications, while deep learning methods were used in 11%. Facial recognition tools were generally successful in identifying rare conditions in dysmorphic patients, with comparable or higher recognition accuracy than clinical experts.
Collapse
Affiliation(s)
- Quentin Hennocq
- Institut Imagine, Paris Descartes University-Sorbonne Paris Cité, Paris, France; Department of Maxillo-Facial Surgery and Plastic Surgery, Hôpital Universitaire Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Centre de Référence des Malformations Rares de la Face et de la Cavité Buccale MAFACE, Filière Maladies Rares TeteCou, Faculté de Médecine, Université de Paris, Paris, France.
| | - Roman Hossein Khonsari
- Institut Imagine, Paris Descartes University-Sorbonne Paris Cité, Paris, France; Department of Maxillo-Facial Surgery and Plastic Surgery, Hôpital Universitaire Necker - Enfants Malades, Assistance Publique - Hôpitaux de Paris, Centre de Référence des Malformations Rares de la Face et de la Cavité Buccale MAFACE, Filière Maladies Rares TeteCou, Faculté de Médecine, Université de Paris, Paris, France
| | - Vincent Benoît
- Institut Imagine, Paris Descartes University-Sorbonne Paris Cité, Paris, France
| | - Marlène Rio
- Institut Imagine, Paris Descartes University-Sorbonne Paris Cité, Paris, France; Department of Genetics, IHU Necker-Enfants Malades, University Paris Descartes, Paris, France
| | - Nicolas Garcelon
- Institut Imagine, Paris Descartes University-Sorbonne Paris Cité, Paris, France
| |
Collapse
|
32
|
Bandini A, Rezaei S, Guarin DL, Kulkarni M, Lim D, Boulos MI, Zinman L, Yunusova Y, Taati B. A New Dataset for Facial Motion Analysis in Individuals With Neurological Disorders. IEEE J Biomed Health Inform 2021; 25:1111-1119. [PMID: 32841132 PMCID: PMC8062040 DOI: 10.1109/jbhi.2020.3019242] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
We present the first public dataset with videos of oro-facial gestures performed by individuals with oro-facial impairment due to neurological disorders, such as amyotrophic lateral sclerosis (ALS) and stroke. Perceptual clinical scores from trained clinicians are provided as metadata. Manual annotation of facial landmarks is also provided for a subset of over 3300 frames. Through extensive experiments with multiple facial landmark detection algorithms, including state-of-the-art convolutional neural network (CNN) models, we demonstrated the presence of bias in the landmark localization accuracy of pre-trained face alignment approaches in our participant groups. The pre-trained models produced higher errors in the two clinical groups compared to age-matched healthy control subjects. We also investigated how this bias changes when the existing models are fine-tuned using data from the target population. The release of this dataset aims to propel the development of face alignment algorithms robust to the presence of oro-facial impairment, support the automatic analysis and recognition of oro-facial gestures, enhance the automatic identification of neurological diseases, as well as the estimation of disease severity from videos and images.
Collapse
|
33
|
Abstract
The inability to move the muscles of the face on one or both sides is known as facial paralysis, which may affect the ability of the patient to speak, blink, swallow saliva, eat, or communicate through natural facial expressions. The well-being of the patient could also be negatively affected. Computer-based systems as a means to detect facial paralysis are important in the development of standardized tools for medical assessment, treatment, and monitoring; additionally, they are expected to provide user-friendly tools for patient monitoring at home. In this work, a methodology to detect facial paralysis in a face photograph is proposed. A system consisting of three modules—facial landmark extraction, facial measure computation, and facial paralysis classification—was designed. Our facial measures aim to identify asymmetry levels within the face elements using facial landmarks, and a binary classifier based on a multi-layer perceptron approach provides an output label. The Weka suite was selected to design the classifier and implement the learning algorithm. Tests on publicly available databases reveal outstanding classification results on images, showing that our methodology that was used to design a binary classifier can be expanded to other databases with great results, even if the participants do not execute similar facial expressions.
Collapse
|
34
|
Glancova A, Do QT, Sanghavi DK, Franco PM, Gopal N, Lehman LM, Dong Y, Pickering BW, Herasevich V. Are We Ready for Video Recognition and Computer Vision in the Intensive Care Unit? A Survey. Appl Clin Inform 2021; 12:120-132. [PMID: 33626583 DOI: 10.1055/s-0040-1722614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022] Open
Abstract
OBJECTIVE Video recording and video recognition (VR) with computer vision have become widely used in many aspects of modern life. Hospitals have employed VR technology for security purposes, however, despite the growing number of studies showing the feasibility of VR software for physiologic monitoring or detection of patient movement, its use in the intensive care unit (ICU) in real-time is sparse and the perception of this novel technology is unknown. The objective of this study is to understand the attitudes of providers, patients, and patient's families toward using VR in the ICU. DESIGN A 10-question survey instrument was used and distributed into two groups of participants: clinicians (MDs, advance practice providers, registered nurses), patients and families (adult patients and patients' relatives). Questions were specifically worded and section for free text-comments created to elicit respondents' thoughts and attitudes on potential issues and barriers toward implementation of VR in the ICU. SETTING The survey was conducted at Mayo Clinic in Minnesota and Florida. RESULTS A total of 233 clinicians' and 50 patients' surveys were collected. Both cohorts favored VR under specific circumstances (e.g., invasive intervention and diagnostic manipulation). Acceptable reasons for VR usage according to clinicians were anticipated positive impact on patient safety (70%), and diagnostic suggestions and decision support (51%). A minority of providers was concerned that artificial intelligence (AI) would replace their job (14%) or erode professional skills (28%). The potential use of VR in lawsuits (81% clinicians) and privacy breaches (59% patients) were major areas of concern. Further identified barriers were lack of trust for AI, deterioration of the patient-clinician rapport. Patients agreed with VR unless it does not reduce nursing care or record sensitive scenarios. CONCLUSION The survey provides valuable information on the acceptance of VR cameras in the critical care setting including an overview of real concerns and attitudes toward the use of VR technology in the ICU.
Collapse
Affiliation(s)
- Alzbeta Glancova
- Department of Anesthesiology, Mayo Clinic, Rochester, Minnesota, United States
| | - Quan T Do
- Department of Anesthesiology, Mayo Clinic, Rochester, Minnesota, United States
| | - Devang K Sanghavi
- Department of Medicine, Mayo Clinic, Jacksonville, Florida, United States
| | - Pablo Moreno Franco
- Department of Medicine, Critical Care, Mayo Clinic, Jacksonville, Florida, United States
| | - Neethu Gopal
- Department of Neurology, Mayo Clinic, Jacksonville, Florida, United States
| | - Lindsey M Lehman
- Mayo Clinic, Critical Care IMP, Rochester, Minnesota, United States
| | - Yue Dong
- Department of Medicine, Mayo Clinic, Rochester, Minnesota, United States
| | - Brian W Pickering
- Department of Anesthesiology, Mayo Clinic, Rochester, Minnesota, United States
| | - Vitaly Herasevich
- Department of Anesthesiology and Medicine, Department of Anesthesiology, Mayo Clinic, Rochester, Minnesota, United States
| |
Collapse
|
35
|
ConvXGB: A new deep learning model for classification problems based on CNN and XGBoost. NUCLEAR ENGINEERING AND TECHNOLOGY 2021. [DOI: 10.1016/j.net.2020.04.008] [Citation(s) in RCA: 33] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
36
|
Abbas A, Yadav V, Smith E, Ramjas E, Rutter SB, Benavidez C, Koesmahargyo V, Zhang L, Guan L, Rosenfield P, Perez-Rodriguez M, Galatzer-Levy IR. Computer Vision-Based Assessment of Motor Functioning in Schizophrenia: Use of Smartphones for Remote Measurement of Schizophrenia Symptomatology. Digit Biomark 2021; 5:29-36. [PMID: 33615120 DOI: 10.1159/000512383] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2020] [Accepted: 10/14/2020] [Indexed: 11/19/2022] Open
Abstract
Introduction Motor abnormalities have been shown to be a distinct component of schizophrenia symptomatology. However, objective and scalable methods for assessment of motor functioning in schizophrenia are lacking. Advancements in machine learning-based digital tools have allowed for automated and remote "digital phenotyping" of disease symptomatology. Here, we assess the performance of a computer vision-based assessment of motor functioning as a characteristic of schizophrenia using video data collected remotely through smartphones. Methods Eighteen patients with schizophrenia and 9 healthy controls were asked to remotely participate in smartphone-based assessments daily for 14 days. Video recorded from the smartphone front-facing camera during these assessments was used to quantify the Euclidean distance of head movement between frames through a pretrained computer vision model. The ability of head movement measurements to distinguish between patients and healthy controls as well as their relationship to schizophrenia symptom severity as measured through traditional clinical scores was assessed. Results The rate of head movement in participants with schizophrenia (1.48 mm/frame) and those without differed significantly (2.50 mm/frame; p = 0.01), and a logistic regression demonstrated that head movement was a significant predictor of schizophrenia diagnosis (p = 0.02). Linear regression between head movement and clinical scores of schizophrenia showed that head movement has a negative relationship with schizophrenia symptom severity (p = 0.04), primarily with negative symptoms of schizophrenia. Conclusions Remote, smartphone-based assessments were able to capture meaningful visual behavior for computer vision-based objective measurement of head movement. The measurements of head movement acquired were able to accurately classify schizophrenia diagnosis and quantify symptom severity in patients with schizophrenia.
Collapse
Affiliation(s)
| | | | - Emma Smith
- Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Elizabeth Ramjas
- Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Sarah B Rutter
- Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | | | | | - Li Zhang
- AiCure, LLC, New York, New York, USA
| | - Lei Guan
- AiCure, LLC, New York, New York, USA
| | - Paul Rosenfield
- Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | | | - Isaac R Galatzer-Levy
- AiCure, LLC, New York, New York, USA.,Psychiatry, New York University School of Medicine, New York, New York, USA
| |
Collapse
|
37
|
Zhuang Y, McDonald MM, Aldridge CM, Hassan MA, Uribe O, Arteaga D, Southerland AM, Rohde GK. Video-Based Facial Weakness Analysis. IEEE Trans Biomed Eng 2021; 68:2698-2705. [PMID: 33406036 DOI: 10.1109/tbme.2021.3049739] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
OBJECTIVE Facial weakness is a common sign of neurological diseases such as Bell's palsy and stroke. However, recognizing facial weakness still remains as a challenge, because it requires experience and neurological training. METHODS We propose a framework for facial weakness detection, which models the temporal dynamics of both shape and appearance-based features of each target frame through a bi-directional long short-term memory network (Bi-LSTM). The system is evaluated on a "in-the-wild"video dataset that is verified by three board-certified neurologists. In addition, three emergency medical services (EMS) personnel and three upper level residents rated the dataset. We compare the evaluation of the proposed algorithm with other comparison methods as well as the human raters. RESULTS Experimental evaluation demonstrates that: (1) the proposed algorithm achieves the accuracy, sensitivity, and specificity of 94.3%, 91.4%, and 95.7%, which outperforms other comparison methods and achieves the equal performance to paramedics; (2) the framework can provide visualizable and interpretable results that increases model transparency and interpretability; (3) a prototype is implemented as a proof-of-concept showcase to show the feasibility of an inexpensive solution for facial weakness detection. CONCLUSION The experiment results suggest that the proposed framework can identify facial weakness effectively. SIGNIFICANCE We provide a proof-of-concept study, showing that such technology could be used by non-neurologists to more readily identify facial weakness in the field, leading to increasing coverage and earlier treatment.
Collapse
|
38
|
Tang C, Zheng W, Zong Y, Qiu N, Lu C, Zhang X, Ke X, Guan C. Automatic Identification of High-Risk Autism Spectrum Disorder: A Feasibility Study Using Video and Audio Data Under the Still-Face Paradigm. IEEE Trans Neural Syst Rehabil Eng 2020; 28:2401-2410. [PMID: 32991285 DOI: 10.1109/tnsre.2020.3027756] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
It is reported that the symptoms of autism spectrum disorder (ASD) could be improved by effective early interventions, which arouses an urgent need for large-scale early identification of ASD. Until now, the screening of ASD has relied on the child psychiatrist to collect medical history and conduct behavioral observations with the help of psychological assessment tools. Such screening measures inevitably have some disadvantages, including strong subjectivity, relying on experts and low-efficiency. With the development of computer science, it is possible to realize a computer-aided screening for ASD and alleviate the disadvantages of manual evaluation. In this study, we propose a behavior-based automated screening method to identify high-risk ASD (HR-ASD) for babies aged 8-24 months. The still-face paradigm (SFP) was used to elicit baby's spontaneous social behavior through a face-to-face interaction, in which a mother was required to maintain a normal interaction to amuse her baby for 2 minutes (a baseline episode) and then suddenly change to the no-reaction and no-expression status with 1 minute (a still-face episode). Here, multiple cues derived from baby's social stress response behavior during the latter episode, including head-movements, facial expressions and vocal characteristics, were statistically analyzed between HR-ASD and typical developmental (TD) groups. An automated identification model of HR-ASD was constructed based on these multi-cue features and the support vector machine (SVM) classifier; moreover, its screening performance was satisfied, for all the accuracy, specificity and sensitivity exceeded 90% on the cases included in this study. The experimental results suggest its feasibility in the early screening of HR-ASD.
Collapse
|
39
|
de Belen RAJ, Bednarz T, Sowmya A, Del Favero D. Computer vision in autism spectrum disorder research: a systematic review of published studies from 2009 to 2019. Transl Psychiatry 2020; 10:333. [PMID: 32999273 PMCID: PMC7528087 DOI: 10.1038/s41398-020-01015-w] [Citation(s) in RCA: 29] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/05/2020] [Revised: 09/04/2020] [Accepted: 09/09/2020] [Indexed: 11/29/2022] Open
Abstract
The current state of computer vision methods applied to autism spectrum disorder (ASD) research has not been well established. Increasing evidence suggests that computer vision techniques have a strong impact on autism research. The primary objective of this systematic review is to examine how computer vision analysis has been useful in ASD diagnosis, therapy and autism research in general. A systematic review of publications indexed on PubMed, IEEE Xplore and ACM Digital Library was conducted from 2009 to 2019. Search terms included ['autis*' AND ('computer vision' OR 'behavio* imaging' OR 'behavio* analysis' OR 'affective computing')]. Results are reported according to PRISMA statement. A total of 94 studies are included in the analysis. Eligible papers are categorised based on the potential biological/behavioural markers quantified in each study. Then, different computer vision approaches that were employed in the included papers are described. Different publicly available datasets are also reviewed in order to rapidly familiarise researchers with datasets applicable to their field and to accelerate both new behavioural and technological work on autism research. Finally, future research directions are outlined. The findings in this review suggest that computer vision analysis is useful for the quantification of behavioural/biological markers which can further lead to a more objective analysis in autism research.
Collapse
Affiliation(s)
| | - Tomasz Bednarz
- School of Art & Design, University of New South Wales, Sydney, NSW, Australia
| | - Arcot Sowmya
- School of Computer Science and Engineering, University of New South Wales, Sydney, NSW, Australia
| | - Dennis Del Favero
- School of Art & Design, University of New South Wales, Sydney, NSW, Australia
| |
Collapse
|
40
|
Xin X, Lin X, Yang S, Zheng X. Pain intensity estimation based on a spatial transformation and attention CNN. PLoS One 2020; 15:e0232412. [PMID: 32822348 PMCID: PMC7444520 DOI: 10.1371/journal.pone.0232412] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2019] [Accepted: 04/14/2020] [Indexed: 02/05/2023] Open
Abstract
Models designed to detect abnormalities that reflect disease from facial structures are an emerging area of research for automated facial analysis, which has important potential value in smart healthcare applications. However, most of the proposed models directly analyze the whole face image containing the background information, and rarely consider the effects of the background and different face regions on the analysis results. Therefore, in view of these effects, we propose an end-to-end attention network with spatial transformation to estimate different pain intensities. In the proposed method, the face image is first provided as input to a spatial transformation network for solving the problem of background interference; then, the attention mechanism is used to adaptively adjust the weights of different face regions of the transformed face image; finally, a convolutional neural network (CNN) containing a Softmax function is utilized to classify the pain levels. The extensive experiments and analysis are conducted on the benchmarking and publicly available database, namely the UNBC-McMaster shoulder pain. More specifically, in order to verify the superiority of our proposed method, the comparisons with the basic CNNs and the-state-of-the-arts are performed, respectively. The experiments show that the introduced spatial transformation and attention mechanism in our method can significantly improve the estimation performances and outperform the-state-of-the-arts.
Collapse
Affiliation(s)
- Xuwu Xin
- The Second Affiliated Hospital of Shantou University Medical College, Shantou, China
| | - Xiaoyan Lin
- The Second Affiliated Hospital of Shantou University Medical College, Shantou, China
| | - Shengfu Yang
- The First Affiliated Hospital of Jinan University, Guangzhou, China
| | - Xin Zheng
- Shantou Chaonan Minsheng Hospital, Shantou, China
| |
Collapse
|
41
|
Islam MZ, Islam MM, Asraf A. A combined deep CNN-LSTM network for the detection of novel coronavirus (COVID-19) using X-ray images. INFORMATICS IN MEDICINE UNLOCKED 2020; 20:100412. [PMID: 32835084 DOI: 10.1101/2020.06.18.20134718] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2020] [Revised: 08/07/2020] [Accepted: 08/07/2020] [Indexed: 05/27/2023] Open
Abstract
Nowadays, automatic disease detection has become a crucial issue in medical science due to rapid population growth. An automatic disease detection framework assists doctors in the diagnosis of disease and provides exact, consistent, and fast results and reduces the death rate. Coronavirus (COVID-19) has become one of the most severe and acute diseases in recent times and has spread globally. Therefore, an automated detection system, as the fastest diagnostic option, should be implemented to impede COVID-19 from spreading. This paper aims to introduce a deep learning technique based on the combination of a convolutional neural network (CNN) and long short-term memory (LSTM) to diagnose COVID-19 automatically from X-ray images. In this system, CNN is used for deep feature extraction and LSTM is used for detection using the extracted feature. A collection of 4575 X-ray images, including 1525 images of COVID-19, were used as a dataset in this system. The experimental results show that our proposed system achieved an accuracy of 99.4%, AUC of 99.9%, specificity of 99.2%, sensitivity of 99.3%, and F1-score of 98.9%. The system achieved desired results on the currently available dataset, which can be further improved when more COVID-19 images become available. The proposed system can help doctors to diagnose and treat COVID-19 patients easily.
Collapse
Affiliation(s)
- Md Zabirul Islam
- Department of Computer Science and Engineering, Khulna University of Engineering & Technology, Khulna, 9203, Bangladesh
| | - Md Milon Islam
- Department of Computer Science and Engineering, Khulna University of Engineering & Technology, Khulna, 9203, Bangladesh
| | - Amanullah Asraf
- Department of Computer Science and Engineering, Khulna University of Engineering & Technology, Khulna, 9203, Bangladesh
| |
Collapse
|
42
|
Islam MZ, Islam MM, Asraf A. A combined deep CNN-LSTM network for the detection of novel coronavirus (COVID-19) using X-ray images. INFORMATICS IN MEDICINE UNLOCKED 2020; 20:100412. [PMID: 32835084 PMCID: PMC7428728 DOI: 10.1016/j.imu.2020.100412] [Citation(s) in RCA: 207] [Impact Index Per Article: 51.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2020] [Revised: 08/07/2020] [Accepted: 08/07/2020] [Indexed: 12/15/2022] Open
Abstract
Nowadays, automatic disease detection has become a crucial issue in medical science due to rapid population growth. An automatic disease detection framework assists doctors in the diagnosis of disease and provides exact, consistent, and fast results and reduces the death rate. Coronavirus (COVID-19) has become one of the most severe and acute diseases in recent times and has spread globally. Therefore, an automated detection system, as the fastest diagnostic option, should be implemented to impede COVID-19 from spreading. This paper aims to introduce a deep learning technique based on the combination of a convolutional neural network (CNN) and long short-term memory (LSTM) to diagnose COVID-19 automatically from X-ray images. In this system, CNN is used for deep feature extraction and LSTM is used for detection using the extracted feature. A collection of 4575 X-ray images, including 1525 images of COVID-19, were used as a dataset in this system. The experimental results show that our proposed system achieved an accuracy of 99.4%, AUC of 99.9%, specificity of 99.2%, sensitivity of 99.3%, and F1-score of 98.9%. The system achieved desired results on the currently available dataset, which can be further improved when more COVID-19 images become available. The proposed system can help doctors to diagnose and treat COVID-19 patients easily.
Collapse
Affiliation(s)
- Md Zabirul Islam
- Department of Computer Science and Engineering, Khulna University of Engineering & Technology, Khulna, 9203, Bangladesh
| | - Md Milon Islam
- Department of Computer Science and Engineering, Khulna University of Engineering & Technology, Khulna, 9203, Bangladesh
| | - Amanullah Asraf
- Department of Computer Science and Engineering, Khulna University of Engineering & Technology, Khulna, 9203, Bangladesh
| |
Collapse
|
43
|
Shabaan M, Arshid K, Yaqub M, Jinchao F, Zia MS, Bojja GR, Iftikhar M, Ghani U, Ambati LS, Munir R. Survey: smartphone-based assessment of cardiovascular diseases using ECG and PPG analysis. BMC Med Inform Decis Mak 2020; 20:177. [PMID: 32727453 PMCID: PMC7392662 DOI: 10.1186/s12911-020-01199-7] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2020] [Accepted: 07/22/2020] [Indexed: 11/17/2022] Open
Abstract
A number of resources, every year, being spent to tackle early detection of cardiac abnormalities which is one of the leading causes of deaths all over the Globe. The challenges for healthcare systems includes early detection, portability and mobility of patients. This paper presents a categorical review of smartphone-based systems that can detect cardiac abnormalities by the analysis of Electrocardiogram (ECG) and Photoplethysmography (PPG) and the limitation and challenges of these system. The ECG based systems can monitor, record and forward signals for analysis and an alarm can be triggered in case of abnormality, however the limitation of smart phone’s processing capabilities, lack of storage and speed of network are major challenges. The systems based on PPG signals are non-invasive and provides mobility and portability. This study aims to critically review the existing systems, their limitation, challenges and possible improvements to serve as a reference for researchers and developers.
Collapse
Affiliation(s)
| | - Kaleem Arshid
- Beijing University of Technology, Chaoyang District, Beijing, China
| | - Muhammad Yaqub
- Beijing University of Technology, Chaoyang District, Beijing, China
| | - Feng Jinchao
- Beijing University of Technology, Chaoyang District, Beijing, China. .,Beijing Laboratory of Advanced Information Networks, Beijing, China.
| | - M Sultan Zia
- The University of Lahore, Gujarat Campus, Gujarat, Pakistan
| | | | | | - Usman Ghani
- Punjab Education Department, Gugarat, Pakistan
| | | | - Rizwan Munir
- Beijing University of Post and Telecommunication, Beijing, China
| |
Collapse
|
44
|
Ismail LI, Hanapiah FA, Belpaeme T, Dambre J, Wyffels F. Analysis of Attention in Child–Robot Interaction Among Children Diagnosed with Cognitive Impairment. Int J Soc Robot 2020. [DOI: 10.1007/s12369-020-00628-x] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/30/2023]
|
45
|
Ahmedt-Aristizabal D, Nguyen K, Denman S, Sarfraz MS, Sridharan S, Dionisio S, Fookes C. Vision-Based Mouth Motion Analysis in Epilepsy: A 3D Perspective. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2019:1625-1629. [PMID: 31946208 DOI: 10.1109/embc.2019.8857656] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Epilepsy monitoring involves the study of videos to assess clinical signs (semiology) to assist with the diagnosis of seizures. Recent advances in the application of vision-based approaches to epilepsy analysis have demonstrated significant potential to automate this assessment. Nevertheless, current proposed computer vision based techniques are unable to accurately quantify specific facial modifications, e.g. mouth motions, which are examined by neurologists to distinguish between seizure types. 2D approaches that analyse facial landmarks have been proposed to quantify mouth motions, however, they are unable to fully represent motions in the mouth and cheeks (ictal pouting) due to a lack of landmarks in the the cheek regions. Additionally, 2D region-based techniques based on the detection of the mouth have limitations when dealing with large pose variations, and thus make a fair comparison between samples difficult due to the variety of poses present. 3D approaches, on the other hand, retain rich information about the shape and appearance of faces, simplifying alignment for comparison between sequences. In this paper, we propose a novel network method based on a 3D reconstruction of the face and deep learning to detect and quantify mouth semiology in our video dataset of 20 seizures, recorded from patients with mesial temporal and extra-temporal lobe epilepsy. The proposed network is capable of distinguishing between seizures of both types of epilepsy. An average classification accuracy of 89% demonstrates the benefits of computer vision and deep learning for clinical applications of non-contact systems to identify semiology commonly encountered in a natural clinical setting.
Collapse
|
46
|
Zhuang Y, McDonald M, Uribe O, Yin X, Parikh D, Southerland AM, Rohde GK. Facial Weakness Analysis and Quantification of Static Images. IEEE J Biomed Health Inform 2020; 24:2260-2267. [PMID: 31944968 DOI: 10.1109/jbhi.2020.2964520] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
Facial weakness is a symptom commonly associated to lack of facial muscle control due to neurological injury. Several diseases are associated with facial weakness such as stroke and Bell's palsy. The use of digital imaging through mobile phones, tablets, personal computers and other devices could provide timely opportunity for detection, which if accurate enough can improve treatment by enabling faster patient triage and recovery progress monitoring. Most of the existing facial weakness detection approaches from static images are based on facial landmarks from which geometric features can be calculated. Landmark-based methods, however, can suffer from inaccuracies in face landmarks localization. In this study, We also experimentally evaluate the performance of several feature extraction methods for measuring facial weakness, including the landmark-based features, as well as intensity-based features on a neurologist-certified dataset that comprises 186 images of normal, 125 images of left facial weakness, and 126 images of right facial weakness. We demonstrate that, for the application of facial weakness detection from single (static) images, approaches that incorporate the Histogram of Oriented Gradients (HoG) features tend to be more accurate.
Collapse
|
47
|
Sepas‐Moghaddam A, Pereira FM, Correia PL. Face recognition: a novel multi‐level taxonomy based survey. IET BIOMETRICS 2019. [DOI: 10.1049/iet-bmt.2019.0001] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022] Open
Affiliation(s)
- Alireza Sepas‐Moghaddam
- Instituto de Telecomunicações, Instituto Superior Técnico – Universidade de LisboaLisbonPortugal
| | - Fernando M. Pereira
- Instituto de Telecomunicações, Instituto Superior Técnico – Universidade de LisboaLisbonPortugal
| | - Paulo Lobato Correia
- Instituto de Telecomunicações, Instituto Superior Técnico – Universidade de LisboaLisbonPortugal
| |
Collapse
|
48
|
Yolcu G, Oztel I, Kazan S, Oz C, Palaniappan K, Lever TE, Bunyak F. Facial expression recognition for monitoring neurological disorders based on convolutional neural network. MULTIMEDIA TOOLS AND APPLICATIONS 2019; 78:31581-31603. [PMID: 35693322 PMCID: PMC9181900 DOI: 10.1007/s11042-019-07959-6] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/06/2018] [Revised: 06/13/2019] [Accepted: 07/10/2019] [Indexed: 06/15/2023]
Abstract
Facial expressions are a significant part of non-verbal communication. Recognizing facial expressions of people with neurological disorders is essential because these people may have lost a significant amount of their verbal communication ability. Such an assessment requires time consuming examination involving medical personnel, which can be quite challenging and expensive. Automated facial expression recognition systems that are low-cost and noninvasive can help experts detect neurological disorders. In this study, an automated facial expression recognition system is developed using a novel deep learning approach. The architecture consists of four-stage networks. The first, second and third networks segment the facial components which are essential for facial expression recognition. Owing to the three networks, an iconize facial image is obtained. The fourth network classifies facial expressions using raw facial images and iconize facial images. This four-stage method combines holistic facial information with local part-based features to achieve more robust facial expression recognition. Preliminary experimental results achieved 94.44% accuracy for facial expression recognition on RaFD database. The proposed system produced 5% improvement than the facial expression recognition system by using raw images. This study presents a quantitative, objective and non-invasive facial expression recognition system to help in the monitoring and diagnosis of neurological disorders influencing facial expressions.
Collapse
Affiliation(s)
- Gozde Yolcu
- Department of Computer Engineering, Sakarya University, 54050 Serdivan, Sakarya, Turkey
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO 65211, USA
| | - Ismail Oztel
- Department of Computer Engineering, Sakarya University, 54050 Serdivan, Sakarya, Turkey
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO 65211, USA
| | - Serap Kazan
- Department of Computer Engineering, Sakarya University, 54050 Serdivan, Sakarya, Turkey
| | - Cemil Oz
- Department of Computer Engineering, Sakarya University, 54050 Serdivan, Sakarya, Turkey
| | - Kannappan Palaniappan
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO 65211, USA
| | - Teresa E. Lever
- Department of Otolaryngology, University of Missouri, Columbia, MO 65211, USA
| | - Filiz Bunyak
- Department of Electrical Engineering and Computer Science, University of Missouri, Columbia, MO 65211, USA
| |
Collapse
|
49
|
Finka LR, Luna SP, Brondani JT, Tzimiropoulos Y, McDonagh J, Farnworth MJ, Ruta M, Mills DS. Geometric morphometrics for the study of facial expressions in non-human animals, using the domestic cat as an exemplar. Sci Rep 2019; 9:9883. [PMID: 31285531 PMCID: PMC6614427 DOI: 10.1038/s41598-019-46330-5] [Citation(s) in RCA: 27] [Impact Index Per Article: 5.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2019] [Accepted: 06/25/2019] [Indexed: 11/09/2022] Open
Abstract
Facial expression is a common channel for the communication of emotion. However, in the case of non-human animals, the analytical methods used to quantify facial expressions can be subjective, relying heavily on extrapolation from human-based systems. Here, we demonstrate how geometric morphometrics can be applied in order to overcome these problems. We used this approach to identify and quantify changes in facial shape associated with pain in a non-human animal species. Our method accommodates individual variability, species-specific facial anatomy, and postural effects. Facial images were captured at four different time points during ovariohysterectomy of domestic short haired cats (n = 29), with time points corresponding to varying intensities of pain. Images were annotated using landmarks specifically chosen for their relationship with underlying musculature, and relevance to cat-specific facial action units. Landmark data were subjected to normalisation before Principal Components (PCs) were extracted to identify key sources of facial shape variation, relative to pain intensity. A significant relationship between PC scores and a well-validated composite measure of post-operative pain in cats (UNESP-Botucatu MCPS tool) was evident, demonstrating good convergent validity between our geometric face model, and other metrics of pain detection. This study lays the foundation for the automatic, objective detection of emotional expressions in a range of non-human animal species.
Collapse
Affiliation(s)
- Lauren R Finka
- School of Life Sciences, Joseph Bank Laboratories, University of Lincoln, Lincoln, LN6 7DL, UK. .,Animal, Rural and Environmental Sciences, Nottingham Trent University, Southwell, NG25 0QF, UK.
| | - Stelio P Luna
- School of Veterinary Medicine and Animal Science, São Paulo State University (Unesp), São Paulo, 18618-970, Brazil
| | - Juliana T Brondani
- School of Veterinary Medicine and Animal Science, São Paulo State University (Unesp), São Paulo, 18618-970, Brazil
| | | | - John McDonagh
- School of Computer Science, University of Nottingham, Nottingham, NG8 1BB, UK
| | - Mark J Farnworth
- Animal, Rural and Environmental Sciences, Nottingham Trent University, Southwell, NG25 0QF, UK
| | - Marcello Ruta
- School of Life Sciences, Joseph Bank Laboratories, University of Lincoln, Lincoln, LN6 7DL, UK
| | - Daniel S Mills
- School of Life Sciences, Joseph Bank Laboratories, University of Lincoln, Lincoln, LN6 7DL, UK
| |
Collapse
|