1
|
Hyun Choi D, Ha Joo Y, Hong Kim K, Ho Park J, Joo H, Kong HJ, Lee H, Jun Song K, Kim S. A Development of a Sound Recognition-Based Cardiopulmonary Resuscitation Training System. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE 2024; 12:550-557. [PMID: 39155923 PMCID: PMC11329221 DOI: 10.1109/jtehm.2024.3433448] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/13/2024] [Revised: 07/06/2024] [Accepted: 07/22/2024] [Indexed: 08/20/2024]
Abstract
The objective of this study was to develop a sound recognition-based cardiopulmonary resuscitation (CPR) training system that is accessible, cost-effective, easy-to-maintain and provides accurate CPR feedback. Beep-CPR, a novel device with accordion squeakers that emit high-pitched sounds during compression, was developed. The sounds emitted by Beep-CPR were recorded using a smartphone, segmented into 2-second audio fragments, and then transformed into spectrograms. A total of 6,065 spectrograms were generated from approximately 40 minutes of audio data, which were then randomly split into training, validation, and test datasets. Each spectrogram was matched with the depth, rate, and release velocity of the compression measured at the same time interval by the ZOLL X Series monitor/defibrillator. Deep learning models utilizing spectrograms as input were trained using transfer learning based on EfficientNet to predict the depth (Depth model), rate (Rate model), and release velocity (Recoil model) of compressions. Results: The mean absolute error (MAE) for the Depth model was 0.30 cm (95% confidence interval [CI]: 0.27-0.33). The MAE of the Rate model was 3.6/min (95% CI: 3.2-3.9). For the Recoil model, the MAE was 2.3 cm/s (95% CI: 2.1-2.5). External validation of the models demonstrated acceptable performance across multiple conditions, including the utilization of a newly-manufactured device, a fatigued device, and evaluation in an environment with altered spatial dimensions. We have developed a novel sound recognition-based CPR training system, that accurately measures compression quality during training. Significance: Beep-CPR is a cost-effective and easy-to-maintain solution that can improve the efficacy of CPR training by facilitating decentralized at-home training with performance feedback.
Collapse
Affiliation(s)
- Dong Hyun Choi
- Department of Biomedical EngineeringSeoul National University College of MedicineJongnoSeoul03080South Korea
| | - Yoon Ha Joo
- Biomedical Research InstituteSeoul National University HospitalJongnoSeoul03080South Korea
- Department of Emergency MedicineSeoul National University HospitalJongnoSeoul03080South Korea
| | - Ki Hong Kim
- Department of Emergency MedicineSeoul National University HospitalJongnoSeoul03080South Korea
| | - Jeong Ho Park
- Department of Emergency MedicineSeoul National University HospitalJongnoSeoul03080South Korea
| | - Hyunjin Joo
- Innovative Medical Technology Research InstituteSeoul National University HospitalJongnoSeoul03080South Korea
| | - Hyoun-Joong Kong
- Department of Transdisciplinary MedicineInnovative Medical Technology Research InstituteSeoul National University HospitalJongnoSeoul03080South Korea
- Department of MedicineSeoul National University College of MedicineJongnoSeoul03080South Korea
| | - Hyunju Lee
- Laboratory of Emergency Medical ServicesBiomedical Research Institute, Seoul National University HospitalJongnoSeoul03080South Korea
| | - Kyoung Jun Song
- Department of Emergency MedicineSeoul Metropolitan Boramae Medical CenterDongjakSeoul07061South Korea
| | - Sungwan Kim
- Department of Biomedical EngineeringSeoul National University College of MedicineJongnoSeoul03080South Korea
| |
Collapse
|
2
|
Rajeshkumar C, Soundar KR. TO-LAB model: Real time Touchless Lung Abnormality detection model using USRP based machine learning algorithm. Technol Health Care 2024:THC240149. [PMID: 38968032 DOI: 10.3233/thc-240149] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/07/2024]
Abstract
BACKGROUND Due to the increasing prevalence of respiratory diseases and the importance of early diagnosis. The need for non-invasive and touchless medical diagnostic solutions has become increasingly crucial in modern healthcare to detect lung abnormalities. OBJECTIVE Existing methods for lung abnormality detection often rely on invasive and time-consuming procedures limiting their effectiveness in real-time diagnosis. This work introduces a novel Touchless Lung Abnormality (TO-LAB) detection model utilizing universal software radio peripherals (USRP) and machine learning algorithms. METHODS The TO-LAB model integrates a blood pressure meter and an RGB-D depth-sensing camera to gather individual data without physical contact. Heart rate (HR) is analyzed through image conversion to IPPG signals, while blood pressure (BP) is obtained via analog conversion from the blood pressure meter. This touchless imaging setup facilitates the extraction of essential signal features crucial for respiratory pattern analysis. Advanced computer vision algorithms like Mel-frequency cepstral coefficients (MFCC) and Principal Component Analysis (PCA) process the acquired data to focus on breathing abnormalities. These features are then combined and inputted into a machine learning-based Multi-class SVM for breathing activity analysis. The Multi-class SVM categorizes breathing abnormalities as normal, shallow, or elevated based on the fused features. The efficiency of this TO-LAB model is evaluated with the simulated and real-time data. RESULTS According to the findings, the proposed TO-LAB model attains the maximum accuracy of 96.15% for real time data; however, the accuracy increases to 99.54% for simulated data for the efficient classification of breathing abnormalities. CONCLUSION From this analysis, our model attains better results in simulated data but it declines the accuracy while processing with real-time data. Moreover, this work has a significant medical impact since it presents a solution to the problem of gathering enough data during the epidemic to create a realistic model with a large dataset.
Collapse
Affiliation(s)
- C Rajeshkumar
- Department of Information Technology, Sri Krishna College of Technology, Coimbatore, India
| | - K Ruba Soundar
- Department of Computer Science and Engineering, Mepco Schlenk Engineering College, Sivakasi, India
| |
Collapse
|
3
|
Ghaderinia M, Abadijoo H, Mahdavian A, Kousha E, Shakibi R, Taheri SMR, Simaee H, Khatibi A, Moosavi-Movahedi AA, Khayamian MA. Smartphone-based device for point-of-care diagnostics of pulmonary inflammation using convolutional neural networks (CNNs). Sci Rep 2024; 14:6912. [PMID: 38519489 PMCID: PMC10959990 DOI: 10.1038/s41598-024-54939-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Accepted: 02/19/2024] [Indexed: 03/25/2024] Open
Abstract
In pulmonary inflammation diseases, like COVID-19, lung involvement and inflammation determine the treatment regime. Respiratory inflammation is typically arisen due to the cytokine storm and the leakage of the vessels for immune cells recruitment. Currently, such a situation is detected by the clinical judgment of a specialist or precisely by a chest CT scan. However, the lack of accessibility to the CT machines in many poor medical centers as well as its expensive service, demands more accessible methods for fast and cheap detection of lung inflammation. Here, we have introduced a novel method for tracing the inflammation and lung involvement in patients with pulmonary inflammation, such as COVID-19, by a simple electrolyte detection in their sputum samples. The presence of the electrolyte in the sputum sample results in the fern-like structures after air-drying. These fern patterns are different in the CT positive and negative cases that are detected by an AI application on a smartphone and using a low-cost and portable mini-microscope. Evaluating 160 patient-derived sputum sample images, this method demonstrated an interesting accuracy of 95%, as confirmed by CT-scan results. This finding suggests that the method has the potential to serve as a promising and reliable approach for recognizing lung inflammatory diseases, such as COVID-19.
Collapse
Affiliation(s)
- Mohammadreza Ghaderinia
- Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran
- Integrated Biophysics and Bioengineering Lab (iBL), Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran
- Nano Electronic Center of Excellence, Nano Bio Electronics Devices Lab, School of Electrical and Computer Engineering, University of Tehran, P.O. Box 14395/515, Tehran, Iran
| | - Hamed Abadijoo
- Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran
- Integrated Biophysics and Bioengineering Lab (iBL), Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran
- Nano Electronic Center of Excellence, Nano Bio Electronics Devices Lab, School of Electrical and Computer Engineering, University of Tehran, P.O. Box 14395/515, Tehran, Iran
| | - Ashkan Mahdavian
- Nano Electronic Center of Excellence, Nano Bio Electronics Devices Lab, School of Electrical and Computer Engineering, University of Tehran, P.O. Box 14395/515, Tehran, Iran
| | - Ebrahim Kousha
- Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran
- Integrated Biophysics and Bioengineering Lab (iBL), Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran
- Nano Electronic Center of Excellence, Nano Bio Electronics Devices Lab, School of Electrical and Computer Engineering, University of Tehran, P.O. Box 14395/515, Tehran, Iran
| | - Reyhaneh Shakibi
- Department of Biophysics, Faculty of Biological Sciences, Tarbiat Modares University, Tehran, Iran
| | - S Mohammad-Reza Taheri
- Groningen university, University medical center Groningen, Antonius Deusinglaan 1, 9713AW, Groningen, The Netherlands
- Condensed Matter National Laboratory, Institute for Research in Fundamental Sciences (IPM), Tehran, Iran
| | - Hossein Simaee
- Cardiac Primary Prevention Research Center, Cardiovascular Diseases Research Institute, Tehran University of Medical Sciences, Tehran, Iran
| | - Ali Khatibi
- Department of Biotechnology, Faculty of Biological Sciences, Alzahra University, Tehran, Iran
| | | | - Mohammad Ali Khayamian
- Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran.
- Integrated Biophysics and Bioengineering Lab (iBL), Institute of Biochemistry and Biophysics, University of Tehran, Tehran, 1417614335, Iran.
| |
Collapse
|
4
|
Kapetanidis P, Kalioras F, Tsakonas C, Tzamalis P, Kontogiannis G, Karamanidou T, Stavropoulos TG, Nikoletseas S. Respiratory Diseases Diagnosis Using Audio Analysis and Artificial Intelligence: A Systematic Review. SENSORS (BASEL, SWITZERLAND) 2024; 24:1173. [PMID: 38400330 PMCID: PMC10893010 DOI: 10.3390/s24041173] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/15/2023] [Revised: 02/03/2024] [Accepted: 02/04/2024] [Indexed: 02/25/2024]
Abstract
Respiratory diseases represent a significant global burden, necessitating efficient diagnostic methods for timely intervention. Digital biomarkers based on audio, acoustics, and sound from the upper and lower respiratory system, as well as the voice, have emerged as valuable indicators of respiratory functionality. Recent advancements in machine learning (ML) algorithms offer promising avenues for the identification and diagnosis of respiratory diseases through the analysis and processing of such audio-based biomarkers. An ever-increasing number of studies employ ML techniques to extract meaningful information from audio biomarkers. Beyond disease identification, these studies explore diverse aspects such as the recognition of cough sounds amidst environmental noise, the analysis of respiratory sounds to detect respiratory symptoms like wheezes and crackles, as well as the analysis of the voice/speech for the evaluation of human voice abnormalities. To provide a more in-depth analysis, this review examines 75 relevant audio analysis studies across three distinct areas of concern based on respiratory diseases' symptoms: (a) cough detection, (b) lower respiratory symptoms identification, and (c) diagnostics from the voice and speech. Furthermore, publicly available datasets commonly utilized in this domain are presented. It is observed that research trends are influenced by the pandemic, with a surge in studies on COVID-19 diagnosis, mobile data acquisition, and remote diagnosis systems.
Collapse
Affiliation(s)
- Panagiotis Kapetanidis
- Computer Engineering and Informatics Department, University of Patras, 26504 Patras, Greece (C.T.); (G.K.); (S.N.)
| | - Fotios Kalioras
- Computer Engineering and Informatics Department, University of Patras, 26504 Patras, Greece (C.T.); (G.K.); (S.N.)
| | - Constantinos Tsakonas
- Computer Engineering and Informatics Department, University of Patras, 26504 Patras, Greece (C.T.); (G.K.); (S.N.)
| | - Pantelis Tzamalis
- Computer Engineering and Informatics Department, University of Patras, 26504 Patras, Greece (C.T.); (G.K.); (S.N.)
| | - George Kontogiannis
- Computer Engineering and Informatics Department, University of Patras, 26504 Patras, Greece (C.T.); (G.K.); (S.N.)
| | - Theodora Karamanidou
- Pfizer Center for Digital Innovation, 55535 Thessaloniki, Greece; (T.K.); (T.G.S.)
| | | | - Sotiris Nikoletseas
- Computer Engineering and Informatics Department, University of Patras, 26504 Patras, Greece (C.T.); (G.K.); (S.N.)
| |
Collapse
|
5
|
Abdul Sattar Shaikh A, Bhargavi MS, Kumar C P. Weighted aggregation through probability based ranking: An optimized federated learning architecture to classify respiratory diseases. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 242:107821. [PMID: 37776709 DOI: 10.1016/j.cmpb.2023.107821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Revised: 08/29/2023] [Accepted: 09/15/2023] [Indexed: 10/02/2023]
Abstract
Background and Objective Respiratory Diseases are one of the leading chronic illnesses in the world according to the reports by World Health Organization. Diagnosing these respiratory diseases is done through auscultation where a medical professional listens to sounds of air in the lungs for anomalies through a stethoscope. This method necessitates extensive experience and can also be misinterpreted by the medical professional. To address this issue, we introduce an AI-based solution that listens to the lung sounds and classifies the respiratory disease detected. Since the research work deals with medical data that is tightly under wraps due to privacy concerns in the medical field, we introduce a Deep learning solution to classify the diseases and a custom Federated learning (FL) approach to further improve the accuracy of the deep learning model and simultaneously maintain data privacy. Federated Learning architecture maintains data privacy and facilitates a distributed learning system for medical infrastructures. Methods The approach utilizes Generative Adversarial Networks (GAN) based Federated learning approach to ensure data privacy. Generative Adversarial Networks generate new data by synthesizing new lung sounds. This new synthesized data is then converted to spectrograms and trained on a neural network to classify four lung diseases, Heart Attack and Normal breathing patterns. Furthermore, to address performance loss during FL, we also propose a new "Weighted Aggregation through Probability-based Ranking (FedWAPR)" algorithm for optimizing the FL aggregation process. The FedWAPR aggregation takes inspiration from exponential distribution function and ranks better performing clients according to it. Results and Conclusion A test accuracy of about 92% was achieved by the trained model while classifying various respiratory diseases and heart failure. Additionally, we developed a novel FedWAPR approach that significantly outperformed the FedAVG approach for the FL aggregate function. A patient can be checked for respiratory diseases using this improved learning approach without the need for extensive sensitive data recording or for making sure the data sample obtained is secure. In a decentralized training runtime, the trained model successfully classifies various respiratory diseases and heart failure using lung sounds with a test accuracy on par with a centralized model.
Collapse
Affiliation(s)
- Abdullah Abdul Sattar Shaikh
- Department of Computer Science and Engineering, Bangalore Institute of Technology, Bangalore, 560004, Karnataka, India.
| | - M S Bhargavi
- Department of Computer Science and Engineering, Bangalore Institute of Technology, Bangalore, 560004, Karnataka, India.
| | - Pavan Kumar C
- Department of Computer Science and Engineering, Indian Institute of Information Technology Dharwad, Dharwad, 580009, Karnataka, India.
| |
Collapse
|
6
|
Alvarado E, Grágeda N, Luzanto A, Mahu R, Wuth J, Mendoza L, Stern RM, Yoma NB. Automatic Detection of Dyspnea in Real Human-Robot Interaction Scenarios. SENSORS (BASEL, SWITZERLAND) 2023; 23:7590. [PMID: 37688044 PMCID: PMC10490721 DOI: 10.3390/s23177590] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 08/20/2023] [Accepted: 08/25/2023] [Indexed: 09/10/2023]
Abstract
A respiratory distress estimation technique for telephony previously proposed by the authors is adapted and evaluated in real static and dynamic HRI scenarios. The system is evaluated with a telephone dataset re-recorded using the robotic platform designed and implemented for this study. In addition, the original telephone training data are modified using an environmental model that incorporates natural robot-generated and external noise sources and reverberant effects using room impulse responses (RIRs). The results indicate that the average accuracy and AUC are just 0.4% less than those obtained with matched training/testing conditions with simulated data. Quite surprisingly, there is not much difference in accuracy and AUC between static and dynamic HRI conditions. Moreover, the beamforming methods delay-and-sum and MVDR lead to average improvement in accuracy and AUC equal to 8% and 2%, respectively, when applied to training and testing data. Regarding the complementarity of time-dependent and time-independent features, the combination of both types of classifiers provides the best joint accuracy and AUC score.
Collapse
Affiliation(s)
- Eduardo Alvarado
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile; (E.A.); (N.G.); (A.L.); (R.M.); (J.W.)
| | - Nicolás Grágeda
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile; (E.A.); (N.G.); (A.L.); (R.M.); (J.W.)
| | - Alejandro Luzanto
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile; (E.A.); (N.G.); (A.L.); (R.M.); (J.W.)
| | - Rodrigo Mahu
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile; (E.A.); (N.G.); (A.L.); (R.M.); (J.W.)
| | - Jorge Wuth
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile; (E.A.); (N.G.); (A.L.); (R.M.); (J.W.)
| | - Laura Mendoza
- Hospital Clínico Universidad de Chile, Santiago 8380420, Chile;
- Clínica Alemana, Santiago 7630000, Chile
| | - Richard M. Stern
- Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213, USA;
| | - Néstor Becerra Yoma
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile; (E.A.); (N.G.); (A.L.); (R.M.); (J.W.)
| |
Collapse
|
7
|
Zayed BA, Ali AN, Elgebaly AA, Talaia NM, Hamed M, Mansour FR. Smartphone-based point-of-care testing of the SARS-CoV-2: A systematic review. SCIENTIFIC AFRICAN 2023; 21:e01757. [PMID: 37351482 PMCID: PMC10256629 DOI: 10.1016/j.sciaf.2023.e01757] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2023] [Revised: 06/03/2023] [Accepted: 06/09/2023] [Indexed: 06/24/2023] Open
Abstract
The severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) virus's worldwide pandemic has highlighted the urgent need for reliable, quick, and affordable diagnostic tests for comprehending and controlling the epidemic by tracking the world population. Given how crucial it is to monitor and manage the pandemic, researchers have recently concentrated on creating quick detection techniques. Although PCR is still the preferred clinical diagnostic test, there is a pressing need for substitutes that are sufficiently rapid and cost-effective to provide a diagnosis at the time of use. The creation of a quick and simple POC equipment is necessary for home testing. Our review's goal is to provide an overview of the many methods utilized to identify SARS-CoV 2 in various samples utilizing portable devices, as well as any potential applications for smartphones in epidemiological research and detection. The point of care (POC) employs a range of microfluidic biosensors based on smartphones, including molecular sensors, immunological biosensors, hybrid biosensors, and imaging biosensors. For example, a number of tools have been created for the diagnosis of COVID-19, based on various theories. Integrated portable devices can be created using loop-mediated isothermal amplification, which combines isothermal amplification methods with colorimetric detection. Electrochemical approaches have been regarded as a potential substitute for optical sensing techniques that utilize fluorescence for detection and as being more beneficial to the Minimizing and simplicity of the tools used for detection, together with techniques that can amplify DNA or RNA under constant temperature conditions, without the need for repeated heating and cooling cycles. Many research have used smartphones for virus detection and data visualization, making these techniques more user-friendly and broadly distributed throughout nations. Overall, our research provides a review of different novel, non-invasive, affordable, and efficient methods for identifying COVID-19 contagious infected people and halting the disease's transmission.
Collapse
Affiliation(s)
- Berlanty A Zayed
- Tanta Student Research Academy, Faculty of Medicine, Tanta University, Tanta 31111, Egypt
| | - Ahmed N Ali
- Tanta Student Research Academy, Faculty of Medicine, Tanta University, Tanta 31111, Egypt
| | - Alaa A Elgebaly
- Tanta Student Research Academy, Faculty of Medicine, Tanta University, Tanta 31111, Egypt
| | - Nourhan M Talaia
- Tanta Student Research Academy, Faculty of Medicine, Tanta University, Tanta 31111, Egypt
| | - Mahmoud Hamed
- Department of Pharmaceutical Analytical Chemistry, Faculty of Pharmacy, Tanta University, Elgeish Street, The Medical Campus of Tanta University, Tanta 31111, Egypt
| | - Fotouh R Mansour
- Department of Pharmaceutical Analytical Chemistry, Faculty of Pharmacy, Tanta University, Elgeish Street, The Medical Campus of Tanta University, Tanta 31111, Egypt
| |
Collapse
|
8
|
Bhattacharya D, Sharma NK, Dutta D, Chetupalli SR, Mote P, Ganapathy S, Chandrakiran C, Nori S, Suhail KK, Gonuguntla S, Alagesan M. Coswara: A respiratory sounds and symptoms dataset for remote screening of SARS-CoV-2 infection. Sci Data 2023; 10:397. [PMID: 37349364 PMCID: PMC10287715 DOI: 10.1038/s41597-023-02266-0] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Accepted: 05/25/2023] [Indexed: 06/24/2023] Open
Abstract
This paper presents the Coswara dataset, a dataset containing diverse set of respiratory sounds and rich meta-data, recorded between April-2020 and February-2022 from 2635 individuals (1819 SARS-CoV-2 negative, 674 positive, and 142 recovered subjects). The respiratory sounds contained nine sound categories associated with variants of breathing, cough and speech. The rich metadata contained demographic information associated with age, gender and geographic location, as well as the health information relating to the symptoms, pre-existing respiratory ailments, comorbidity and SARS-CoV-2 test status. Our study is the first of its kind to manually annotate the audio quality of the entire dataset (amounting to 65 hours) through manual listening. The paper summarizes the data collection procedure, demographic, symptoms and audio data information. A COVID-19 classifier based on bi-directional long short-term (BLSTM) architecture, is trained and evaluated on the different population sub-groups contained in the dataset to understand the bias/fairness of the model. This enabled the analysis of the impact of gender, geographic location, date of recording, and language proficiency on the COVID-19 detection performance.
Collapse
Affiliation(s)
| | - Neeraj Kumar Sharma
- Mehta Family School of Data Science and Artificial Intelligence, Indian Institute of Technology Guwahati, Guwahati, India
| | - Debottam Dutta
- Department of Electrical Engineering, Indian Institute of Science, Bangalore, India
| | | | - Pravin Mote
- Department of Electrical Engineering, Indian Institute of Science, Bangalore, India
| | - Sriram Ganapathy
- Department of Electrical Engineering, Indian Institute of Science, Bangalore, India.
| | | | - Sahiti Nori
- Ramaiah Medical College Hospital, Bangalore, India
| | - K K Suhail
- Ramaiah Medical College Hospital, Bangalore, India
| | | | - Murali Alagesan
- PSG Institute of Medical Sciences and Research, Coimbatore, India
| |
Collapse
|
9
|
Erol Doğan G, Uzbaş B. Diagnosis of COVID-19 from blood parameters using convolutional neural network. Soft comput 2023; 27:1-16. [PMID: 37362276 PMCID: PMC10225057 DOI: 10.1007/s00500-023-08508-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/10/2023] [Indexed: 06/28/2023]
Abstract
Asymptomatically presenting COVID-19 complicates the detection of infected individuals. Additionally, the virus changes too many genomic variants, which increases the virus's ability to spread. Because there isn't a specific treatment for COVID-19 in a short time, the essential goal is to reduce the virulence of the disease. Blood parameters, which contain essential clinical information about infectious diseases and are easy to access, have an important place in COVID-19 detection. The convolutional neural network (CNN) architecture, which is popular in image processing, produces highly successful results for COVID-19 detection models. When the literature is examined, it is seen that COVID-19 studies with CNN are generally done using lung images. In this study, one-dimensional (1D) blood parameters data were converted into two-dimensional (2D) image data after preprocessing, and COVID-19 detection was made with CNN. The t-distributed stochastic neighbor embedding method was applied to transfer the feature vectors to the 2D plane. All data were framed with convex hull and minimum bounding rectangle algorithms to obtain image data. The image data obtained by pixel mapping was presented to the developed 3-line CNN architecture. This study proposes an effective and successful model by providing a combination of low-cost and rapidly-accessible blood parameters and CNN architecture making image data processing highly successful for COVID-19 detection. Ultimately, COVID-19 detection was made with a success rate of 94.85%. This study has brought a new perspective to COVID-19 detection studies by obtaining 2D image data from 1D COVID-19 blood parameters and using CNN.
Collapse
Affiliation(s)
| | - Betül Uzbaş
- Computer Engineering Department, Konya Technical University, Konya, Turkey
| |
Collapse
|
10
|
Manzella F, Pagliarini G, Sciavicco G, Stan IE. The voice of COVID-19: Breath and cough recording classification with temporal decision trees and random forests. Artif Intell Med 2023; 137:102486. [PMID: 36868683 PMCID: PMC9904537 DOI: 10.1016/j.artmed.2022.102486] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2022] [Revised: 12/27/2022] [Accepted: 12/28/2022] [Indexed: 02/05/2023]
Abstract
Symbolic learning is the logic-based approach to machine learning, and its mission is to provide algorithms and methodologies to extract logical information from data and express it in an interpretable way. Interval temporal logic has been recently proposed as a suitable tool for symbolic learning, specifically via the design of an interval temporal logic decision tree extraction algorithm. In order to improve their performances, interval temporal decision trees can be embedded into interval temporal random forests, mimicking the corresponding schema at the propositional level. In this article we consider a dataset of cough and breath sample recordings of volunteer subjects, labeled with their COVID-19 status, originally collected by the University of Cambridge. By interpreting such recordings as multivariate time series, we study the problem of their automated classification using interval temporal decision trees and forests. While this problem has been approached with the same dataset as well as with other datasets, in all cases, non-symbolic learning methods (usually, deep learning-based) have been applied to solve it; in this article we apply a symbolic approach, and show that it does not only outperform the state-of-the-art obtained with the same dataset, but its results are also superior to those of most non-symbolic techniques applied on other datasets. As an added bonus, thanks to the symbolic nature of our approach, we are also able to extract explicit knowledge to help physicians characterize typical COVID-positive cough and breath.
Collapse
Affiliation(s)
- F Manzella
- Department of Mathematics and Computer Science, University of Ferrara, Italy.
| | - G Pagliarini
- Department of Mathematics and Computer Science, University of Ferrara, Italy.
| | - G Sciavicco
- Department of Mathematics and Computer Science, University of Ferrara, Italy.
| | - I E Stan
- Department of Mathematics and Computer Science, University of Ferrara, Italy.
| |
Collapse
|
11
|
Alvarado E, Grágeda N, Luzanto A, Mahu R, Wuth J, Mendoza L, Yoma NB. Dyspnea Severity Assessment Based on Vocalization Behavior with Deep Learning on the Telephone. SENSORS (BASEL, SWITZERLAND) 2023; 23:2441. [PMID: 36904646 PMCID: PMC10007248 DOI: 10.3390/s23052441] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Revised: 02/08/2023] [Accepted: 02/17/2023] [Indexed: 06/18/2023]
Abstract
In this paper, a system to assess dyspnea with the mMRC scale, on the phone, via deep learning, is proposed. The method is based on modeling the spontaneous behavior of subjects while pronouncing controlled phonetization. These vocalizations were designed, or chosen, to deal with the stationary noise suppression of cellular handsets, to provoke different rates of exhaled air, and to stimulate different levels of fluency. Time-independent and time-dependent engineered features were proposed and selected, and a k-fold scheme with double validation was adopted to select the models with the greatest potential for generalization. Moreover, score fusion methods were also investigated to optimize the complementarity of the controlled phonetizations and features that were engineered and selected. The results reported here were obtained from 104 participants, where 34 corresponded to healthy individuals and 70 were patients with respiratory conditions. The subjects' vocalizations were recorded with a telephone call (i.e., with an IVR server). The system provided an accuracy of 59% (i.e., estimating the correct mMRC), a root mean square error equal to 0.98, false positive rate of 6%, false negative rate of 11%, and an area under the ROC curve equal to 0.97. Finally, a prototype was developed and implemented, with an ASR-based automatic segmentation scheme, to estimate dyspnea on line.
Collapse
Affiliation(s)
- Eduardo Alvarado
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile
| | - Nicolás Grágeda
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile
| | - Alejandro Luzanto
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile
| | - Rodrigo Mahu
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile
| | - Jorge Wuth
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile
| | - Laura Mendoza
- Clinical Hospital, University of Chile, Santiago 8380420, Chile
| | - Néstor Becerra Yoma
- Speech Processing and Transmission Laboratory, Electrical Engineering Department, University of Chile, Santiago 8370451, Chile
| |
Collapse
|
12
|
A Performance Study of CNN Architectures for the Autonomous Detection of COVID-19 Symptoms Using Cough and Breathing. COMPUTERS 2023. [DOI: 10.3390/computers12020044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/22/2023]
Abstract
Deep learning (DL) methods have the potential to be used for detecting COVID-19 symptoms. However, the rationale for which DL method to use and which symptoms to detect has not yet been explored. In this paper, we present the first performance study which compares various convolutional neural network (CNN) architectures for the autonomous preliminary COVID-19 detection of cough and/or breathing symptoms. We compare and analyze residual networks (ResNets), visual geometry Groups (VGGs), Alex neural networks (AlexNet), densely connected networks (DenseNet), squeeze neural networks (SqueezeNet), and COVID-19 identification ResNet (CIdeR) architectures to investigate their classification performance. We uniquely train and validate both unimodal and multimodal CNN architectures using the EPFL and Cambridge datasets. Performance comparison across all modes and datasets showed that the VGG19 and DenseNet-201 achieved the highest unimodal and multimodal classification performance. VGG19 and DensNet-201 had high F1 scores (0.94 and 0.92) for unimodal cough classification on the Cambridge dataset, compared to the next highest F1 score for ResNet (0.79), with comparable F1 scores to ResNet for the larger EPFL cough dataset. They also had consistently high accuracy, recall, and precision. For multimodal detection, VGG19 and DenseNet-201 had the highest F1 scores (0.91) compared to the other CNN structures (≤0.90), with VGG19 also having the highest accuracy and recall. Our investigation provides the foundation needed to select the appropriate deep CNN method to utilize for non-contact early COVID-19 detection.
Collapse
|
13
|
Erol G, Uzbaş B, Yücelbaş C, Yücelbaş Ş. Analyzing the effect of data preprocessing techniques using machine learning algorithms on the diagnosis of COVID-19. CONCURRENCY AND COMPUTATION : PRACTICE & EXPERIENCE 2022; 34:e7393. [PMID: 36714180 PMCID: PMC9874401 DOI: 10.1002/cpe.7393] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/11/2022] [Revised: 07/18/2022] [Accepted: 09/08/2022] [Indexed: 06/18/2023]
Abstract
Real-time polymerase chain reaction (RT-PCR) known as the swab test is a diagnostic test that can diagnose COVID-19 disease through respiratory samples in the laboratory. Due to the rapid spread of the coronavirus around the world, the RT-PCR test has become insufficient to get fast results. For this reason, the need for diagnostic methods to fill this gap has arisen and machine learning studies have started in this area. On the other hand, studying medical data is a challenging area because the data it contains is inconsistent, incomplete, difficult to scale, and very large. Additionally, some poor clinical decisions, irrelevant parameters, and limited medical data adversely affect the accuracy of studies performed. Therefore, considering the availability of datasets containing COVID-19 blood parameters, which are less in number than other medical datasets today, it is aimed to improve these existing datasets. In this direction, to obtain more consistent results in COVID-19 machine learning studies, the effect of data preprocessing techniques on the classification of COVID-19 data was investigated in this study. In this study primarily, encoding categorical feature and feature scaling processes were applied to the dataset with 15 features that contain blood data of 279 patients, including gender and age information. Then, the missingness of the dataset was eliminated by using both K-nearest neighbor algorithm (KNN) and chain equations multiple value assignment (MICE) methods. Data balancing has been done with synthetic minority oversampling technique (SMOTE), which is a data balancing method. The effect of data preprocessing techniques on ensemble learning algorithms bagging, AdaBoost, random forest and on popular classifier algorithms KNN classifier, support vector machine, logistic regression, artificial neural network, and decision tree classifiers have been analyzed. The highest accuracies obtained with the bagging classifier were 83.42% and 83.74% with KNN and MICE imputations by applying SMOTE, respectively. On the other hand, the highest accuracy ratio reached with the same classifier without SMOTE was 83.91% for the KNN imputation. In conclusion, certain data preprocessing techniques are examined comparatively and the effect of these data preprocessing techniques on success is presented and the importance of the right combination of data preprocessing to achieve success has been demonstrated by experimental studies.
Collapse
Affiliation(s)
- Gizemnur Erol
- Konya Technical UniversitySoftware Engineering DepartmentKonyaTurkey
| | - Betül Uzbaş
- Konya Technical UniversityComputer Engineering DepartmentKonyaTurkey
| | - Cüneyt Yücelbaş
- Tarsus UniversityElectronics and Automation Department, Mersin‐Tarsus OIZ Vocational School of Technical SciencesMersinTurkey
| | - Şule Yücelbaş
- Tarsus UniversityComputer Engineering DepartmentMersinTurkey
| |
Collapse
|
14
|
Lanjewar MG, Shaikh AY, Parab J. Cloud-based COVID-19 disease prediction system from X-Ray images using convolutional neural network on smartphone. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:1-30. [PMID: 36467434 PMCID: PMC9684956 DOI: 10.1007/s11042-022-14232-w] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 11/01/2022] [Accepted: 11/04/2022] [Indexed: 06/17/2023]
Abstract
COVID-19 has engulfed over 200 nations through human-to-human transmission, either directly or indirectly. Reverse Transcription-polymerase Chain Reaction (RT-PCR) has been endorsed as a standard COVID-19 diagnostic procedure but has caveats such as low sensitivity, the need for a skilled workforce, and is time-consuming. Coronaviruses show significant manifestation in Chest X-Ray (CX-Ray) images and, thus, can be a viable option for an alternate COVID-19 diagnostic strategy. An automatic COVID-19 detection system can be developed to detect the disease, thus reducing strain on the healthcare system. This paper discusses a real-time Convolutional Neural Network (CNN) based system for COVID-19 illness prediction from CX-Ray images on the cloud. The implemented CNN model displays exemplary results, with training accuracy being 99.94% and validation accuracy reaching 98.81%. The confusion matrix was utilized to assess the models' outcome and achieved 99% precision, 98% recall, 99% F1 score, 100% training area under the curve (AUC) and 98.3% validation AUC. The same CX-Ray dataset was also employed to predict the COVID-19 disease with deep Convolution Neural Networks (DCNN), such as ResNet50, VGG19, InceptonV3, and Xception. The prediction outcome demonstrated that the present CNN was more capable than the DCNN models. The efficient CNN model was deployed to the Platform as a Service (PaaS) cloud.
Collapse
Affiliation(s)
- Madhusudan G. Lanjewar
- School of Physical and Applied Sciences, Goa University, Taleigao Plateau, Goa, 403206 India
| | - Arman Yusuf Shaikh
- School of Physical and Applied Sciences, Goa University, Taleigao Plateau, Goa, 403206 India
| | - Jivan Parab
- School of Physical and Applied Sciences, Goa University, Taleigao Plateau, Goa, 403206 India
| |
Collapse
|
15
|
Rapid Convolutional Neural Networks for Gram-Stained Image Classification at Inference Time on Mobile Devices: Empirical Study from Transfer Learning to Optimization. Biomedicines 2022; 10:biomedicines10112808. [PMID: 36359328 PMCID: PMC9688012 DOI: 10.3390/biomedicines10112808] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2022] [Revised: 10/23/2022] [Accepted: 10/28/2022] [Indexed: 11/06/2022] Open
Abstract
Despite the emergence of mobile health and the success of deep learning (DL), deploying production-ready DL models to resource-limited devices remains challenging. Especially, during inference time, the speed of DL models becomes relevant. We aimed to accelerate inference time for Gram-stained analysis, which is a tedious and manual task involving microorganism detection on whole slide images. Three DL models were optimized in three steps: transfer learning, pruning and quantization and then evaluated on two Android smartphones. Most convolutional layers (≥80%) had to be retrained for adaptation to the Gram-stained classification task. The combination of pruning and quantization demonstrated its utility to reduce the model size and inference time without compromising model quality. Pruning mainly contributed to model size reduction by 15×, while quantization reduced inference time by 3× and decreased model size by 4×. The combination of two reduced the baseline model by an overall factor of 46×. Optimized models were smaller than 6 MB and were able to process one image in <0.6 s on a Galaxy S10. Our findings demonstrate that methods for model compression are highly relevant for the successful deployment of DL solutions to resource-limited devices.
Collapse
|
16
|
Aleixandre JG, Elgendi M, Menon C. The Use of Audio Signals for Detecting COVID-19: A Systematic Review. SENSORS (BASEL, SWITZERLAND) 2022; 22:8114. [PMID: 36365811 PMCID: PMC9653621 DOI: 10.3390/s22218114] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/16/2022] [Revised: 10/17/2022] [Accepted: 10/20/2022] [Indexed: 06/16/2023]
Abstract
A systematic review on the topic of automatic detection of COVID-19 using audio signals was performed. A total of 48 papers were obtained after screening 659 records identified in the PubMed, IEEE Xplore, Embase, and Google Scholar databases. The reviewed studies employ a mixture of open-access and self-collected datasets. Because COVID-19 has only recently been investigated, there is a limited amount of available data. Most of the data are crowdsourced, which motivated a detailed study of the various pre-processing techniques used by the reviewed studies. Although 13 of the 48 identified papers show promising results, several have been performed with small-scale datasets (<200). Among those papers, convolutional neural networks and support vector machine algorithms were the best-performing methods. The analysis of the extracted features showed that Mel-frequency cepstral coefficients and zero-crossing rate continue to be the most popular choices. Less common alternatives, such as non-linear features, have also been proven to be effective. The reported values for sensitivity range from 65.0% to 99.8% and those for accuracy from 59.0% to 99.8%.
Collapse
Affiliation(s)
- José Gómez Aleixandre
- Biomedical and Mobile Health Technology Lab, ETH Zurich, 8008 Zurich, Switzerland
- Department of Physics, ETH Zurich, 8093 Zurich, Switzerland
| | - Mohamed Elgendi
- Biomedical and Mobile Health Technology Lab, ETH Zurich, 8008 Zurich, Switzerland
| | - Carlo Menon
- Biomedical and Mobile Health Technology Lab, ETH Zurich, 8008 Zurich, Switzerland
| |
Collapse
|
17
|
Adams J. Artificial Intelligence as an Enabler of Achieving Primary Care + Public Health = 1. Int J Public Health 2022; 67:1605257. [PMID: 36312319 PMCID: PMC9596781 DOI: 10.3389/ijph.2022.1605257] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2022] [Accepted: 09/26/2022] [Indexed: 11/16/2022] Open
|
18
|
Kasim N, Bachner-Hinenzon N, Brikman S, Cheshin O, Adler D, Dori G. A comparison of the power of breathing sounds signals acquired with a smart stethoscope from a cohort of COVID-19 patients at peak disease, and pre-discharge from the hospital. Biomed Signal Process Control 2022; 78:103920. [PMID: 35785024 PMCID: PMC9234039 DOI: 10.1016/j.bspc.2022.103920] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2021] [Revised: 06/14/2022] [Accepted: 06/18/2022] [Indexed: 11/24/2022]
Abstract
Objectives To characterize the frequencies of breathing sounds signals (BS) in COVID-19 patients at peak disease and pre-discharge from hospitalization using a Smart stethoscope. Methods Prospective cohort study conducted during the first COVID-19 wave (April-August 2020) in Israel. COVID-19 patients (n = 19) were validated by SARS-Cov-2 PCR test. The healthy control group was composed of 153 volunteers who stated that they were healthy. Power of BS was calculated in the frequency ranges of 0–20, 0–200, and 0–2000 Hz. Results The power calculated over frequency ranges 0–20, 20–200, and 200–2000 Hz contributed approximately 45%, 45%, and 10% to the total power calculated over the range 0–2000 Hz, respectively. Total power calculated from the right side of the back showed an increase of 45–80% during peak disease compared with the healthy controls (p < 0.05). The power calculated over the back, in the infrasound range, 0–20 Hz, and not in the 20–2000 Hz range, was greater for the healthy controls than for patients. Using all 3 ranges of frequencies for distinguishing peak disease from healthy controls resulted in sensitivity and specificity of 84% and 91%, respectively. Omitting the 0–20 Hz range resulted in sensitivity and specificity of 74% and 67%, respectively. Discussion The BS power acquired from COVID-19 patients at peak disease was significantly greater than that at pre-discharge from the hospital. The infrasound range had a significant contribution to the total power. Although the source of the infrasound is not presently clear, it may serve as an automated diagnostic tool when more clinical experience is gained with this method.
Collapse
Affiliation(s)
- Nour Kasim
- Department of Internal Medicine E and Corona, HaEmek Medical Center, Afula, Israel
| | | | - Shay Brikman
- Department of Internal Medicine E and Corona, HaEmek Medical Center, Afula, Israel
- Faculty of Medicine, Technion -Israel Institute of Technology, Haifa, Israel
| | - Ori Cheshin
- Department of Internal Medicine E and Corona, HaEmek Medical Center, Afula, Israel
| | | | - Guy Dori
- Department of Internal Medicine E and Corona, HaEmek Medical Center, Afula, Israel
- Faculty of Medicine, Technion -Israel Institute of Technology, Haifa, Israel
| |
Collapse
|
19
|
Aly M, Alotaibi NS. A novel deep learning model to detect COVID-19 based on wavelet features extracted from Mel-scale spectrogram of patients' cough and breathing sounds. INFORMATICS IN MEDICINE UNLOCKED 2022; 32:101049. [PMID: 35989705 PMCID: PMC9375256 DOI: 10.1016/j.imu.2022.101049] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2022] [Revised: 08/08/2022] [Accepted: 08/09/2022] [Indexed: 10/26/2022] Open
Abstract
The goal of this paper is to classify the various cough and breath sounds of COVID-19 artefacts in the signals from dynamic real-life environments. The main reason for choosing cough and breath sounds than other common symptoms to detect COVID-19 patients from the comfort of their homes, so that they do not overload the Medicare system and therefore do not unwittingly spread the disease by regularly monitoring themselves. The presented model includes two main phases. The first phase is the sound-to-image transformation, which is improved by the Mel-scale spectrogram approach. The second phase consists of extraction of features and classification using nine deep transfer models (ResNet18/34/50/100/101, GoogLeNet, SqueezeNet, MobileNetv2, and NasNetmobile). The dataset contains information data from almost 1600 people (1185 Male and 415 Female) from all over the world. Our classification model is the most accurate, its accuracy is 99.2% according to the SGDM optimizer. The accuracy is good enough that a large set of labelled cough and breath data may be used to check the possibility for generalization. The results demonstrate that ResNet18 is the best stable model for classifying cough and breath tones from a restricted dataset, with a sensitivity of 98.3% and a specificity of 97.8%. Finally, the presented model is shown to be more trustworthy and accurate than any other present model. Cough and breath study accuracy is promising enough to put extrapolation and generalization to the test.
Collapse
Affiliation(s)
- Mohammed Aly
- Department of Artificial Intelligence, Faculty of Computers and Artificial Intelligence, Egyptian Russian University, Badr City, 11829, Cairo, Egypt
| | - Nouf Saeed Alotaibi
- Department of Computer Science, College of Science, Shaqra University, Shaqra City, 11961, Saudi Arabia
| |
Collapse
|
20
|
Alkhodari M, Widatalla N, Wahbah M, Al Sakaji R, Funamoto K, Krishnan A, Kimura Y, Khandoker AH. Deep learning identifies cardiac coupling between mother and fetus during gestation. Front Cardiovasc Med 2022; 9:926965. [PMID: 35966548 PMCID: PMC9372367 DOI: 10.3389/fcvm.2022.926965] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2022] [Accepted: 06/29/2022] [Indexed: 11/18/2022] Open
Abstract
In the last two decades, stillbirth has caused around 2 million fetal deaths worldwide. Although current ultrasound tools are reliably used for the assessment of fetal growth during pregnancy, it still raises safety issues on the fetus, requires skilled providers, and has economic concerns in less developed countries. Here, we propose deep coherence, a novel artificial intelligence (AI) approach that relies on 1 min non-invasive electrocardiography (ECG) to explain the association between maternal and fetal heartbeats during pregnancy. We validated the performance of this approach using a trained deep learning tool on a total of 941 one minute maternal-fetal R-peaks segments collected from 172 pregnant women (20–40 weeks). The high accuracy achieved by the tool (90%) in identifying coupling scenarios demonstrated the potential of using AI as a monitoring tool for frequent evaluation of fetal development. The interpretability of deep learning was significant in explaining synchronization mechanisms between the maternal and fetal heartbeats. This study could potentially pave the way toward the integration of automated deep learning tools in clinical practice to provide timely and continuous fetal monitoring while reducing triage, side-effects, and costs associated with current clinical devices.
Collapse
Affiliation(s)
- Mohanad Alkhodari
- Department of Biomedical Engineering, Healthcare Engineering Innovation Center, Khalifa University, Abu Dhabi, United Arab Emirates
- *Correspondence: Mohanad Alkhodari
| | - Namareq Widatalla
- Graduate School of Biomedical Engineering, Tohoku University, Sendai, Japan
| | - Maisam Wahbah
- Department of Biomedical Engineering, Healthcare Engineering Innovation Center, Khalifa University, Abu Dhabi, United Arab Emirates
| | - Raghad Al Sakaji
- Department of Biomedical Engineering, Healthcare Engineering Innovation Center, Khalifa University, Abu Dhabi, United Arab Emirates
| | - Kiyoe Funamoto
- Department of Biomedical Engineering, Healthcare Engineering Innovation Center, Khalifa University, Abu Dhabi, United Arab Emirates
| | - Anita Krishnan
- Division of Cardiology, Children's National Hospital, Washington, DC, United States
| | - Yoshitaka Kimura
- Department of Maternal and Child Health Care Medical Science, Tohoku University Graduate School of Medicine, Sendai, Japan
| | - Ahsan H. Khandoker
- Department of Biomedical Engineering, Healthcare Engineering Innovation Center, Khalifa University, Abu Dhabi, United Arab Emirates
- Ahsan H. Khandoker
| |
Collapse
|
21
|
Development and Validation of a Multimodal-Based Prognosis and Intervention Prediction Model for COVID-19 Patients in a Multicenter Cohort. SENSORS 2022; 22:s22135007. [PMID: 35808502 PMCID: PMC9269794 DOI: 10.3390/s22135007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/01/2022] [Revised: 06/29/2022] [Accepted: 06/29/2022] [Indexed: 02/04/2023]
Abstract
The ability to accurately predict the prognosis and intervention requirements for treating highly infectious diseases, such as COVID-19, can greatly support the effective management of patients, especially in resource-limited settings. The aim of the study is to develop and validate a multimodal artificial intelligence (AI) system using clinical findings, laboratory data and AI-interpreted features of chest X-rays (CXRs), and to predict the prognosis and the required interventions for patients diagnosed with COVID-19, using multi-center data. In total, 2282 real-time reverse transcriptase polymerase chain reaction-confirmed COVID-19 patients’ initial clinical findings, laboratory data and CXRs were retrospectively collected from 13 medical centers in South Korea, between January 2020 and June 2021. The prognostic outcomes collected included intensive care unit (ICU) admission and in-hospital mortality. Intervention outcomes included the use of oxygen (O2) supplementation, mechanical ventilation and extracorporeal membrane oxygenation (ECMO). A deep learning algorithm detecting 10 common CXR abnormalities (DLAD-10) was used to infer the initial CXR taken. A random forest model with a quantile classifier was used to predict the prognostic and intervention outcomes, using multimodal data. The area under the receiver operating curve (AUROC) values for the single-modal model, using clinical findings, laboratory data and the outputs from DLAD-10, were 0.742 (95% confidence interval [CI], 0.696−0.788), 0.794 (0.745−0.843) and 0.770 (0.724−0.815), respectively. The AUROC of the combined model, using clinical findings, laboratory data and DLAD-10 outputs, was significantly higher at 0.854 (0.820−0.889) than that of all other models (p < 0.001, using DeLong’s test). In the order of importance, age, dyspnea, consolidation and fever were significant clinical variables for prediction. The most predictive DLAD-10 output was consolidation. We have shown that a multimodal AI model can improve the performance of predicting both the prognosis and intervention in COVID-19 patients, and this could assist in effective treatment and subsequent resource management. Further, image feature extraction using an established AI engine with well-defined clinical outputs, and combining them with different modes of clinical data, could be a useful way of creating an understandable multimodal prediction model.
Collapse
|