1
|
Zhao Q, Geng S, Wang B, Sun Y, Nie W, Bai B, Yu C, Zhang F, Tang G, Zhang D, Zhou Y, Liu J, Hong S. Deep Learning in Heart Sound Analysis: From Techniques to Clinical Applications. HEALTH DATA SCIENCE 2024; 4:0182. [PMID: 39387057 PMCID: PMC11461928 DOI: 10.34133/hds.0182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/12/2023] [Revised: 08/09/2024] [Accepted: 08/13/2024] [Indexed: 10/12/2024]
Abstract
Importance: Heart sound auscultation is a routinely used physical examination in clinical practice to identify potential cardiac abnormalities. However, accurate interpretation of heart sounds requires specialized training and experience, which limits its generalizability. Deep learning, a subset of machine learning, involves training artificial neural networks to learn from large datasets and perform complex tasks with intricate patterns. Over the past decade, deep learning has been successfully applied to heart sound analysis, achieving remarkable results and accumulating substantial heart sound data for model training. Although several reviews have summarized deep learning algorithms for heart sound analysis, there is a lack of comprehensive summaries regarding the available heart sound data and the clinical applications. Highlights: This review will compile the commonly used heart sound datasets, introduce the fundamentals and state-of-the-art techniques in heart sound analysis and deep learning, and summarize the current applications of deep learning for heart sound analysis, along with their limitations and areas for future improvement. Conclusions: The integration of deep learning into heart sound analysis represents a significant advancement in clinical practice. The growing availability of heart sound datasets and the continuous development of deep learning techniques contribute to the improvement and broader clinical adoption of these models. However, ongoing research is needed to address existing challenges and refine these technologies for broader clinical use.
Collapse
Affiliation(s)
- Qinghao Zhao
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | | | - Boya Wang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Gastrointestinal Oncology,
Peking University Cancer Hospital and Institute, Beijing, China
| | - Yutong Sun
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | - Wenchang Nie
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | - Baochen Bai
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | - Chao Yu
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | - Feng Zhang
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | - Gongzheng Tang
- National Institute of Health Data Science,
Peking University, Beijing, China
- Institute of Medical Technology,
Health Science Center of Peking University, Beijing, China
| | | | - Yuxi Zhou
- Department of Computer Science,
Tianjin University of Technology, Tianjin, China
- DCST, BNRist, RIIT, Institute of Internet Industry,
Tsinghua University, Beijing, China
| | - Jian Liu
- Department of Cardiology,
Peking University People’s Hospital, Beijing, China
| | - Shenda Hong
- National Institute of Health Data Science,
Peking University, Beijing, China
- Institute of Medical Technology,
Health Science Center of Peking University, Beijing, China
| |
Collapse
|
2
|
Tao R, Li H, Lu J, Huang Y, Wang Y, Lu W, Shao X, Zhou J, Yu X. DDLA: a double deep latent autoencoder for diabetic retinopathy diagnose based on continuous glucose sensors. Med Biol Eng Comput 2024; 62:3089-3106. [PMID: 38775870 DOI: 10.1007/s11517-024-03120-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 05/04/2024] [Indexed: 09/07/2024]
Abstract
The current diagnosis of diabetic retinopathy is based on fundus images and clinical experience. However, considering the ineffectiveness and non-portability of medical devices, we aimed to develop a diagnostic model for diabetic retinopathy based on glucose series data from the wearable continuous glucose monitoring system. Therefore, this study developed a novel method, i.e., double deep latent autoencoder, for exploring glycemic variability influence from multi-day glucose data for diabetic retinopathy. Specifically, the model proposed in this research could encode continuous glucose sensor data with non-continuous and variable length via the integration of a data reorganization module and a novel encoding module with fragmented-missing-wise objective function. Additionally, the model implements a double deep autoencoder, which integrated convolutional neural network, long short-term memory, to jointly capturing the inter-day and intra-day glucose latent features from glucose series. The effectiveness of the proposed model is evaluated through a cross-validation method to clinical datasets of 765 type 2 diabetes patients. The proposed method achieves the highest accuracy value (0.89), precision value (0.88), and F1 score (0.73). The results suggest that our model can be used to remotely diagnose and screen for diabetic retinopathy by learning potential features of glucose series data collected by wearable continuous glucose monitoring systems.
Collapse
Affiliation(s)
- Rui Tao
- College of Information Science and Engineering, Northeastern University, NO. 3-11 Wenhua Road, Shenyang, 110819, Liaoning, China
| | - Hongru Li
- College of Information Science and Engineering, Northeastern University, NO. 3-11 Wenhua Road, Shenyang, 110819, Liaoning, China
| | - Jingyi Lu
- Department of Endocrinology and Metabolism, Shanghai Clinical Center for Diabetes, Shanghai Diabetes Institute, Shanghai Key Laboratory of Diabetes Mellitus, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, 600 Yishan Road, Shanghai, 200233, China
| | - Youhe Huang
- College of Information Science and Engineering, Northeastern University, NO. 3-11 Wenhua Road, Shenyang, 110819, Liaoning, China
| | - Yaxin Wang
- Department of Endocrinology and Metabolism, Shanghai Clinical Center for Diabetes, Shanghai Diabetes Institute, Shanghai Key Laboratory of Diabetes Mellitus, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, 600 Yishan Road, Shanghai, 200233, China
| | - Wei Lu
- Department of Endocrinology and Metabolism, Shanghai Clinical Center for Diabetes, Shanghai Diabetes Institute, Shanghai Key Laboratory of Diabetes Mellitus, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, 600 Yishan Road, Shanghai, 200233, China
| | - Xiaopeng Shao
- College of Information Science and Engineering, Northeastern University, NO. 3-11 Wenhua Road, Shenyang, 110819, Liaoning, China
| | - Jian Zhou
- Department of Endocrinology and Metabolism, Shanghai Clinical Center for Diabetes, Shanghai Diabetes Institute, Shanghai Key Laboratory of Diabetes Mellitus, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, 600 Yishan Road, Shanghai, 200233, China.
| | - Xia Yu
- College of Information Science and Engineering, Northeastern University, NO. 3-11 Wenhua Road, Shenyang, 110819, Liaoning, China.
| |
Collapse
|
3
|
Centracchio J, Parlato S, Esposito D, Andreozzi E. Accurate Localization of First and Second Heart Sounds via Template Matching in Forcecardiography Signals. SENSORS (BASEL, SWITZERLAND) 2024; 24:1525. [PMID: 38475062 DOI: 10.3390/s24051525] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Revised: 02/21/2024] [Accepted: 02/23/2024] [Indexed: 03/14/2024]
Abstract
Cardiac auscultation is an essential part of physical examination and plays a key role in the early diagnosis of many cardiovascular diseases. The analysis of phonocardiography (PCG) recordings is generally based on the recognition of the main heart sounds, i.e., S1 and S2, which is not a trivial task. This study proposes a method for an accurate recognition and localization of heart sounds in Forcecardiography (FCG) recordings. FCG is a novel technique able to measure subsonic vibrations and sounds via small force sensors placed onto a subject's thorax, allowing continuous cardio-respiratory monitoring. In this study, a template-matching technique based on normalized cross-correlation was used to automatically recognize heart sounds in FCG signals recorded from six healthy subjects at rest. Distinct templates were manually selected from each FCG recording and used to separately localize S1 and S2 sounds, as well as S1-S2 pairs. A simultaneously recorded electrocardiography (ECG) trace was used for performance evaluation. The results show that the template matching approach proved capable of separately classifying S1 and S2 sounds in more than 96% of all heartbeats. Linear regression, correlation, and Bland-Altman analyses showed that inter-beat intervals were estimated with high accuracy. Indeed, the estimation error was confined within 10 ms, with negligible impact on heart rate estimation. Heart rate variability (HRV) indices were also computed and turned out to be almost comparable with those obtained from ECG. The preliminary yet encouraging results of this study suggest that the template matching approach based on normalized cross-correlation allows very accurate heart sounds localization and inter-beat intervals estimation.
Collapse
Affiliation(s)
- Jessica Centracchio
- Department of Electrical Engineering and Information Technologies, University of Naples Federico II, Via Claudio, 21, I-80125 Naples, Italy
| | - Salvatore Parlato
- Department of Electrical Engineering and Information Technologies, University of Naples Federico II, Via Claudio, 21, I-80125 Naples, Italy
| | - Daniele Esposito
- Department of Electrical Engineering and Information Technologies, University of Naples Federico II, Via Claudio, 21, I-80125 Naples, Italy
| | - Emilio Andreozzi
- Department of Electrical Engineering and Information Technologies, University of Naples Federico II, Via Claudio, 21, I-80125 Naples, Italy
| |
Collapse
|
4
|
Nijhawan R, Ansari SA, Kumar S, Alassery F, El-Kenawy SM. Gun identification from gunshot audios for secure public places using transformer learning. Sci Rep 2022; 12:13300. [PMID: 35918405 PMCID: PMC9345922 DOI: 10.1038/s41598-022-17497-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Accepted: 07/26/2022] [Indexed: 11/09/2022] Open
Abstract
Increased mass shootings and terrorist activities severely impact society mentally and physically. Development of real-time and cost-effective automated weapon detection systems increases a sense of safety in public. Most of the previously proposed methods were vision-based. They visually analyze the presence of a gun in a camera frame. This research focuses on gun-type (rifle, handgun, none) detection based on the audio of its shot. Mel-frequency-based audio features have been used. We compared both convolution-based and fully self-attention-based (transformers) architectures. We found transformer architecture generalizes better on audio features. Experimental results using the proposed transformer methodology on audio clips of gunshots show classification accuracy of 93.87%, with training loss and validation loss of 0.2509 and 0.1991, respectively. Based on experiments, we are convinced that our model can effectively be used as both a standalone system and in association with visual gun-detection systems for better security.
Collapse
Affiliation(s)
- Rahul Nijhawan
- School of Computer Science, University of Petroleum and Energy Studies, Dehradun, Uttarakhand, India
| | | | - Sunil Kumar
- School of Computer Science, University of Petroleum and Energy Studies, Dehradun, Uttarakhand, India.
| | - Fawaz Alassery
- Department of Computer Engineering, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif, 21944, Saudi Arabia
| | - Sayed M El-Kenawy
- Department of Communications and Electronics, Delta Higher Institute of Engineering and Technology, Mansoura, 35111, Egypt.,Faculty of Artificial Intelligence, Delta University for Science and Technology, Mansoura, 35712, Egypt
| |
Collapse
|
5
|
Ren Z, Qian K, Dong F, Dai Z, Nejdl W, Yamamoto Y, Schuller BW. Deep attention-based neural networks for explainable heart sound classification. MACHINE LEARNING WITH APPLICATIONS 2022. [DOI: 10.1016/j.mlwa.2022.100322] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023] Open
|
6
|
Emokpae LE, Emokpae RN, Bowry E, Bin Saif J, Mahmud M, Lalouani W, Younis M, Joyner RL. A wearable multi-modal acoustic system for breathing analysis. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2022; 151:1033. [PMID: 35232111 PMCID: PMC8942111 DOI: 10.1121/10.0009487] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/13/2021] [Revised: 01/16/2022] [Accepted: 01/18/2022] [Indexed: 06/14/2023]
Abstract
Chronic obstructive pulmonary disease (COPD) is the third leading cause of death worldwide with over 3 × 106 deaths in 2019. Such an alarming figure becomes frightening when combined with the number of lost lives resulting from COVID-caused respiratory failure. Because COPD exacerbations identified early can commonly be treated at home, early symptom detections may enable a major reduction of COPD patient readmission and associated healthcare costs; this is particularly important during pandemics such as COVID-19 in which healthcare facilities are overwhelmed. The standard adjuncts used to assess lung function (e.g., spirometry, plethysmography, and CT scan) are expensive, time consuming, and cannot be used in remote patient monitoring of an acute exacerbation. In this paper, a wearable multi-modal system for breathing analysis is presented, which can be used in quantifying various airflow obstructions. The wearable multi-modal electroacoustic system employs a body area sensor network with each sensor-node having a multi-modal sensing capability, such as a digital stethoscope, electrocardiogram monitor, thermometer, and goniometer. The signal-to-noise ratio (SNR) of the resulting acoustic spectrum is used as a measure of breathing intensity. The results are shown from data collected from over 35 healthy subjects and 3 COPD subjects, demonstrating a positive correlation of SNR values to the health-scale score.
Collapse
Affiliation(s)
- Lloyd E Emokpae
- LASARRUS Clinic and Research Center, Baltimore, Maryland 21220, USA
| | - Roland N Emokpae
- LASARRUS Clinic and Research Center, Baltimore, Maryland 21220, USA
| | - Ese Bowry
- LASARRUS Clinic and Research Center, Baltimore, Maryland 21220, USA
| | - Jaeed Bin Saif
- Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland 21250, USA
| | - Muntasir Mahmud
- Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland 21250, USA
| | - Wassila Lalouani
- Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland 21250, USA
| | - Mohamed Younis
- Department of Computer Science and Electrical Engineering, University of Maryland, Baltimore County, Baltimore, Maryland 21250, USA
| | - Robert L Joyner
- Richard A. Henson Research Institute, TidalHealth Peninsula Regional, Salisbury, Maryland 21801, USA
| |
Collapse
|
7
|
Dong F, Qian K, Ren Z, Baird A, Li X, Dai Z, Dong B, Metze F, Yamamoto Y, Schuller B. Machine Listening for Heart Status Monitoring: Introducing and Benchmarking HSS - the Heart Sounds Shenzhen Corpus. IEEE J Biomed Health Inform 2019; 24:2082-2092. [PMID: 31765322 DOI: 10.1109/jbhi.2019.2955281] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/28/2024]
Abstract
Auscultation of the heart is a widely studied technique, which requires precise hearing from practitioners as a means of distinguishing subtle differences in heart-beat rhythm. This technique is popular due to its non-invasive nature, and can be an early diagnosis aid for a range of cardiac conditions. Machine listening approaches can support this process, monitoring continuously and allowing for a representation of both mild and chronic heart conditions. Despite this potential, relevant databases and benchmark studies are scarce. In this paper, we introduce our publicly accessible database, the Heart Sounds Shenzhen Corpus (HSS), which was first released during the recent INTERSPEECH 2018 ComParE Heart Sound sub-challenge. Additionally, we provide a survey of machine learning work in the area of heart sound recognition, as well as a benchmark for HSS utilising standard acoustic features and machine learning models. At best our support vector machine with Log Mel features achieves 49.7% unweighted average recall on a three category task (normal, mild, moderate/severe).
Collapse
|