1
|
Cansiz B, Kilinc CU, Serbes G. Tunable Q-factor wavelet transform based lung signal decomposition and statistical feature extraction for effective lung disease classification. Comput Biol Med 2024; 178:108698. [PMID: 38861896 DOI: 10.1016/j.compbiomed.2024.108698] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Revised: 05/07/2024] [Accepted: 06/01/2024] [Indexed: 06/13/2024]
Abstract
The auscultation is a non-invasive and cost-effective method used for the diagnosis of lung diseases, which are one of the leading causes of death worldwide. However, the efficacy of the auscultation suffers from the limitations of the analog stethoscopes and the subjective nature of human interpretation. To overcome these limitations, the accurate diagnosis of these diseases by employing the computer based automated algorithms applied to the digitized lung sounds has been studied for the last decades. This study proposes a novel approach that uses a Tunable Q-factor Wavelet Transform (TQWT) based statistical feature extraction followed by individual and ensemble learning model training with the aim of lung disease classification. During the learning stage various machine learning algorithms are utilized as the individual learners as well as the hard and soft voting fusion approaches are employed for performance enhancement with the aid of the predictions of individual models. For an objective evaluation of the proposed approach, the study was structured into two main tasks that were investigated in detail by using several sub-tasks to comparison with state-of-the-art studies. Among the sub-tasks which investigates patient-based classification, the highest accuracy obtained for the binary classification was achieved as 97.63% (healthy vs. non-healthy), while accuracy values up to 66.32% for three-class classification (obstructive-related, restrictive-related, and healthy), and 53.42% for five-class classification (asthma, chronic obstructive pulmonary disease, interstitial lung disease, pulmonary infection, and healthy) were obtained. Regarding the other sub-task, which investigates sample-based classification, the proposed approach was superior to almost all previous findings. The proposed method underscores the potential of TQWT based signal decomposition that leverages the power of its adaptive time-frequency resolution property satisfied by Q-factor adjustability. The obtained results are very promising and the proposed approach paves the way for more accurate and automated digital auscultation techniques in clinical settings.
Collapse
Affiliation(s)
- Berke Cansiz
- Department of Biomedical Engineering, Yildiz Technical University, Esenler, Istanbul 34220, Turkey
| | - Coskuvar Utkan Kilinc
- Department of Biomedical Engineering, Yildiz Technical University, Esenler, Istanbul 34220, Turkey
| | - Gorkem Serbes
- Department of Biomedical Engineering, Yildiz Technical University, Esenler, Istanbul 34220, Turkey.
| |
Collapse
|
2
|
Sabry AH, I. Dallal Bashi O, Nik Ali N, Mahmood Al Kubaisi Y. Lung disease recognition methods using audio-based analysis with machine learning. Heliyon 2024; 10:e26218. [PMID: 38420389 PMCID: PMC10900411 DOI: 10.1016/j.heliyon.2024.e26218] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Revised: 12/11/2023] [Accepted: 02/08/2024] [Indexed: 03/02/2024] Open
Abstract
The use of computer-based automated approaches and improvements in lung sound recording techniques have made lung sound-based diagnostics even better and devoid of subjectivity errors. Using a computer to evaluate lung sound features more thoroughly with the use of analyzing changes in lung sound behavior, recording measurements, suppressing the presence of noise contaminations, and graphical representations are all made possible by computer-based lung sound analysis. This paper starts with a discussion of the need for this research area, providing an overview of the field and the motivations behind it. Following that, it details the survey methodology used in this work. It presents a discussion on the elements of sound-based lung disease classification using machine learning algorithms. This includes commonly prior considered datasets, feature extraction techniques, pre-processing methods, artifact removal methods, lung-heart sound separation, deep learning algorithms, and wavelet transform of lung audio signals. The study introduces studies that review lung screening including a summary table of these references and discusses the literature gaps in the existing studies. It is concluded that the use of sound-based machine learning in the classification of respiratory diseases has promising results. While we believe this material will prove valuable to physicians and researchers exploring sound-signal-based machine learning, large-scale investigations remain essential to solidify the findings and foster wider adoption within the medical community.
Collapse
Affiliation(s)
- Ahmad H. Sabry
- Department of Medical Instrumentation Engineering Techniques, Shatt Al-Arab University College, Basra, Iraq
| | - Omar I. Dallal Bashi
- Medical Technical Institute, Northern Technical University, 95G2+P34, Mosul, 41002, Iraq
| | - N.H. Nik Ali
- School of Electrical Engineering, College of Engineering, Universiti Teknologi MARA, 40450 Shah Alam, Selangor, Malaysia
| | - Yasir Mahmood Al Kubaisi
- Department of Sustainability Management, Dubai Academic Health Corporation, Dubai, 4545, United Arab Emirates
| |
Collapse
|
3
|
Abdul Sattar Shaikh A, Bhargavi MS, Kumar C P. Weighted aggregation through probability based ranking: An optimized federated learning architecture to classify respiratory diseases. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 242:107821. [PMID: 37776709 DOI: 10.1016/j.cmpb.2023.107821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Revised: 08/29/2023] [Accepted: 09/15/2023] [Indexed: 10/02/2023]
Abstract
Background and Objective Respiratory Diseases are one of the leading chronic illnesses in the world according to the reports by World Health Organization. Diagnosing these respiratory diseases is done through auscultation where a medical professional listens to sounds of air in the lungs for anomalies through a stethoscope. This method necessitates extensive experience and can also be misinterpreted by the medical professional. To address this issue, we introduce an AI-based solution that listens to the lung sounds and classifies the respiratory disease detected. Since the research work deals with medical data that is tightly under wraps due to privacy concerns in the medical field, we introduce a Deep learning solution to classify the diseases and a custom Federated learning (FL) approach to further improve the accuracy of the deep learning model and simultaneously maintain data privacy. Federated Learning architecture maintains data privacy and facilitates a distributed learning system for medical infrastructures. Methods The approach utilizes Generative Adversarial Networks (GAN) based Federated learning approach to ensure data privacy. Generative Adversarial Networks generate new data by synthesizing new lung sounds. This new synthesized data is then converted to spectrograms and trained on a neural network to classify four lung diseases, Heart Attack and Normal breathing patterns. Furthermore, to address performance loss during FL, we also propose a new "Weighted Aggregation through Probability-based Ranking (FedWAPR)" algorithm for optimizing the FL aggregation process. The FedWAPR aggregation takes inspiration from exponential distribution function and ranks better performing clients according to it. Results and Conclusion A test accuracy of about 92% was achieved by the trained model while classifying various respiratory diseases and heart failure. Additionally, we developed a novel FedWAPR approach that significantly outperformed the FedAVG approach for the FL aggregate function. A patient can be checked for respiratory diseases using this improved learning approach without the need for extensive sensitive data recording or for making sure the data sample obtained is secure. In a decentralized training runtime, the trained model successfully classifies various respiratory diseases and heart failure using lung sounds with a test accuracy on par with a centralized model.
Collapse
Affiliation(s)
- Abdullah Abdul Sattar Shaikh
- Department of Computer Science and Engineering, Bangalore Institute of Technology, Bangalore, 560004, Karnataka, India.
| | - M S Bhargavi
- Department of Computer Science and Engineering, Bangalore Institute of Technology, Bangalore, 560004, Karnataka, India.
| | - Pavan Kumar C
- Department of Computer Science and Engineering, Indian Institute of Information Technology Dharwad, Dharwad, 580009, Karnataka, India.
| |
Collapse
|
4
|
Pessoa D, Rocha BM, Strodthoff C, Gomes M, Rodrigues G, Petmezas G, Cheimariotis GA, Kilintzis V, Kaimakamis E, Maglaveras N, Marques A, Frerichs I, Carvalho PD, Paiva RP. BRACETS: Bimodal repository of auscultation coupled with electrical impedance thoracic signals. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 240:107720. [PMID: 37544061 DOI: 10.1016/j.cmpb.2023.107720] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Revised: 06/27/2023] [Accepted: 07/10/2023] [Indexed: 08/08/2023]
Abstract
BACKGROUND AND OBJECTIVE Respiratory diseases are among the most significant causes of morbidity and mortality worldwide, causing substantial strain on society and health systems. Over the last few decades, there has been increasing interest in the automatic analysis of respiratory sounds and electrical impedance tomography (EIT). Nevertheless, no publicly available databases with both respiratory sound and EIT data are available. METHODS In this work, we have assembled the first open-access bimodal database focusing on the differential diagnosis of respiratory diseases (BRACETS: Bimodal Repository of Auscultation Coupled with Electrical Impedance Thoracic Signals). It includes simultaneous recordings of single and multi-channel respiratory sounds and EIT. Furthermore, we have proposed several machine learning-based baseline systems for automatically classifying respiratory diseases in six distinct evaluation tasks using respiratory sound and EIT (A1, A2, A3, B1, B2, B3). These tasks included classifying respiratory diseases at sample and subject levels. The performance of the classification models was evaluated using a 5-fold cross-validation scheme (with subject isolation between folds). RESULTS The resulting database consists of 1097 respiratory sounds and 795 EIT recordings acquired from 78 adult subjects in two countries (Portugal and Greece). In the task of automatically classifying respiratory diseases, the baseline classification models have achieved the following average balanced accuracy: Task A1 - 77.9±13.1%; Task A2 - 51.6±9.7%; Task A3 - 38.6±13.1%; Task B1 - 90.0±22.4%; Task B2 - 61.4±11.8%; Task B3 - 50.8±10.6%. CONCLUSION The creation of this database and its public release will aid the research community in developing automated methodologies to assess and monitor respiratory function, and it might serve as a benchmark in the field of digital medicine for managing respiratory diseases. Moreover, it could pave the way for creating multi-modal robust approaches for that same purpose.
Collapse
Affiliation(s)
- Diogo Pessoa
- University of Coimbra Centre for Informatics and Systems of the University of Coimbra, Department of Informatics Engineering, 3030-290 Coimbra, Portugal.
| | - Bruno Machado Rocha
- University of Coimbra Centre for Informatics and Systems of the University of Coimbra, Department of Informatics Engineering, 3030-290 Coimbra, Portugal
| | - Claas Strodthoff
- Department of Anesthesiology, and Intensive Care Medicine, University Medical Center Schleswig-Holstein Campus Kiel, Kiel 24105, Schleswig-Holstein, Germany
| | - Maria Gomes
- Lab3R - Respiratory Research and Rehabilitation Laboratory, School of Health Sciences (ESSUA), University of Aveiro, 3810-193 Aveiro, Portugal
| | - Guilherme Rodrigues
- Lab3R - Respiratory Research and Rehabilitation Laboratory, School of Health Sciences (ESSUA), University of Aveiro, 3810-193 Aveiro, Portugal
| | - Georgios Petmezas
- 2nd Department of Obstetrics and Gynaecology, The Medical School, 54124 Thessaloniki, Greece
| | | | - Vassilis Kilintzis
- 2nd Department of Obstetrics and Gynaecology, The Medical School, 54124 Thessaloniki, Greece
| | - Evangelos Kaimakamis
- 1st Intensive Care Unit, "G. Papanikolaou" General Hospital of Thessaloniki, 57010 Pilea Hortiatis, Greece
| | - Nicos Maglaveras
- 2nd Department of Obstetrics and Gynaecology, The Medical School, 54124 Thessaloniki, Greece
| | - Alda Marques
- Lab3R - Respiratory Research and Rehabilitation Laboratory, School of Health Sciences (ESSUA), University of Aveiro, 3810-193 Aveiro, Portugal; Institute of Biomedicine (iBiMED), University of Aveiro, 3810-193 Aveiro, Portugal
| | - Inéz Frerichs
- Department of Anesthesiology, and Intensive Care Medicine, University Medical Center Schleswig-Holstein Campus Kiel, Kiel 24105, Schleswig-Holstein, Germany
| | - Paulo de Carvalho
- University of Coimbra Centre for Informatics and Systems of the University of Coimbra, Department of Informatics Engineering, 3030-290 Coimbra, Portugal
| | - Rui Pedro Paiva
- University of Coimbra Centre for Informatics and Systems of the University of Coimbra, Department of Informatics Engineering, 3030-290 Coimbra, Portugal
| |
Collapse
|
5
|
Prabhakar SK, Won DO. HISET: Hybrid interpretable strategies with ensemble techniques for respiratory sound classification. Heliyon 2023; 9:e18466. [PMID: 37554776 PMCID: PMC10404967 DOI: 10.1016/j.heliyon.2023.e18466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Revised: 07/18/2023] [Accepted: 07/18/2023] [Indexed: 08/10/2023] Open
Abstract
The human respiratory systems can be affected by several diseases and it is associated with distinctive sounds. For advanced biomedical signal processing, one of the most complex issues is automated respiratory sound classification. In this research, five Hybrid Interpretable Strategies with Ensemble Techniques (HISET) which are quite interesting and robust are proposed for the purpose of respiratory sounds classification. The first approach is termed as an Ensemble GSSR technique which utilizes L 2 Granger Analysis and the proposed Supportive Ensemble Empirical Mode Decomposition (SEEMD) technique and then Support Vector Machine based Recursive Feature Elimination (SVM-RFE) is used for feature selection and followed by classification with Machine Learning (ML) classifiers. The second approach proposed is the implementation of a novel Realm Revamping Sparse Representation Classification (RR-SRC) technique and third approach proposed is a Distance Metric dependent Variational Mode Decomposition (DM-VMD) with Extreme Learning Machine (ELM) classification process. The fourth approach proposed is with the usage of Harris Hawks Optimization (HHO) with a Scaling Factor based Pliable Differential Evolution (SFPDE) algorithm termed as HHO-SFPDE and it is classified with ML classifiers. The fifth or the final approach proposed analyzes the application of dimensionality reduction techniques with the proposed Gray Wolf Optimization based Support Vector Classification (GWO-SVC) and another parallel approach utilizes a similar kind of analysis with the Grasshopper Optimization Algorithm (GOA) based Sparse Autoencoder. The results are examined for ICBHI dataset and the best results are shown for the 2-class classification when the analysis is carried out with Manhattan distance-based VMD-ELM reporting an accuracy of 95.39%, and for 3-class classification Euclidean distance-based VMD-ELM reported an accuracy of 90.61% and for 4-class classification, Manhattan distance-based VMD-ELM reported an accuracy of 89.27%.
Collapse
Affiliation(s)
- Sunil Kumar Prabhakar
- Department of Artificial Intelligence Convergence, Hallym University, Chuncheon, Gangwon-do, South Korea
| | - Dong-Ok Won
- Department of Artificial Intelligence Convergence, Hallym University, Chuncheon, Gangwon-do, South Korea
| |
Collapse
|
6
|
Sfayyih AH, Sulaiman N, Sabry AH. A review on lung disease recognition by acoustic signal analysis with deep learning networks. JOURNAL OF BIG DATA 2023; 10:101. [PMID: 37333945 PMCID: PMC10259357 DOI: 10.1186/s40537-023-00762-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Accepted: 05/08/2023] [Indexed: 06/20/2023]
Abstract
Recently, assistive explanations for difficulties in the health check area have been made viable thanks in considerable portion to technologies like deep learning and machine learning. Using auditory analysis and medical imaging, they also increase the predictive accuracy for prompt and early disease detection. Medical professionals are thankful for such technological support since it helps them manage further patients because of the shortage of skilled human resources. In addition to serious illnesses like lung cancer and respiratory diseases, the plurality of breathing difficulties is gradually rising and endangering society. Because early prediction and immediate treatment are crucial for respiratory disorders, chest X-rays and respiratory sound audio are proving to be quite helpful together. Compared to related review studies on lung disease classification/detection using deep learning algorithms, only two review studies based on signal analysis for lung disease diagnosis have been conducted in 2011 and 2018. This work provides a review of lung disease recognition with acoustic signal analysis with deep learning networks. We anticipate that physicians and researchers working with sound-signal-based machine learning will find this material beneficial.
Collapse
Affiliation(s)
- Alyaa Hamel Sfayyih
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Universiti Putra Malaysia, 43400 Serdang, Malaysia
| | - Nasri Sulaiman
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Universiti Putra Malaysia, 43400 Serdang, Malaysia
| | - Ahmad H. Sabry
- Department of Computer Engineering, Al-Nahrain University, Al Jadriyah Bridge, 64074 Baghdad, Iraq
| |
Collapse
|
7
|
Sfayyih AH, Sabry AH, Jameel SM, Sulaiman N, Raafat SM, Humaidi AJ, Kubaiaisi YMA. Acoustic-Based Deep Learning Architectures for Lung Disease Diagnosis: A Comprehensive Overview. Diagnostics (Basel) 2023; 13:diagnostics13101748. [PMID: 37238233 DOI: 10.3390/diagnostics13101748] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2023] [Revised: 05/04/2023] [Accepted: 05/11/2023] [Indexed: 05/28/2023] Open
Abstract
Lung auscultation has long been used as a valuable medical tool to assess respiratory health and has gotten a lot of attention in recent years, notably following the coronavirus epidemic. Lung auscultation is used to assess a patient's respiratory role. Modern technological progress has guided the growth of computer-based respiratory speech investigation, a valuable tool for detecting lung abnormalities and diseases. Several recent studies have reviewed this important area, but none are specific to lung sound-based analysis with deep-learning architectures from one side and the provided information was not sufficient for a good understanding of these techniques. This paper gives a complete review of prior deep-learning-based architecture lung sound analysis. Deep-learning-based respiratory sound analysis articles are found in different databases including the Plos, ACM Digital Libraries, Elsevier, PubMed, MDPI, Springer, and IEEE. More than 160 publications were extracted and submitted for assessment. This paper discusses different trends in pathology/lung sound, the common features for classifying lung sounds, several considered datasets, classification methods, signal processing techniques, and some statistical information based on previous study findings. Finally, the assessment concludes with a discussion of potential future improvements and recommendations.
Collapse
Affiliation(s)
- Alyaa Hamel Sfayyih
- Department of Electrical and Electronic Engineering, Faculty of Engineering, University Putra Malaysia, Serdang 43400, Malaysia
| | - Ahmad H Sabry
- Department of Computer Engineering, Al-Nahrain University Al Jadriyah Bridge, Baghdad 64074, Iraq
| | | | - Nasri Sulaiman
- Department of Electrical and Electronic Engineering, Faculty of Engineering, University Putra Malaysia, Serdang 43400, Malaysia
| | - Safanah Mudheher Raafat
- Department of Control and Systems Engineering, University of Technology, Baghdad 10011, Iraq
| | - Amjad J Humaidi
- Department of Control and Systems Engineering, University of Technology, Baghdad 10011, Iraq
| | - Yasir Mahmood Al Kubaiaisi
- Department of Sustainability Management, Dubai Academic Health Corporation, Dubai 4545, United Arab Emirates
| |
Collapse
|
8
|
Zhang M, Li M, Guo L, Liu J. A Low-Cost AI-Empowered Stethoscope and a Lightweight Model for Detecting Cardiac and Respiratory Diseases from Lung and Heart Auscultation Sounds. SENSORS (BASEL, SWITZERLAND) 2023; 23:2591. [PMID: 36904794 PMCID: PMC10007545 DOI: 10.3390/s23052591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/26/2022] [Revised: 02/11/2023] [Accepted: 02/17/2023] [Indexed: 06/18/2023]
Abstract
Cardiac and respiratory diseases are the primary causes of health problems. If we can automate anomalous heart and lung sound diagnosis, we can improve the early detection of disease and enable the screening of a wider population than possible with manual screening. We propose a lightweight yet powerful model for simultaneous lung and heart sound diagnosis, which is deployable in an embedded low-cost device and is valuable in remote areas or developing countries where Internet access may not be available. We trained and tested the proposed model with the ICBHI and the Yaseen datasets. The experimental results showed that our 11-class prediction model could achieve 99.94% accuracy, 99.84% precision, 99.89% specificity, 99.66% sensitivity, and 99.72% F1 score. We designed a digital stethoscope (around USD 5) and connected it to a low-cost, single-board-computer Raspberry Pi Zero 2W (around USD 20), on which our pretrained model can be smoothly run. This AI-empowered digital stethoscope is beneficial for anyone in the medical field, as it can automatically provide diagnostic results and produce digital audio records for further analysis.
Collapse
Affiliation(s)
- Miao Zhang
- School of Mathematics, Shandong University, Jinan 250100, China
- School of Mathematics and Statistics, Shandong University, Weihai 264200, China
| | - Min Li
- School of Mathematics and Statistics, Shandong University, Weihai 264200, China
| | - Liang Guo
- School of Mathematics and Statistics, Shandong University, Weihai 264200, China
- Data Science Institute, Shandong University, Jinan 250100, China
| | - Jianya Liu
- School of Mathematics, Shandong University, Jinan 250100, China
- Data Science Institute, Shandong University, Jinan 250100, China
| |
Collapse
|
9
|
Cinyol F, Baysal U, Köksal D, Babaoğlu E, Ulaşlı SS. Incorporating support vector machine to the classification of respiratory sounds by Convolutional Neural Network. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
10
|
Alqudah AM, Qazan S, Obeidat YM. Deep learning models for detecting respiratory pathologies from raw lung auscultation sounds. Soft comput 2022; 26:13405-13429. [PMID: 36186666 PMCID: PMC9510581 DOI: 10.1007/s00500-022-07499-6] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/09/2022] [Indexed: 11/23/2022]
Abstract
In recent years deep learning models improve the diagnosis performance of many diseases especially respiratory diseases. This paper will propose an evaluation for the performance of different deep learning models associated with the raw lung auscultation sounds in detecting respiratory pathologies to help in providing diagnostic of respiratory pathologies in digital recorded respiratory sounds. Also, we will find out the best deep learning model for this task. In this paper, three different deep learning models have been evaluated on non-augmented and augmented datasets, where two different datasets have been utilized to generate four different sub-datasets. The results show that all the proposed deep learning methods were successful and achieved high performance in classifying the raw lung sounds, the methods were applied on different datasets and used either augmentation or non-augmentation. Among all proposed deep learning models, the CNN–LSTM model was the best model in all datasets for both augmentation and non-augmentation cases. The accuracy of CNN–LSTM model using non-augmentation was 99.6%, 99.8%, 82.4%, and 99.4% for datasets 1, 2, 3, and 4, respectively, and using augmentation was 100%, 99.8%, 98.0%, and 99.5% for datasets 1, 2, 3, and 4, respectively. While the augmentation process successfully helps the deep learning models in enhancing their performance on the testing datasets with a notable value. Moreover, the hybrid model that combines both CNN and LSTM techniques performed better than models that are based only on one of these techniques, this mainly refers to the use of CNN for automatic deep features extraction from lung sound while LSTM is used for classification.
Collapse
Affiliation(s)
- Ali Mohammad Alqudah
- Department of Biomedical Systems and Informatics Engineering, Hijjawi Faculty for Engineering Technology, Yarmouk University, Irbid, Jordan
| | - Shoroq Qazan
- Department of Computer Engineering, Hijjawi Faculty for Engineering Technology, Yarmouk University, Irbid, Jordan
| | - Yusra M Obeidat
- Department of Electronic Engineering, Hijjawi Faculty for Engineering Technology, Yarmouk University, Irbid, Jordan
| |
Collapse
|
11
|
Saldanha J, Chakraborty S, Patil S, Kotecha K, Kumar S, Nayyar A. Data augmentation using Variational Autoencoders for improvement of respiratory disease classification. PLoS One 2022; 17:e0266467. [PMID: 35960763 PMCID: PMC9374267 DOI: 10.1371/journal.pone.0266467] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/21/2021] [Accepted: 03/22/2022] [Indexed: 11/19/2022] Open
Abstract
Computerized auscultation of lung sounds is gaining importance today with the availability of lung sounds and its potential in overcoming the limitations of traditional diagnosis methods for respiratory diseases. The publicly available ICBHI respiratory sounds database is severely imbalanced, making it difficult for a deep learning model to generalize and provide reliable results. This work aims to synthesize respiratory sounds of various categories using variants of Variational Autoencoders like Multilayer Perceptron VAE (MLP-VAE), Convolutional VAE (CVAE) Conditional VAE and compare the influence of augmenting the imbalanced dataset on the performance of various lung sound classification models. We evaluated the quality of the synthetic respiratory sounds’ quality using metrics such as Fréchet Audio Distance (FAD), Cross-Correlation and Mel Cepstral Distortion. Our results showed that MLP-VAE achieved an average FAD of 12.42 over all classes, whereas Convolutional VAE and Conditional CVAE achieved an average FAD of 11.58 and 11.64 for all classes, respectively. A significant improvement in the classification performance metrics was observed upon augmenting the imbalanced dataset for certain minority classes and marginal improvement for the other classes. Hence, our work shows that deep learning-based lung sound classification models are not only a promising solution over traditional methods but can also achieve a significant performance boost upon augmenting an imbalanced training set.
Collapse
Affiliation(s)
- Jane Saldanha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Shaunak Chakraborty
- Dept. of Computer Science, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Shruti Patil
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
- * E-mail:
| | - Ketan Kotecha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Satish Kumar
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Anand Nayyar
- Graduate School (Computer Science), Faculty of Information Technology, Duy Tan University, Da Nang, Vietnam
| |
Collapse
|
12
|
Hassan Naqvi SZ, Choudhry MA. Embedded system design for classification of COPD and pneumonia patients by lung sound analysis. BIOMED ENG-BIOMED TE 2022; 67:201-218. [PMID: 35405045 DOI: 10.1515/bmt-2022-0011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2022] [Accepted: 03/17/2022] [Indexed: 11/15/2022]
Abstract
Chronic obstructive pulmonary disease (COPD) and pneumonia are lethal pulmonary illnesses with equivocal nature of abnormal pulmonic acoustics. Using lung sound signals, the classification of pulmonary abnormalities is a difficult task. A standalone system was conceived for screening COPD and Pneumonia patients through signal processing and machine learning methodologies. The proposed system will assist practitioners and pulmonologists in the accurate classification of disease. In this research work, ICBHI's and self-collected lung sound (LS) databases are used to investigate COPD and pneumonia patient. In this scheme, empirical mode decomposition (EMD), discrete wavelet transform (DWT), and analysis of variance (ANOVA) techniques are employed for segmentation, noise elimination, and feature selection, respectively. To overcome the inherent limitation of ICBHI's LS database, the adaptive synthetic (ADASYN) sampling technique is used to eradicate class imbalance. Lung sound features are used to train fine Gaussian support vector machine (FG-SVM) for classification of COPD, pneumonia, and heathy healthy subjects. This machine learning scheme is implemented on low cost and portable Raspberry pi 3 model B+ (Cortex-A53 (ARMv8) 64-bit SoC @ 1.4 GHz through hardware-supported language. Resultant hardware is capable of screening COPD and pneumonia patients accurately and assist health professionals.
Collapse
Affiliation(s)
- Syed Zohaib Hassan Naqvi
- Department of Electronics Engineering, University of Engineering and Technology Taxila, Taxila, Pakistan
| | - Mohmmad Ahmad Choudhry
- Department of Electrical Engineering, University of Engineering and Technology Taxila, Taxila, Pakistan
| |
Collapse
|
13
|
Tariq Z, Shah SK, Lee Y. Feature-Based Fusion Using CNN for Lung and Heart Sound Classification. SENSORS 2022; 22:s22041521. [PMID: 35214424 PMCID: PMC8875944 DOI: 10.3390/s22041521] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/29/2021] [Revised: 02/10/2022] [Accepted: 02/14/2022] [Indexed: 02/04/2023]
Abstract
Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of data is a considerable pitfall for improving the performance of deep learning. In this paper, we propose a novel feature-based fusion network called FDC-FS for classifying heart and lung sounds. The FDC-FS framework aims to effectively transfer learning from three different deep neural network models built from audio datasets. The innovation of the proposed transfer learning relies on the transformation from audio data to image vectors and from three specific models to one fused model that would be more suitable for deep learning. We used two publicly available datasets for this study, i.e., lung sound data from ICHBI 2017 challenge and heart challenge data. We applied data augmentation techniques, such as noise distortion, pitch shift, and time stretching, dealing with some data issues in these datasets. Importantly, we extracted three unique features from the audio samples, i.e., Spectrogram, MFCC, and Chromagram. Finally, we built a fusion of three optimal convolutional neural network models by feeding the image feature vectors transformed from audio features. We confirmed the superiority of the proposed fusion model compared to the state-of-the-art works. The highest accuracy we achieved with FDC-FS is 99.1% with Spectrogram-based lung sound classification while 97% for Spectrogram and Chromagram based heart sound classification.
Collapse
|
14
|
Alkhodari M, Khandoker AH. Detection of COVID-19 in smartphone-based breathing recordings: A pre-screening deep learning tool. PLoS One 2022; 17:e0262448. [PMID: 35025945 PMCID: PMC8758005 DOI: 10.1371/journal.pone.0262448] [Citation(s) in RCA: 21] [Impact Index Per Article: 10.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Accepted: 12/24/2021] [Indexed: 12/14/2022] Open
Abstract
This study was sought to investigate the feasibility of using smartphone-based breathing sounds within a deep learning framework to discriminate between COVID-19, including asymptomatic, and healthy subjects. A total of 480 breathing sounds (240 shallow and 240 deep) were obtained from a publicly available database named Coswara. These sounds were recorded by 120 COVID-19 and 120 healthy subjects via a smartphone microphone through a website application. A deep learning framework was proposed herein that relies on hand-crafted features extracted from the original recordings and from the mel-frequency cepstral coefficients (MFCC) as well as deep-activated features learned by a combination of convolutional neural network and bi-directional long short-term memory units (CNN-BiLSTM). The statistical analysis of patient profiles has shown a significant difference (p-value: 0.041) for ischemic heart disease between COVID-19 and healthy subjects. The Analysis of the normal distribution of the combined MFCC values showed that COVID-19 subjects tended to have a distribution that is skewed more towards the right side of the zero mean (shallow: 0.59±1.74, deep: 0.65±4.35, p-value: <0.001). In addition, the proposed deep learning approach had an overall discrimination accuracy of 94.58% and 92.08% using shallow and deep recordings, respectively. Furthermore, it detected COVID-19 subjects successfully with a maximum sensitivity of 94.21%, specificity of 94.96%, and area under the receiver operating characteristic (AUROC) curves of 0.90. Among the 120 COVID-19 participants, asymptomatic subjects (18 subjects) were successfully detected with 100.00% accuracy using shallow recordings and 88.89% using deep recordings. This study paves the way towards utilizing smartphone-based breathing sounds for the purpose of COVID-19 detection. The observations found in this study were promising to suggest deep learning and smartphone-based breathing sounds as an effective pre-screening tool for COVID-19 alongside the current reverse-transcription polymerase chain reaction (RT-PCR) assay. It can be considered as an early, rapid, easily distributed, time-efficient, and almost no-cost diagnosis technique complying with social distancing restrictions during COVID-19 pandemic.
Collapse
Affiliation(s)
- Mohanad Alkhodari
- Healthcare Engineering Innovation Center (HEIC), Department of Biomedical Engineering, Khalifa University, Abu Dhabi, UAE
| | - Ahsan H. Khandoker
- Healthcare Engineering Innovation Center (HEIC), Department of Biomedical Engineering, Khalifa University, Abu Dhabi, UAE
| |
Collapse
|
15
|
Haider NS, Behera A. Computerized lung sound based classification of asthma and chronic obstructive pulmonary disease (COPD). Biocybern Biomed Eng 2022. [DOI: 10.1016/j.bbe.2021.12.004] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2022]
|
16
|
Stokes K, Castaldo R, Franzese M, Salvatore M, Fico G, Pokvic LG, Badnjevic A, Pecchia L. A machine learning model for supporting symptom-based referral and diagnosis of bronchitis and pneumonia in limited resource settings. Biocybern Biomed Eng 2021. [DOI: 10.1016/j.bbe.2021.09.002] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
|
17
|
KAYA B, ÖNAL M. A CNN Based Method for Detecting Covid-19 from CT Images. COMPUTER SCIENCE 2021. [DOI: 10.53070/bbd.990793] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
|
18
|
Fraiwan M, Fraiwan L, Khassawneh B, Ibnian A. A dataset of lung sounds recorded from the chest wall using an electronic stethoscope. Data Brief 2021; 35:106913. [PMID: 33732827 PMCID: PMC7937981 DOI: 10.1016/j.dib.2021.106913] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2021] [Revised: 02/13/2021] [Accepted: 02/23/2021] [Indexed: 11/30/2022] Open
Abstract
The advancement of stethoscope technology has enabled high quality recording of patient sounds. We used an electronic stethoscope to record lung sounds from healthy and unhealthy subjects. The dataset includes sounds from seven ailments (i.e., asthma, heart failure, pneumonia, bronchitis, pleural effusion, lung fibrosis, and chronic obstructive pulmonary disease (COPD)) as well as normal breathing sounds. The dataset presented in this article contains the audio recordings from the examination of the chest wall at various vantage points. The stethoscope placement on the subject was determined by the specialist physician performing the diagnosis. Each recording was replicated three times corresponding to various frequency filters that emphasize certain bodily sounds. The dataset can be used for the development of automated methods that detect pulmonary diseases from lung sounds or identify the correct type of lung sound. The same methods can also be applied to the study of heart sounds.
Collapse
Affiliation(s)
- Mohammad Fraiwan
- Department of Computer Engineering, Jordan University of Science and Technology, P.O. Box 3030, Irbid 22110, Jordan
| | - Luay Fraiwan
- Department of Biomedical Engineering, Jordan University of Science and Technology, P.O. Box 3030, Irbid 22110, Jordan
| | - Basheer Khassawneh
- Department of Internal Medicine, Jordan University of Science and Technology, P.O. Box 3030, Irbid 22110, Jordan
| | - Ali Ibnian
- Department of Internal Medicine, Jordan University of Science and Technology, P.O. Box 3030, Irbid 22110, Jordan
| |
Collapse
|
19
|
Mishra NK, Singh P, Joshi SD. Automated detection of COVID-19 from CT scan using convolutional neural network. Biocybern Biomed Eng 2021; 41:572-588. [PMID: 33967366 PMCID: PMC8084624 DOI: 10.1016/j.bbe.2021.04.006] [Citation(s) in RCA: 24] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2021] [Revised: 03/09/2021] [Accepted: 04/15/2021] [Indexed: 01/01/2023]
Abstract
Under the prevailing circumstances of the global pandemic of COVID-19, early diagnosis and accurate detection of COVID-19 through tests/screening and, subsequently, isolation of the infected people would be a proactive measure. Artificial intelligence (AI) based solutions, using Convolutional Neural Network (CNN) and exploiting the Deep Learning model's diagnostic capabilities, have been studied in this paper. Transfer Learning approach, based on VGG16 and ResNet50 architectures, has been used to develop an algorithm to detect COVID-19 from CT scan images consisting of Healthy (Normal), COVID-19, and Pneumonia categories. This paper adopts data augmentation and fine-tuning techniques to improve and optimize the VGG16 and ResNet50 model. Further, stratified 5-fold cross-validation has been conducted to test the robustness and effectiveness of the model. The proposed model performs exceptionally well in case of binary classification (COVID-19 vs. Normal) with an average classification accuracy of more than 99% in both VGG16 and ResNet50 based models. In multiclass classification (COVID-19 vs. Normal vs. Pneumonia), the proposed model achieves an average classification accuracy of 86.74% and 88.52% using VGG16 and ResNet50 architectures as baseline, respectively. Experimental results show that the proposed model achieves superior performance and can be used for automated detection of COVID-19 from CT scans.
Collapse
Affiliation(s)
| | - Pushpendra Singh
- Department of ECE, National Institute of Technology Hamirpur, India
| | - Shiv Dutt Joshi
- Department of Electrical Engineering, Indian Institute of Technology Delhi, India
| |
Collapse
|