1
|
Majzoobi F, Khodabakhshi MB, Jamasb S, Goudarzi S. ConvLSNet: A lightweight architecture based on ConvLSTM model for the classification of pulmonary conditions using multichannel lung sound recordings. Artif Intell Med 2024; 154:102922. [PMID: 38924864 DOI: 10.1016/j.artmed.2024.102922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2023] [Revised: 06/19/2024] [Accepted: 06/21/2024] [Indexed: 06/28/2024]
Abstract
Characterization of lung sounds (LS) is indispensable for diagnosing respiratory pathology. Although conventional neural networks (NNs) have been widely employed for the automatic diagnosis of lung sounds, deep neural networks can potentially be more useful than conventional NNs by allowing accurate classification without requiring preprocessing and feature extraction. Utilizing the long short-term memory (LSTM) layers to reveal the sequence-based properties of the LS time series, a novel architecture consisting of a cascade of convolutional long short-term memory (ConvLSTM) and LSTM layers, namely ConvLSNet is developed, which permits highly accurate diagnosis of pulmonary disease states. By modeling the multichannel lung sounds through the ConvLSTM layer, the proposed ConvLSNet architecture can concurrently deal with the spatial and temporal properties of the six-channel LS recordings without heavy preprocessing or data transformation. Notably, the proposed model achieves a classification accuracy of 97.4 % based on LS data corresponding to three pulmonary conditions, namely asthma, COPD, and the healthy state. Compared with architectures consisting exclusively of CNN or LSTM layers, as well as those employing a cascade integration of 2DCNN and LSTM layers, the proposed ConvLSNet architecture exhibited the highest classification accuracy, while imposing the lowest computational cost as quantified by the number of parameters, training time, and learning rate.
Collapse
Affiliation(s)
- Faezeh Majzoobi
- Biomedical Engineering Department, Hamedan University of Technology, Hamedan, Iran
| | | | - Shahriar Jamasb
- Biomedical Engineering Department, Hamedan University of Technology, Hamedan, Iran
| | - Sobhan Goudarzi
- Physical Science Platform, Sunnybrook Research Institute, Toronto, ON M4N 3M5, Canada
| |
Collapse
|
2
|
Khan R, Khan SU, Saeed U, Koo IS. Auscultation-Based Pulmonary Disease Detection through Parallel Transformation and Deep Learning. Bioengineering (Basel) 2024; 11:586. [PMID: 38927822 PMCID: PMC11200393 DOI: 10.3390/bioengineering11060586] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Revised: 06/05/2024] [Accepted: 06/06/2024] [Indexed: 06/28/2024] Open
Abstract
Respiratory diseases are among the leading causes of death, with many individuals in a population frequently affected by various types of pulmonary disorders. Early diagnosis and patient monitoring (traditionally involving lung auscultation) are essential for the effective management of respiratory diseases. However, the interpretation of lung sounds is a subjective and labor-intensive process that demands considerable medical expertise, and there is a good chance of misclassification. To address this problem, we propose a hybrid deep learning technique that incorporates signal processing techniques. Parallel transformation is applied to adventitious respiratory sounds, transforming lung sound signals into two distinct time-frequency scalograms: the continuous wavelet transform and the mel spectrogram. Furthermore, parallel convolutional autoencoders are employed to extract features from scalograms, and the resulting latent space features are fused into a hybrid feature pool. Finally, leveraging a long short-term memory model, a feature from the latent space is used as input for classifying various types of respiratory diseases. Our work is evaluated using the ICBHI-2017 lung sound dataset. The experimental findings indicate that our proposed method achieves promising predictive performance, with average values for accuracy, sensitivity, specificity, and F1-score of 94.16%, 89.56%, 99.10%, and 89.56%, respectively, for eight-class respiratory diseases; 79.61%, 78.55%, 92.49%, and 78.67%, respectively, for four-class diseases; and 85.61%, 83.44%, 83.44%, and 84.21%, respectively, for binary-class (normal vs. abnormal) lung sounds.
Collapse
Affiliation(s)
- Rehan Khan
- Department of Electrical Electronic and Computer Engineering, University of Ulsan, Ulsan 44610, Republic of Korea; (R.K.); (S.U.K.)
| | - Shafi Ullah Khan
- Department of Electrical Electronic and Computer Engineering, University of Ulsan, Ulsan 44610, Republic of Korea; (R.K.); (S.U.K.)
| | - Umer Saeed
- Research Centre for Intelligent Healthcare, Coventry University, Coventry CV1 5FB, UK;
| | - In-Soo Koo
- Department of Electrical Electronic and Computer Engineering, University of Ulsan, Ulsan 44610, Republic of Korea; (R.K.); (S.U.K.)
| |
Collapse
|
3
|
Semmad A, Bahoura M. Comparative study of respiratory sounds classification methods based on cepstral analysis and artificial neural networks. Comput Biol Med 2024; 171:108190. [PMID: 38387384 DOI: 10.1016/j.compbiomed.2024.108190] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2023] [Revised: 01/30/2024] [Accepted: 02/18/2024] [Indexed: 02/24/2024]
Abstract
In this paper, we investigated and evaluated various machine learning-based approaches for automatically detecting wheezing sounds. We conducted a comprehensive comparison of these proposed systems, assessing their classification performance through metrics such as Sensitivity, Specificity, and Accuracy. The main approach to developing a machine learning-based system for classifying respiratory sounds involved the combination of a technique for extracting features from an unknown input sound with a classification method to determine its belonging class. The characterization techniques used in this study are based on the cepstral analysis, which was extensively employed in the automatic speech recognition field. While MFCC (Mel-Frequency Cepstral Coefficients) feature extraction methods are commonly used in respiratory sounds classification, our study introduces a novelty by employing GFCC (Gammatone-Frequency Cepstral Coefficients) and BFCC (Bark-Frequency Cepstral Coefficients) for this purpose. For the classification task, we employed two types of neural networks: the MLP (Multilayer Perceptron), a feedforward neural network, and a variant of the LSTM (Long Short-Term Memory) recurrent neural network called BiLSTM (Bidirectional LSTM). The proposed classification systems are evaluated using a database consisting of 497 wheezing segments and 915 normal respiratory segments, which are recorded from individuals diagnosticated with asthma and individuals without any respiratory issues, respectively. The highest classification performance was achieved by the BFCC-BiLSTM model, which demonstrated an exceptional accuracy rate of 99.8%.
Collapse
Affiliation(s)
- Abdelkrim Semmad
- Department of Engineering, Université du Québec à Rimouski, 300, allée des Ursulines, Rimouski, Qc, Canada, G5L 3A1.
| | - Mohammed Bahoura
- Department of Engineering, Université du Québec à Rimouski, 300, allée des Ursulines, Rimouski, Qc, Canada, G5L 3A1.
| |
Collapse
|
4
|
Sabry AH, I. Dallal Bashi O, Nik Ali N, Mahmood Al Kubaisi Y. Lung disease recognition methods using audio-based analysis with machine learning. Heliyon 2024; 10:e26218. [PMID: 38420389 PMCID: PMC10900411 DOI: 10.1016/j.heliyon.2024.e26218] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Revised: 12/11/2023] [Accepted: 02/08/2024] [Indexed: 03/02/2024] Open
Abstract
The use of computer-based automated approaches and improvements in lung sound recording techniques have made lung sound-based diagnostics even better and devoid of subjectivity errors. Using a computer to evaluate lung sound features more thoroughly with the use of analyzing changes in lung sound behavior, recording measurements, suppressing the presence of noise contaminations, and graphical representations are all made possible by computer-based lung sound analysis. This paper starts with a discussion of the need for this research area, providing an overview of the field and the motivations behind it. Following that, it details the survey methodology used in this work. It presents a discussion on the elements of sound-based lung disease classification using machine learning algorithms. This includes commonly prior considered datasets, feature extraction techniques, pre-processing methods, artifact removal methods, lung-heart sound separation, deep learning algorithms, and wavelet transform of lung audio signals. The study introduces studies that review lung screening including a summary table of these references and discusses the literature gaps in the existing studies. It is concluded that the use of sound-based machine learning in the classification of respiratory diseases has promising results. While we believe this material will prove valuable to physicians and researchers exploring sound-signal-based machine learning, large-scale investigations remain essential to solidify the findings and foster wider adoption within the medical community.
Collapse
Affiliation(s)
- Ahmad H. Sabry
- Department of Medical Instrumentation Engineering Techniques, Shatt Al-Arab University College, Basra, Iraq
| | - Omar I. Dallal Bashi
- Medical Technical Institute, Northern Technical University, 95G2+P34, Mosul, 41002, Iraq
| | - N.H. Nik Ali
- School of Electrical Engineering, College of Engineering, Universiti Teknologi MARA, 40450 Shah Alam, Selangor, Malaysia
| | - Yasir Mahmood Al Kubaisi
- Department of Sustainability Management, Dubai Academic Health Corporation, Dubai, 4545, United Arab Emirates
| |
Collapse
|
5
|
Mang LD, González Martínez FD, Martinez Muñoz D, García Galán S, Cortina R. Classification of Adventitious Sounds Combining Cochleogram and Vision Transformers. SENSORS (BASEL, SWITZERLAND) 2024; 24:682. [PMID: 38276373 PMCID: PMC10818433 DOI: 10.3390/s24020682] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/27/2023] [Revised: 01/13/2024] [Accepted: 01/19/2024] [Indexed: 01/27/2024]
Abstract
Early identification of respiratory irregularities is critical for improving lung health and reducing global mortality rates. The analysis of respiratory sounds plays a significant role in characterizing the respiratory system's condition and identifying abnormalities. The main contribution of this study is to investigate the performance when the input data, represented by cochleogram, is used to feed the Vision Transformer (ViT) architecture, since this input-classifier combination is the first time it has been applied to adventitious sound classification to our knowledge. Although ViT has shown promising results in audio classification tasks by applying self-attention to spectrogram patches, we extend this approach by applying the cochleogram, which captures specific spectro-temporal features of adventitious sounds. The proposed methodology is evaluated on the ICBHI dataset. We compare the classification performance of ViT with other state-of-the-art CNN approaches using spectrogram, Mel frequency cepstral coefficients, constant-Q transform, and cochleogram as input data. Our results confirm the superior classification performance combining cochleogram and ViT, highlighting the potential of ViT for reliable respiratory sound classification. This study contributes to the ongoing efforts in developing automatic intelligent techniques with the aim to significantly augment the speed and effectiveness of respiratory disease detection, thereby addressing a critical need in the medical field.
Collapse
Affiliation(s)
- Loredana Daria Mang
- Department of Telecommunication Engineering, University of Jaen, 23700 Linares, Spain; (F.D.G.M.); (D.M.M.); (S.G.G.)
| | | | - Damian Martinez Muñoz
- Department of Telecommunication Engineering, University of Jaen, 23700 Linares, Spain; (F.D.G.M.); (D.M.M.); (S.G.G.)
| | - Sebastián García Galán
- Department of Telecommunication Engineering, University of Jaen, 23700 Linares, Spain; (F.D.G.M.); (D.M.M.); (S.G.G.)
| | - Raquel Cortina
- Department of Computer Science, University of Oviedo, 33003 Oviedo, Spain;
| |
Collapse
|
6
|
Kim H, Koh D, Jung Y, Han H, Kim J, Joo Y. Breathing sounds analysis system for early detection of airway problems in patients with a tracheostomy tube. Sci Rep 2023; 13:21029. [PMID: 38030682 PMCID: PMC10687247 DOI: 10.1038/s41598-023-47904-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Accepted: 11/20/2023] [Indexed: 12/01/2023] Open
Abstract
To prevent immediate mortality in patients with a tracheostomy tube, it is essential to ensure timely suctioning or replacement of the tube. Breathing sounds at the entrance of tracheostomy tubes were recorded with a microphone and analyzed using a spectrogram to detect airway problems. The sounds were classified into three categories based on the waveform of the spectrogram according to the obstacle status: normal breathing sounds (NS), vibrant breathing sounds (VS) caused by movable obstacles, and sharp breathing sounds (SS) caused by fixed obstacles. A total of 3950 breathing sounds from 23 patients were analyzed. Despite neither the patients nor the medical staff recognizing any airway problems, the number and percentage of NS, VS, and SS were 1449 (36.7%), 1313 (33.2%), and 1188 (30.1%), respectively. Artificial intelligence (AI) was utilized to automatically classify breathing sounds. MobileNet and Inception_v3 exhibited the highest sensitivity and specificity scores of 0.9441 and 0.9414, respectively. When classifying into three categories, ResNet_50 showed the highest accuracy of 0.9027, and AlexNet showed the highest accuracy of 0.9660 in abnormal sounds. Classifying breathing sounds into three categories is very useful in deciding whether to suction or change the tracheostomy tubes, and AI can accomplish this with high accuracy.
Collapse
Affiliation(s)
- Hyunbum Kim
- Department of Otorhinolaryngology-Head and Neck Surgery, College of Medicine, The Catholic University of Korea, 2 Sosa-dong, Wonmi-gu, Bucheon, Kyounggi-do, 14647, Republic of Korea
| | - Daeyeon Koh
- School of Mechanical Engineering, Yonsei University, 50 Yonsei-Ro, Seodaemun-Gu, Seoul, 03722, Republic of Korea
| | - Yohan Jung
- School of Mechanical Engineering, Yonsei University, 50 Yonsei-Ro, Seodaemun-Gu, Seoul, 03722, Republic of Korea
| | - Hyunjun Han
- School of Mechanical Engineering, Yonsei University, 50 Yonsei-Ro, Seodaemun-Gu, Seoul, 03722, Republic of Korea
| | - Jongbaeg Kim
- School of Mechanical Engineering, Yonsei University, 50 Yonsei-Ro, Seodaemun-Gu, Seoul, 03722, Republic of Korea.
| | - Younghoon Joo
- Department of Otorhinolaryngology-Head and Neck Surgery, College of Medicine, The Catholic University of Korea, 2 Sosa-dong, Wonmi-gu, Bucheon, Kyounggi-do, 14647, Republic of Korea.
| |
Collapse
|
7
|
Garcia-Mendez JP, Lal A, Herasevich S, Tekin A, Pinevich Y, Lipatov K, Wang HY, Qamar S, Ayala IN, Khapov I, Gerberi DJ, Diedrich D, Pickering BW, Herasevich V. Machine Learning for Automated Classification of Abnormal Lung Sounds Obtained from Public Databases: A Systematic Review. Bioengineering (Basel) 2023; 10:1155. [PMID: 37892885 PMCID: PMC10604310 DOI: 10.3390/bioengineering10101155] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Revised: 09/15/2023] [Accepted: 09/26/2023] [Indexed: 10/29/2023] Open
Abstract
Pulmonary auscultation is essential for detecting abnormal lung sounds during physical assessments, but its reliability depends on the operator. Machine learning (ML) models offer an alternative by automatically classifying lung sounds. ML models require substantial data, and public databases aim to address this limitation. This systematic review compares characteristics, diagnostic accuracy, concerns, and data sources of existing models in the literature. Papers published from five major databases between 1990 and 2022 were assessed. Quality assessment was accomplished with a modified QUADAS-2 tool. The review encompassed 62 studies utilizing ML models and public-access databases for lung sound classification. Artificial neural networks (ANN) and support vector machines (SVM) were frequently employed in the ML classifiers. The accuracy ranged from 49.43% to 100% for discriminating abnormal sound types and 69.40% to 99.62% for disease class classification. Seventeen public databases were identified, with the ICBHI 2017 database being the most used (66%). The majority of studies exhibited a high risk of bias and concerns related to patient selection and reference standards. Summarizing, ML models can effectively classify abnormal lung sounds using publicly available data sources. Nevertheless, inconsistent reporting and methodologies pose limitations to advancing the field, and therefore, public databases should adhere to standardized recording and labeling procedures.
Collapse
Affiliation(s)
- Juan P. Garcia-Mendez
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Amos Lal
- Department of Medicine, Division of Pulmonary and Critical Care Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Svetlana Herasevich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Aysun Tekin
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Yuliya Pinevich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
- Department of Cardiac Anesthesiology and Intensive Care, Republican Clinical Medical Center, 223052 Minsk, Belarus
| | - Kirill Lipatov
- Division of Pulmonary Medicine, Mayo Clinic Health Systems, Essentia Health, Duluth, MN 55805, USA
| | - Hsin-Yi Wang
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
- Department of Anesthesiology, Taipei Veterans General Hospital, National Yang Ming Chiao Tung University, Taipei 11217, Taiwan
- Department of Biomedical Sciences and Engineering, National Central University, Taoyuan 320317, Taiwan
| | - Shahraz Qamar
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Ivan N. Ayala
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Ivan Khapov
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | | | - Daniel Diedrich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Brian W. Pickering
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| |
Collapse
|
8
|
Sethi AK, Muddaloor P, Anvekar P, Agarwal J, Mohan A, Singh M, Gopalakrishnan K, Yadav A, Adhikari A, Damani D, Kulkarni K, Aakre CA, Ryu AJ, Iyer VN, Arunachalam SP. Digital Pulmonology Practice with Phonopulmography Leveraging Artificial Intelligence: Future Perspectives Using Dual Microwave Acoustic Sensing and Imaging. SENSORS (BASEL, SWITZERLAND) 2023; 23:5514. [PMID: 37420680 DOI: 10.3390/s23125514] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/31/2023] [Revised: 06/01/2023] [Accepted: 06/05/2023] [Indexed: 07/09/2023]
Abstract
Respiratory disorders, being one of the leading causes of disability worldwide, account for constant evolution in management technologies, resulting in the incorporation of artificial intelligence (AI) in the recording and analysis of lung sounds to aid diagnosis in clinical pulmonology practice. Although lung sound auscultation is a common clinical practice, its use in diagnosis is limited due to its high variability and subjectivity. We review the origin of lung sounds, various auscultation and processing methods over the years and their clinical applications to understand the potential for a lung sound auscultation and analysis device. Respiratory sounds result from the intra-pulmonary collision of molecules contained in the air, leading to turbulent flow and subsequent sound production. These sounds have been recorded via an electronic stethoscope and analyzed using back-propagation neural networks, wavelet transform models, Gaussian mixture models and recently with machine learning and deep learning models with possible use in asthma, COVID-19, asbestosis and interstitial lung disease. The purpose of this review was to summarize lung sound physiology, recording technologies and diagnostics methods using AI for digital pulmonology practice. Future research and development in recording and analyzing respiratory sounds in real time could revolutionize clinical practice for both the patients and the healthcare personnel.
Collapse
Affiliation(s)
- Arshia K Sethi
- GIH Artificial Intelligence Laboratory (GAIL), Division of Gastroenterology and Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Pratyusha Muddaloor
- GIH Artificial Intelligence Laboratory (GAIL), Division of Gastroenterology and Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | | | - Joshika Agarwal
- Division of Pulmonary and Critical Care Medicine, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Anmol Mohan
- GIH Artificial Intelligence Laboratory (GAIL), Division of Gastroenterology and Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | | | - Keerthy Gopalakrishnan
- GIH Artificial Intelligence Laboratory (GAIL), Division of Gastroenterology and Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
- Microwave Engineering and Imaging Laboratory (MEIL), Division of Gastroenterology & Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Ashima Yadav
- Department of Cardiovascular Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Aakriti Adhikari
- GIH Artificial Intelligence Laboratory (GAIL), Division of Gastroenterology and Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Devanshi Damani
- Department of Cardiovascular Medicine, Mayo Clinic, Rochester, MN 55905, USA
- Department of Internal Medicine, Texas Tech University Health Science Center, El Paso, TX 79995, USA
| | - Kanchan Kulkarni
- INSERM, Centre de Recherche Cardio-Thoracique de Bordeaux, University of Bordeaux, U1045, F-33000 Bordeaux, France
- IHU Liryc, Heart Rhythm Disease Institute, Fondation Bordeaux Université, F-33600 Pessac, France
| | | | - Alexander J Ryu
- Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Vivek N Iyer
- Division of Pulmonary and Critical Care Medicine, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Shivaram P Arunachalam
- GIH Artificial Intelligence Laboratory (GAIL), Division of Gastroenterology and Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
- Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
- Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
- Microwave Engineering and Imaging Laboratory (MEIL), Division of Gastroenterology & Hepatology, Department of Medicine, Mayo Clinic, Rochester, MN 55905, USA
| |
Collapse
|
9
|
Sfayyih AH, Sulaiman N, Sabry AH. A review on lung disease recognition by acoustic signal analysis with deep learning networks. JOURNAL OF BIG DATA 2023; 10:101. [PMID: 37333945 PMCID: PMC10259357 DOI: 10.1186/s40537-023-00762-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Accepted: 05/08/2023] [Indexed: 06/20/2023]
Abstract
Recently, assistive explanations for difficulties in the health check area have been made viable thanks in considerable portion to technologies like deep learning and machine learning. Using auditory analysis and medical imaging, they also increase the predictive accuracy for prompt and early disease detection. Medical professionals are thankful for such technological support since it helps them manage further patients because of the shortage of skilled human resources. In addition to serious illnesses like lung cancer and respiratory diseases, the plurality of breathing difficulties is gradually rising and endangering society. Because early prediction and immediate treatment are crucial for respiratory disorders, chest X-rays and respiratory sound audio are proving to be quite helpful together. Compared to related review studies on lung disease classification/detection using deep learning algorithms, only two review studies based on signal analysis for lung disease diagnosis have been conducted in 2011 and 2018. This work provides a review of lung disease recognition with acoustic signal analysis with deep learning networks. We anticipate that physicians and researchers working with sound-signal-based machine learning will find this material beneficial.
Collapse
Affiliation(s)
- Alyaa Hamel Sfayyih
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Universiti Putra Malaysia, 43400 Serdang, Malaysia
| | - Nasri Sulaiman
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Universiti Putra Malaysia, 43400 Serdang, Malaysia
| | - Ahmad H. Sabry
- Department of Computer Engineering, Al-Nahrain University, Al Jadriyah Bridge, 64074 Baghdad, Iraq
| |
Collapse
|
10
|
Sfayyih AH, Sabry AH, Jameel SM, Sulaiman N, Raafat SM, Humaidi AJ, Kubaiaisi YMA. Acoustic-Based Deep Learning Architectures for Lung Disease Diagnosis: A Comprehensive Overview. Diagnostics (Basel) 2023; 13:diagnostics13101748. [PMID: 37238233 DOI: 10.3390/diagnostics13101748] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2023] [Revised: 05/04/2023] [Accepted: 05/11/2023] [Indexed: 05/28/2023] Open
Abstract
Lung auscultation has long been used as a valuable medical tool to assess respiratory health and has gotten a lot of attention in recent years, notably following the coronavirus epidemic. Lung auscultation is used to assess a patient's respiratory role. Modern technological progress has guided the growth of computer-based respiratory speech investigation, a valuable tool for detecting lung abnormalities and diseases. Several recent studies have reviewed this important area, but none are specific to lung sound-based analysis with deep-learning architectures from one side and the provided information was not sufficient for a good understanding of these techniques. This paper gives a complete review of prior deep-learning-based architecture lung sound analysis. Deep-learning-based respiratory sound analysis articles are found in different databases including the Plos, ACM Digital Libraries, Elsevier, PubMed, MDPI, Springer, and IEEE. More than 160 publications were extracted and submitted for assessment. This paper discusses different trends in pathology/lung sound, the common features for classifying lung sounds, several considered datasets, classification methods, signal processing techniques, and some statistical information based on previous study findings. Finally, the assessment concludes with a discussion of potential future improvements and recommendations.
Collapse
Affiliation(s)
- Alyaa Hamel Sfayyih
- Department of Electrical and Electronic Engineering, Faculty of Engineering, University Putra Malaysia, Serdang 43400, Malaysia
| | - Ahmad H Sabry
- Department of Computer Engineering, Al-Nahrain University Al Jadriyah Bridge, Baghdad 64074, Iraq
| | | | - Nasri Sulaiman
- Department of Electrical and Electronic Engineering, Faculty of Engineering, University Putra Malaysia, Serdang 43400, Malaysia
| | - Safanah Mudheher Raafat
- Department of Control and Systems Engineering, University of Technology, Baghdad 10011, Iraq
| | - Amjad J Humaidi
- Department of Control and Systems Engineering, University of Technology, Baghdad 10011, Iraq
| | - Yasir Mahmood Al Kubaiaisi
- Department of Sustainability Management, Dubai Academic Health Corporation, Dubai 4545, United Arab Emirates
| |
Collapse
|
11
|
Mang L, Canadas-Quesada F, Carabias-Orti J, Combarro E, Ranilla J. Cochleogram-based adventitious sounds classification using convolutional neural networks. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104555] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023]
|
12
|
Monitoring of Sleep Breathing States Based on Audio Sensor Utilizing Mel-Scale Features in Home Healthcare. JOURNAL OF HEALTHCARE ENGINEERING 2023; 2023:6197564. [PMID: 36818388 PMCID: PMC9935909 DOI: 10.1155/2023/6197564] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/04/2022] [Revised: 07/03/2022] [Accepted: 11/24/2022] [Indexed: 02/11/2023]
Abstract
Sleep-related breathing disorders (SBDs) will lead to poor sleep quality and increase the risk of cardiovascular and cerebrovascular diseases which may cause death in serious cases. This paper aims to detect breathing states related to SBDs by breathing sound signals. A moment waveform analysis is applied to locate and segment the breathing cycles. As the core of our study, a set of useful features of breathing signal is proposed based on Mel frequency cepstrum analysis. Finally, the normal and abnormal sleep breathing states can be distinguished by the extracted Mel-scale indexes. Young healthy testers and patients who suffered from obstructive sleep apnea are tested utilizing the proposed method. The average accuracy for detecting abnormal breathing states can reach 93.1%. It will be helpful to prevent SBDs and improve the sleep quality of home healthcare.
Collapse
|
13
|
Cinyol F, Baysal U, Köksal D, Babaoğlu E, Ulaşlı SS. Incorporating support vector machine to the classification of respiratory sounds by Convolutional Neural Network. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
14
|
Singh RB, Zhuang H. Measurements, Analysis, Classification, and Detection of Gunshot and Gunshot-like Sounds. SENSORS (BASEL, SWITZERLAND) 2022; 22:9170. [PMID: 36501869 PMCID: PMC9737970 DOI: 10.3390/s22239170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/03/2022] [Revised: 11/17/2022] [Accepted: 11/22/2022] [Indexed: 06/17/2023]
Abstract
Gun violence has been on the rise in recent years. To help curb the downward spiral of this negative influence in communities, machine learning strategies on gunshot detection can be developed and deployed. After outlining the procedure by which a typical type of gunshot-like sounds were measured, this paper focuses on the analysis of feature importance pertaining to gunshot and gunshot-like sounds. The random forest mean decrease in impurity and the SHapley Additive exPlanations feature importance analysis were employed for this task. From the feature importance analysis, feature reduction was then carried out. Via the Mel-frequency cepstral coefficients feature extraction process on 1-sec audio clips, these extracted features were then reduced to a more manageable quantity using the above-mentioned feature reduction processes. These reduced features were sent to a random forest classifier. The SHapley Additive exPlanations feature importance output was compared to that of the mean decrease in impurity feature importance. The results show what Mel-frequency cepstral coefficients features are important in discriminating gunshot sounds and various gunshot-like sounds. Together with the feature importance/reduction processes, the recent uniform manifold approximation and projection method was used to compare the closeness of various gunshot-like sounds to gunshot sounds in the feature space. Finally, the approach presented in this paper provides people with a viable means to make gunshot sounds more discernible from other sounds.
Collapse
|
15
|
Ghulam Nabi F, Sundaraj K, Shahid Iqbal M, Shafiq M, Planiappan R. A telemedicine software application for asthma severity levels identification using wheeze sounds classification. Biocybern Biomed Eng 2022. [DOI: 10.1016/j.bbe.2022.11.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
|
16
|
Alqudah AM, Qazan S, Obeidat YM. Deep learning models for detecting respiratory pathologies from raw lung auscultation sounds. Soft comput 2022; 26:13405-13429. [PMID: 36186666 PMCID: PMC9510581 DOI: 10.1007/s00500-022-07499-6] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/09/2022] [Indexed: 11/23/2022]
Abstract
In recent years deep learning models improve the diagnosis performance of many diseases especially respiratory diseases. This paper will propose an evaluation for the performance of different deep learning models associated with the raw lung auscultation sounds in detecting respiratory pathologies to help in providing diagnostic of respiratory pathologies in digital recorded respiratory sounds. Also, we will find out the best deep learning model for this task. In this paper, three different deep learning models have been evaluated on non-augmented and augmented datasets, where two different datasets have been utilized to generate four different sub-datasets. The results show that all the proposed deep learning methods were successful and achieved high performance in classifying the raw lung sounds, the methods were applied on different datasets and used either augmentation or non-augmentation. Among all proposed deep learning models, the CNN–LSTM model was the best model in all datasets for both augmentation and non-augmentation cases. The accuracy of CNN–LSTM model using non-augmentation was 99.6%, 99.8%, 82.4%, and 99.4% for datasets 1, 2, 3, and 4, respectively, and using augmentation was 100%, 99.8%, 98.0%, and 99.5% for datasets 1, 2, 3, and 4, respectively. While the augmentation process successfully helps the deep learning models in enhancing their performance on the testing datasets with a notable value. Moreover, the hybrid model that combines both CNN and LSTM techniques performed better than models that are based only on one of these techniques, this mainly refers to the use of CNN for automatic deep features extraction from lung sound while LSTM is used for classification.
Collapse
Affiliation(s)
- Ali Mohammad Alqudah
- Department of Biomedical Systems and Informatics Engineering, Hijjawi Faculty for Engineering Technology, Yarmouk University, Irbid, Jordan
| | - Shoroq Qazan
- Department of Computer Engineering, Hijjawi Faculty for Engineering Technology, Yarmouk University, Irbid, Jordan
| | - Yusra M Obeidat
- Department of Electronic Engineering, Hijjawi Faculty for Engineering Technology, Yarmouk University, Irbid, Jordan
| |
Collapse
|
17
|
Neili Z, Sundaraj K. A comparative study of the spectrogram, scalogram, melspectrogram and gammatonegram time-frequency representations for the classification of lung sounds using the ICBHI database based on CNNs. BIOMED ENG-BIOMED TE 2022; 67:367-390. [PMID: 35926850 DOI: 10.1515/bmt-2022-0180] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2022] [Accepted: 06/21/2022] [Indexed: 11/15/2022]
Abstract
In lung sound classification using deep learning, many studies have considered the use of short-time Fourier transform (STFT) as the most commonly used 2D representation of the input data. Consequently, STFT has been widely used as an analytical tool, but other versions of the representation have also been developed. This study aims to evaluate and compare the performance of the spectrogram, scalogram, melspectrogram and gammatonegram representations, and provide comparative information to users regarding the suitability of these time-frequency (TF) techniques in lung sound classification. Lung sound signals used in this study were obtained from the ICBHI 2017 respiratory sound database. These lung sound recordings were converted into images of spectrogram, scalogram, melspectrogram and gammatonegram TF representations respectively. The four types of images were fed separately into the VGG16, ResNet-50 and AlexNet deep-learning architectures. Network performances were analyzed and compared based on accuracy, precision, recall and F1-score. The results of the analysis on the performance of the four representations using these three commonly used CNN deep-learning networks indicate that the generated gammatonegram and scalogram TF images coupled with ResNet-50 achieved maximum classification accuracies.
Collapse
Affiliation(s)
- Zakaria Neili
- Electronics Department, University of Badji Mokhtar Annaba, Annaba, Algeria
| | - Kenneth Sundaraj
- Faculty of Electronics and Computer Engineering, Universiti Teknikal Malaysia Melaka, Melaka, Malaysia
| |
Collapse
|
18
|
Purnomo AT, Komariah KS, Lin DB, Hendria WF, Sin BK, Ahmadi N. Non-Contact Supervision of COVID-19 Breathing Behaviour With FMCW Radar and Stacked Ensemble Learning Model in Real-Time. IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS 2022; 16:664-678. [PMID: 35853073 PMCID: PMC9647724 DOI: 10.1109/tbcas.2022.3192359] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 03/30/2022] [Accepted: 06/24/2022] [Indexed: 06/15/2023]
Abstract
A respiratory disorder that attacks COVID-19 patients requires intensive supervision of medical practitioners during the isolation period. A non-contact monitoring device will be a suitable solution for reducing the spread risk of the virus while monitoring the COVID-19 patient. This study uses Frequency-Modulated Continuous Wave (FMCW) radar and Machine Learning (ML) to obtain respiratory information and analyze respiratory signals, respectively. Multiple subjects in a room can be detected simultaneously by calculating the Angle of Arrival (AoA) of the received signal and utilizing the Multiple Input Multiple Output (MIMO) of FMCW radar. Fast Fourier Transform (FFT) and some signal processing are implemented to obtain a breathing waveform. ML helps the system to analyze the respiratory signals automatically. This paper also compares the performance of several ML algorithms such as Multinomial Logistic Regression (MLR), Decision Tree (DT), Random Forest (RF), Support Vector Machine (SVM), eXtreme Gradient Boosting (XGB), Light Gradient Boosting Machine (LGBM), CatBoosting (CB) Classifier, Multilayer Perceptron (MLP), and three proposed stacked ensemble models, namely Stacked Ensemble Classifier (SEC), Boosting Tree-based Stacked Classifier (BTSC), and Neural Stacked Ensemble Model (NSEM) to obtain the best ML model. The results show that the NSEM algorithm achieves the best performance with 97.1% accuracy. In the real-time implementation, the system could simultaneously detect several objects with different breathing characteristics and classify the respiratory signals into five different classes.
Collapse
Affiliation(s)
- Ariana Tulus Purnomo
- Department of Electronic and Computer EngineeringNational Taiwan University of Science and TechnologyTaipei10607Taiwan
| | - Kokoy Siti Komariah
- Department of AI Convergence and the Division of Computer Engineering (respectively)Pukyong National UniversityBusan48513Republic of Korea
| | - Ding-Bing Lin
- Department of Electronic and Computer EngineeringNational Taiwan University of Science and TechnologyTaipei10607Taiwan
| | - Willy Fitra Hendria
- Department of Intelligent Mechatronics EngineeringSejong UniversitySeoul05006Republic of Korea
| | - Bong-Kee Sin
- Department of AI Convergence and the Division of Computer Engineering (respectively)Pukyong National UniversityBusan48513Republic of Korea
| | - Nur Ahmadi
- Center for Artificial Intelligence (U-CoE AI-VLB), School of Electrical Engineering and InformaticsBandung Institute of TechnologyBandung40132Indonesia
| |
Collapse
|
19
|
Extraction of low-dimensional features for single-channel common lung sound classification. Med Biol Eng Comput 2022; 60:1555-1568. [DOI: 10.1007/s11517-022-02552-w] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2021] [Accepted: 03/09/2022] [Indexed: 11/27/2022]
|
20
|
A temporal dependency feature in lower dimension for lung sound signal classification. Sci Rep 2022; 12:7889. [PMID: 35551232 PMCID: PMC9098886 DOI: 10.1038/s41598-022-11726-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2022] [Accepted: 04/04/2022] [Indexed: 11/20/2022] Open
Abstract
Respiratory sounds are expressed as nonlinear and nonstationary signals, whose unpredictability makes it difficult to extract significant features for classification. Static cepstral coefficients such as Mel-frequency cepstral coefficients (MFCCs), have been used for classification of lung sound signals. However, they are modeled in high-dimensional hyperspectral space, and also lose temporal dependency information. Therefore, we propose shifted \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{mathrsfs}
\usepackage{upgreek}
\setlength{\oddsidemargin}{-69pt}
\begin{document}$$\delta $$\end{document}δ-cepstral coefficients in lower-subspace (SDC-L) as a novel feature for lung sound classification. It preserves temporal dependency information of multiple frames nearby same to original SDC, and improves feature extraction by reducing the hyperspectral dimension. We modified EMD algorithm by adding a stopping rule to objectively select a finite number of intrinsic mode functions (IMFs). The performances of SDC-L were evaluated with three machine learning techniques (support vector machine (SVM), k-nearest neighbor (k-NN) and random forest (RF)) and two deep learning algorithms (multilayer perceptron (MLP) and convolutional neural network (cNN)) and one hybrid deep learning algorithm combining cNN with long short term memory (LSTM) in terms of accuracy, precision, recall and F1-score. We found that the first 2 IMFs were enough to construct our feature. SVM, MLP and a hybrid deep learning algorithm (cNN plus LSTM) outperformed with SDC-L, and the other classifiers achieved equivalent results with all features. Our findings show that SDC-L is a promising feature for the classification of lung sound signals.
Collapse
|
21
|
Nguyen T, Pernkopf F. Lung Sound Classification Using Co-tuning and Stochastic Normalization. IEEE Trans Biomed Eng 2022; 69:2872-2882. [PMID: 35254969 DOI: 10.1109/tbme.2022.3156293] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Computational methods for lung sound analysis are beneficial for computer-aided diagnosis support, storage and monitoring in critical care. In this paper, we use pre-trained ResNet models as backbone architectures for classification of adventitious lung sounds and respiratory diseases. The learned representation of the pre-trained model is transferred by using vanilla fine-tuning, co-tuning, stochastic normalization and the combination of the co-tuning and stochastic normalization techniques. Furthermore, data augmentation in both time domain and time-frequency domain is used to account for the class imbalance of the ICBHI and our multi-channel lung sound dataset. Additionally, we introduce spectrum correction to account for the variations of the recording device properties on the ICBHI dataset. Empirically, our proposed systems mostly outperform all state-of-the-art lung sound classification systems for the adventitious lung sounds and respiratory diseases of both datasets.
Collapse
|
22
|
Fraiwan M, Fraiwan L, Alkhodari M, Hassanin O. Recognition of pulmonary diseases from lung sounds using convolutional neural networks and long short-term memory. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING 2022; 13:4759-4771. [PMID: 33841584 PMCID: PMC8019351 DOI: 10.1007/s12652-021-03184-y] [Citation(s) in RCA: 28] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/12/2020] [Accepted: 03/25/2021] [Indexed: 05/03/2023]
Abstract
UNLABELLED In this paper, a study is conducted to explore the ability of deep learning in recognizing pulmonary diseases from electronically recorded lung sounds. The selected data-set included a total of 103 patients obtained from locally recorded stethoscope lung sounds acquired at King Abdullah University Hospital, Jordan University of Science and Technology, Jordan. In addition, 110 patients data were added to the data-set from the Int. Conf. on Biomedical Health Informatics publicly available challenge database. Initially, all signals were checked to have a sampling frequency of 4 kHz and segmented into 5 s segments. Then, several preprocessing steps were undertaken to ensure smoother and less noisy signals. These steps included wavelet smoothing, displacement artifact removal, and z-score normalization. The deep learning network architecture consisted of two stages; convolutional neural networks and bidirectional long short-term memory units. The training of the model was evaluated based on a k-fold cross-validation scheme of tenfolds using several performance evaluation metrics including Cohen's kappa, accuracy, sensitivity, specificity, precision, and F1-score. The developed algorithm achieved the highest average accuracy of 99.62% with a precision of 98.85% in classifying patients based on the pulmonary disease types using CNN + BDLSTM. Furthermore, a total agreement of 98.26% was obtained between the predictions and original classes within the training scheme. This study paves the way towards implementing deep learning models in clinical settings to assist clinicians in decision making related to the recognition of pulmonary diseases. SUPPLEMENTARY INFORMATION The online version supplementary material available at 10.1007/s12652-021-03184-y.
Collapse
Affiliation(s)
- M. Fraiwan
- Department of Computer Engineering, Jordan University of Science and Technology, P.O. Box 3030, Irbid, 22110 Jordan
| | - L. Fraiwan
- Department of Biomedical Engineering, Jordan University of Science and Technology, P.O. Box 3030, Irbid, 22110 Jordan
| | - M. Alkhodari
- Department of Electrical and Computer Engineering, Abu Dhabi University, Abu Dhabi, UAE
| | - O. Hassanin
- Department of Electrical and Computer Engineering, Abu Dhabi University, Abu Dhabi, UAE
| |
Collapse
|
23
|
Zulfiqar R, Majeed F, Irfan R, Rauf HT, Benkhelifa E, Belkacem AN. Abnormal Respiratory Sounds Classification Using Deep CNN Through Artificial Noise Addition. Front Med (Lausanne) 2021; 8:714811. [PMID: 34869413 PMCID: PMC8635523 DOI: 10.3389/fmed.2021.714811] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2021] [Accepted: 10/07/2021] [Indexed: 11/29/2022] Open
Abstract
Respiratory sound (RS) attributes and their analyses structure a fundamental piece of pneumonic pathology, and it gives symptomatic data regarding a patient's lung. A couple of decades back, doctors depended on their hearing to distinguish symptomatic signs in lung audios by utilizing the typical stethoscope, which is usually considered a cheap and secure method for examining the patients. Lung disease is the third most ordinary cause of death worldwide, so; it is essential to classify the RS abnormality accurately to overcome the death rate. In this research, we have applied Fourier analysis for the visual inspection of abnormal respiratory sounds. Spectrum analysis was done through Artificial Noise Addition (ANA) in conjunction with different deep convolutional neural networks (CNN) to classify the seven abnormal respiratory sounds—both continuous (CAS) and discontinuous (DAS). The proposed framework contains an adaptive mechanism of adding a similar type of noise to unhealthy respiratory sounds. ANA makes sound features enough reach to be identified more accurately than the respiratory sounds without ANA. The obtained results using the proposed framework are superior to previous techniques since we simultaneously considered the seven different abnormal respiratory sound classes.
Collapse
Affiliation(s)
- Rizwana Zulfiqar
- Faculty of Computing and Information Technology, University of Gujrat, Gujrat, Pakistan
| | - Fiaz Majeed
- Faculty of Computing and Information Technology, University of Gujrat, Gujrat, Pakistan
| | - Rizwana Irfan
- Department of Information Technology, College of Computing and Information Technology at Khulais, University of Jeddah, Jeddah, Saudi Arabia
| | - Hafiz Tayyab Rauf
- Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, Stoke-on-Trent, United Kingdom
| | - Elhadj Benkhelifa
- Cloud Computing and Applications Reseach Lab, Staffordshire University, Stoke-on-Trent, United Kingdom
| | - Abdelkader Nasreddine Belkacem
- Department of Computer and Network Engineering, College of Information Technology, UAE University, Al Ain, United Arab Emirates
| |
Collapse
|
24
|
Baghel N, Nangia V, Dutta MK. ALSD-Net: Automatic lung sounds diagnosis network from pulmonary signals. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06302-1] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
25
|
Gammatonegram based triple classification of lung sounds using deep convolutional neural network with transfer learning. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.102947] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
26
|
Shuvo SB, Ali SN, Swapnil SI, Hasan T, Bhuiyan MIH. A Lightweight CNN Model for Detecting Respiratory Diseases From Lung Auscultation Sounds Using EMD-CWT-Based Hybrid Scalogram. IEEE J Biomed Health Inform 2021; 25:2595-2603. [PMID: 33373309 DOI: 10.1109/jbhi.2020.3048006] [Citation(s) in RCA: 33] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Listening to lung sounds through auscultation is vital in examining the respiratory system for abnormalities. Automated analysis of lung auscultation sounds can be beneficial to the health systems in low-resource settings where there is a lack of skilled physicians. In this work, we propose a lightweight convolutional neural network (CNN) architecture to classify respiratory diseases from individual breath cycles using hybrid scalogram-based features of lung sounds. The proposed feature-set utilizes the empirical mode decomposition (EMD) and the continuous wavelet transform (CWT). The performance of the proposed scheme is studied using a patient independent train-validation-test set from the publicly available ICBHI 2017 lung sound dataset. Employing the proposed framework, weighted accuracy scores of 98.92% for three-class chronic classification and 98.70% for six-class pathological classification are achieved, which outperform well-known and much larger VGG16 in terms of accuracy by absolute margins of 1.10% and 1.11%, respectively. The proposed CNN model also outperforms other contemporary lightweight models while being computationally comparable.
Collapse
|
27
|
GTCC-based BiLSTM deep-learning framework for respiratory sound classification using empirical mode decomposition. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-06295-x] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
28
|
De La Torre Cruz J, Cañadas Quesada FJ, Ruiz Reyes N, García Galán S, Carabias Orti JJ, Peréz Chica G. Monophonic and Polyphonic Wheezing Classification Based on Constrained Low-Rank Non-Negative Matrix Factorization. SENSORS (BASEL, SWITZERLAND) 2021; 21:1661. [PMID: 33670892 PMCID: PMC7957792 DOI: 10.3390/s21051661] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/25/2021] [Revised: 02/17/2021] [Accepted: 02/22/2021] [Indexed: 11/21/2022]
Abstract
The appearance of wheezing sounds is widely considered by physicians as a key indicator to detect early pulmonary disorders or even the severity associated with respiratory diseases, as occurs in the case of asthma and chronic obstructive pulmonary disease. From a physician's point of view, monophonic and polyphonic wheezing classification is still a challenging topic in biomedical signal processing since both types of wheezes are sinusoidal in nature. Unlike most of the classification algorithms in which interference caused by normal respiratory sounds is not addressed in depth, our first contribution proposes a novel Constrained Low-Rank Non-negative Matrix Factorization (CL-RNMF) approach, never applied to classification of wheezing as far as the authors' knowledge, which incorporates several constraints (sparseness and smoothness) and a low-rank configuration to extract the wheezing spectral content, minimizing the acoustic interference from normal respiratory sounds. The second contribution automatically analyzes the harmonic structure of the energy distribution associated with the estimated wheezing spectrogram to classify the type of wheezing. Experimental results report that: (i) the proposed method outperforms the most recent and relevant state-of-the-art wheezing classification method by approximately 8% in accuracy; (ii) unlike state-of-the-art methods based on classifiers, the proposed method uses an unsupervised approach that does not require any training.
Collapse
Affiliation(s)
- Juan De La Torre Cruz
- Department of Telecommunication Engineering, University of Jaen, Campus Cientifico-Tecnologico de Linares, Avda. de la Universidad, s/n, Linares, 23700 Jaen, Spain; (F.J.C.Q.); (N.R.R.); (S.G.G.); (J.J.C.O.)
| | - Francisco Jesús Cañadas Quesada
- Department of Telecommunication Engineering, University of Jaen, Campus Cientifico-Tecnologico de Linares, Avda. de la Universidad, s/n, Linares, 23700 Jaen, Spain; (F.J.C.Q.); (N.R.R.); (S.G.G.); (J.J.C.O.)
| | - Nicolás Ruiz Reyes
- Department of Telecommunication Engineering, University of Jaen, Campus Cientifico-Tecnologico de Linares, Avda. de la Universidad, s/n, Linares, 23700 Jaen, Spain; (F.J.C.Q.); (N.R.R.); (S.G.G.); (J.J.C.O.)
| | - Sebastián García Galán
- Department of Telecommunication Engineering, University of Jaen, Campus Cientifico-Tecnologico de Linares, Avda. de la Universidad, s/n, Linares, 23700 Jaen, Spain; (F.J.C.Q.); (N.R.R.); (S.G.G.); (J.J.C.O.)
| | - Julio José Carabias Orti
- Department of Telecommunication Engineering, University of Jaen, Campus Cientifico-Tecnologico de Linares, Avda. de la Universidad, s/n, Linares, 23700 Jaen, Spain; (F.J.C.Q.); (N.R.R.); (S.G.G.); (J.J.C.O.)
| | - Gerardo Peréz Chica
- Pneumology Clinical Management Unit of the University Hospital of Jaen, Av. del Ejercito Espanol, 10, 23007 Jaen, Spain;
| |
Collapse
|
29
|
Fraiwan L, Hassanin O, Fraiwan M, Khassawneh B, Ibnian AM, Alkhodari M. Automatic identification of respiratory diseases from stethoscopic lung sound signals using ensemble classifiers. Biocybern Biomed Eng 2021. [DOI: 10.1016/j.bbe.2020.11.003] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
30
|
Nguyen T, Pernkopf F. Lung Sound Classification Using Snapshot Ensemble of Convolutional Neural Networks. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2020:760-763. [PMID: 33018097 DOI: 10.1109/embc44109.2020.9176076] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
We propose a robust and efficient lung sound classification system using a snapshot ensemble of convolutional neural networks (CNNs). A robust CNN architecture is used to extract high-level features from log mel spectrograms. The CNN architecture is trained on a cosine cycle learning rate schedule. Capturing the best model of each training cycle allows to obtain multiple models settled on various local optima from cycle to cycle at the cost of training a single mode. Therefore, the snapshot ensemble boosts performance of the proposed system while keeping the drawback of expensive training of ensembles moderate. To deal with the class-imbalance of the dataset, temporal stretching and vocal tract length perturbation (VTLP) for data augmentation and the focal loss objective are used. Empirically, our system outperforms state-of-the-art systems for the prediction task of four classes (normal, crackles, wheezes, and both crackles and wheezes) and two classes (normal and abnormal (i.e. crackles, wheezes, and both crackles and wheezes)) and achieves 78.4% and 83.7% ICBHI specific micro-averaged accuracy, respectively. The average accuracy is repeated on ten random splittings of 80% training and 20% testing data using the ICBHI 2017 dataset of respiratory cycles.
Collapse
|
31
|
Tsai KH, Wang WC, Cheng CH, Tsai CY, Wang JK, Lin TH, Fang SH, Chen LC, Tsao Y. Blind Monaural Source Separation on Heart and Lung Sounds Based on Periodic-Coded Deep Autoencoder. IEEE J Biomed Health Inform 2020; 24:3203-3214. [PMID: 32795973 DOI: 10.1109/jbhi.2020.3016831] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Auscultation is the most efficient way to diagnose cardiovascular and respiratory diseases. To reach accurate diagnoses, a device must be able to recognize heart and lung sounds from various clinical situations. However, the recorded chest sounds are mixed by heart and lung sounds. Thus, effectively separating these two sounds is critical in the pre-processing stage. Recent advances in machine learning have progressed on monaural source separations, but most of the well-known techniques require paired mixed sounds and individual pure sounds for model training. As the preparation of pure heart and lung sounds is difficult, special designs must be considered to derive effective heart and lung sound separation techniques. In this study, we proposed a novel periodicity-coded deep auto-encoder (PC-DAE) approach to separate mixed heart-lung sounds in an unsupervised manner via the assumption of different periodicities between heart rate and respiration rate. The PC-DAE benefits from deep-learning-based models by extracting representative features and considers the periodicity of heart and lung sounds to carry out the separation. We evaluated PC-DAE on two datasets. The first one includes sounds from the Student Auscultation Manikin (SAM), and the second is prepared by recording chest sounds in real-world conditions. Experimental results indicate that PC-DAE outperforms several well-known separation works in terms of standardized evaluation metrics. Moreover, waveforms and spectrograms demonstrate the effectiveness of PC-DAE compared to existing approaches. It is also confirmed that by using the proposed PC-DAE as a pre-processing stage, the heart sound recognition accuracies can be notably boosted. The experimental results confirmed the effectiveness of PC-DAE and its potential to be used in clinical applications.
Collapse
|
32
|
Acharya J, Basu A. Deep Neural Network for Respiratory Sound Classification in Wearable Devices Enabled by Patient Specific Model Tuning. IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS 2020; 14:535-544. [PMID: 32191898 DOI: 10.1109/tbcas.2020.2981172] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
The primary objective of this paper is to build classification models and strategies to identify breathing sound anomalies (wheeze, crackle) for automated diagnosis of respiratory and pulmonary diseases. In this work we propose a deep CNN-RNN model that classifies respiratory sounds based on Mel-spectrograms. We also implement a patient specific model tuning strategy that first screens respiratory patients and then builds patient specific classification models using limited patient data for reliable anomaly detection. Moreover, we devise a local log quantization strategy for model weights to reduce the memory footprint for deployment in memory constrained systems such as wearable devices. The proposed hybrid CNN-RNN model achieves a score of [Formula: see text] on four-class classification of breathing cycles for ICBHI'17 scientific challenge respiratory sound database. When the model is re-trained with patient specific data, it produces a score of [Formula: see text] for leave-one-out validation. The proposed weight quantization technique achieves ≈ 4 × reduction in total memory cost without loss of performance. The main contribution of the paper is as follows: Firstly, the proposed model is able to achieve state of the art score on the ICBHI'17 dataset. Secondly, deep learning models are shown to successfully learn domain specific knowledge when pre-trained with breathing data and produce significantly superior performance compared to generalized models. Finally, local log quantization of trained weights is shown to be able to reduce the memory requirement significantly. This type of patient-specific re-training strategy can be very useful in developing reliable long-term automated patient monitoring systems particularly in wearable healthcare solutions.
Collapse
|
33
|
Pramono RXA, Imtiaz SA, Rodriguez-Villegas E. Evaluation of Mel-Frequency Cepstrum for Wheeze Analysis. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2019:4686-4689. [PMID: 31946908 DOI: 10.1109/embc.2019.8857848] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Monitoring of wheezes is an integral part of managing Chronic Respiratory Diseases such as asthma and Chronic Obstructive Pulmonary Disease (COPD). Recently, there is a growing interest in automatic detection of wheezes and the use of Mel-Frequency Cepstral Coefficients (MFCC) have been shown to achieve encouraging detection performance. While the successful use of MFCC for identifying wheezes has been demonstrated, it is not clear which MFCC coefficients are actually useful for detecting wheezes. The objective of this paper is to characterize and study the effectiveness of individual coefficients in discriminating between wheezes and normal respiratory sounds. The coefficients have been evaluated in terms of histogram dissimilarity and linear separability. Further, a comparison between the use of single coefficient against other combinations of coefficients is also presented. The results demonstrate MFCC-2 coefficient to be significantly more effective than all the other coefficients in discriminating between wheezes and normal respiratory sounds sampled at 8000 Hz.
Collapse
|
34
|
McNulty J, Reilly RB, Taylor TE, O'Dwyer SM, Costello RW, Zigel Y. Automatic Audio-Based Classification of Patient Inhaler Use: A Pharmacy Based Study. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2019:2606-2609. [PMID: 31946430 DOI: 10.1109/embc.2019.8857132] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
Chronic respiratory diseases may be controlled through the delivery of medication to the airways and lungs using an inhaler. However, adherence to correct inhaler technique is poor, which impedes patients from receiving maximum clinical benefit from their medication. In this study, the Inhaler Compliance Assessment device was employed to record audio of patients using a Diskus dry powder inhaler. An algorithm that classifies inhaler sounds (blister, inhalation, interference) was developed to automatically assess patient adherence from these inhaler audio recordings. The presented algorithm employed audio-based signal processing methods and statistical modeling in the form of quadratic discriminant analysis (QDA). A total of 350 audio recordings were obtained from 70 patients. The acquired audio dataset was split evenly for training and testing. A total accuracy of 85.35% was obtained (testing dataset) for this 3-class classification system. A sensitivity of 89.22% and 70% was obtained for inhalation and blister detection respectively. This approach may have significant clinical impact by providing healthcare professionals with an efficient, objective method of monitoring patient adherence to inhaler treatment.
Collapse
|
35
|
Demir F, Sengur A, Bajaj V. Convolutional neural networks based efficient approach for classification of lung diseases. Health Inf Sci Syst 2019; 8:4. [PMID: 31915523 PMCID: PMC6928168 DOI: 10.1007/s13755-019-0091-3] [Citation(s) in RCA: 59] [Impact Index Per Article: 11.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2019] [Accepted: 10/28/2019] [Indexed: 11/30/2022] Open
Abstract
Treatment of lung diseases, which are the third most common cause of death in the world, is of great importance in the medical field. Many studies using lung sounds recorded with stethoscope have been conducted in the literature in order to diagnose the lung diseases with artificial intelligence-compatible devices and to assist the experts in their diagnosis. In this paper, ICBHI 2017 database which includes different sample frequencies, noise and background sounds was used for the classification of lung sounds. The lung sound signals were initially converted to spectrogram images by using time–frequency method. The short time Fourier transform (STFT) method was considered as time–frequency transformation. Two deep learning based approaches were used for lung sound classification. In the first approach, a pre-trained deep convolutional neural networks (CNN) model was used for feature extraction and a support vector machine (SVM) classifier was used in classification of the lung sounds. In the second approach, the pre-trained deep CNN model was fine-tuned (transfer learning) via spectrogram images for lung sound classification. The accuracies of the proposed methods were tested by using the ten-fold cross validation. The accuracies for the first and second proposed methods were 65.5% and 63.09%, respectively. The obtained accuracies were then compared with some of the existing results and it was seen that obtained scores were better than the other results.
Collapse
Affiliation(s)
- Fatih Demir
- 1Electrical and Electronics Engineering Dept., Technology Faculty, Firat University, Elazig, Turkey
| | - Abdulkadir Sengur
- 1Electrical and Electronics Engineering Dept., Technology Faculty, Firat University, Elazig, Turkey
| | | |
Collapse
|
36
|
Chen H, Yuan X, Li J, Pei Z, Zheng X. Automatic Multi-Level In-Exhale Segmentation and Enhanced Generalized S-Transform for wheezing detection. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2019; 178:163-173. [PMID: 31416545 DOI: 10.1016/j.cmpb.2019.06.024] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/09/2019] [Revised: 06/09/2019] [Accepted: 06/25/2019] [Indexed: 06/10/2023]
Abstract
BACKGROUND AND OBJECTIVE Wheezing is a common symptom of patients caused by asthma and chronic obstructive pulmonary diseases. Wheezing detection identifies wheezing lung sounds and helps physicians in diagnosis, monitoring, and treatment of pulmonary diseases. Different from the traditional way to detect wheezing sounds using digital image process methods, automatic wheezing detection uses computerized tools or algorithms to objectively and accurately assess and evaluate lung sounds. We propose an innovative machine learning-based approach for wheezing detection. The phases of the respiratory sounds are separated automatically and the wheezing features are extracted accordingly to improve the classification accuracy. METHODS To enhance the features of wheezing for classification, the Adaptive Multi-Level In-Exhale Segmentation (AMIE_SEG) is proposed to automatically and precisely segment the respiratory sounds into inspiratory and expiratory phases. Furthermore, the Enhanced Generalized S-Transform (EGST) is proposed to extract the wheezing features. The highlighted features of wheezing improve the accuracy of wheezing detection with machine learning-based classifiers. RESULTS To evaluate the novelty and superiority of the proposed AMIE_SEG and EGST for wheezing detection, we employ three machine learning-based classifiers, Support Vector Machine (SVM), Extreme Learning Machine (ELM) and K-Nearest Neighbor (KNN), with public datasets at segment level and record level respectively. According to the experimental results, the proposed method performs the best using the KNN classifier at segment level, with the measured accuracy, sensitivity, specificity as 98.62%, 95.9% and 99.3% in average respectively. On the other aspect, at record level, the three classifiers perform excellent, with the accuracy, sensitivity, specificity up to 99.52%, 100% and 99.27% respectively. We validate the method with public respiratory sounds dataset. CONCLUSION The comparison results indicate the very good performance of the proposed methods for long-term wheezing monitoring and telemedicine.
Collapse
Affiliation(s)
- Hai Chen
- Faculty of Information Technology, Macau University of Science and Technology, Taipa, Macau; School of Information Technology, Beijing Normal University, Zhuhai, Zhuhai, China.
| | - Xiaochen Yuan
- Faculty of Information Technology, Macau University of Science and Technology, Taipa, Macau.
| | - Jianqing Li
- Faculty of Information Technology, Macau University of Science and Technology, Taipa, Macau.
| | - Zhiyuan Pei
- School of Information Technology, Beijing Normal University, Zhuhai, Zhuhai, China.
| | - Xiaobin Zheng
- Fifth Affiliated Hospital of Sun Yat-sen University, Zhuhai, China.
| |
Collapse
|
37
|
Development of Machine Learning for Asthmatic and Healthy Voluntary Cough Sounds: A Proof of Concept Study. APPLIED SCIENCES-BASEL 2019. [DOI: 10.3390/app9142833] [Citation(s) in RCA: 20] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Abstract
(1) Background: Cough is a major presentation in childhood asthma. Here, we aim to develop a machine-learning based cough sound classifier for asthmatic and healthy children. (2) Methods: Children less than 16 years old were randomly recruited in a Children’s Hospital, from February 2017 to April 2018, and were divided into 2 cohorts—healthy children and children with acute asthma presenting with cough. Children with other concurrent respiratory conditions were excluded in the asthmatic cohort. Demographic data, duration of cough, and history of respiratory status were obtained. Children were instructed to produce voluntary cough sounds. These clinically labeled cough sounds were randomly divided into training and testing sets. Audio features such as Mel-Frequency Cepstral Coefficients and Constant-Q Cepstral Coefficients were extracted. Using a training set, a classification model was developed with Gaussian Mixture Model–Universal Background Model (GMM-UBM). Its predictive performance was tested using the test set against the physicians’ labels. (3) Results: Asthmatic cough sounds from 89 children (totaling 1192 cough sounds) and healthy coughs from 89 children (totaling 1140 cough sounds) were analyzed. The sensitivity and specificity of the audio-based classification model was 82.81% and 84.76%, respectively, when differentiating coughs from asthmatic children versus coughs from ‘healthy’ children. (4) Conclusion: Audio-based classification using machine learning is a potentially useful technique in assisting the differentiation of asthmatic cough sounds from healthy voluntary cough sounds in children.
Collapse
|
38
|
Nabi FG, Sundaraj K, Lam CK. Identification of asthma severity levels through wheeze sound characterization and classification using integrated power features. Biomed Signal Process Control 2019. [DOI: 10.1016/j.bspc.2019.04.018] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
39
|
Ghulam Nabi F, Sundaraj K, Chee Kiang L, Palaniappan R, Sundaraj S. Wheeze sound analysis using computer-based techniques: a systematic review. ACTA ACUST UNITED AC 2019; 64:1-28. [PMID: 29087951 DOI: 10.1515/bmt-2016-0219] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2016] [Accepted: 08/24/2017] [Indexed: 11/15/2022]
Abstract
Wheezes are high pitched continuous respiratory acoustic sounds which are produced as a result of airway obstruction. Computer-based analyses of wheeze signals have been extensively used for parametric analysis, spectral analysis, identification of airway obstruction, feature extraction and diseases or pathology classification. While this area is currently an active field of research, the available literature has not yet been reviewed. This systematic review identified articles describing wheeze analyses using computer-based techniques on the SCOPUS, IEEE Xplore, ACM, PubMed and Springer and Elsevier electronic databases. After a set of selection criteria was applied, 41 articles were selected for detailed analysis. The findings reveal that 1) computerized wheeze analysis can be used for the identification of disease severity level or pathology, 2) further research is required to achieve acceptable rates of identification on the degree of airway obstruction with normal breathing, 3) analysis using combinations of features and on subgroups of the respiratory cycle has provided a pathway to classify various diseases or pathology that stem from airway obstruction.
Collapse
Affiliation(s)
- Fizza Ghulam Nabi
- School of Mechatronic Engineering, Universiti Malaysia Perlis (UniMAP), 02600 Arau, Perlis, Malaysia, Phone: +601111519452
| | - Kenneth Sundaraj
- Faculty of Electronics and Computer Engineering, Universiti Teknikal Malaysia Melaka (UTeM), 76100 Durian Tunggal, Melaka, Malaysia
| | - Lam Chee Kiang
- School of Mechatronic Engineering, Universiti Malaysia Perlis (UniMAP), 02600 Arau, Perlis, Malaysia
| | - Rajkumar Palaniappan
- School of Electronics Engineering, Vellore Institute of Technology (VIT), Tamil Nadu 632014, India
| | - Sebastian Sundaraj
- Department of Anesthesiology, Hospital Tengku Ampuan Rahimah (HTAR), 41200 Klang, Selangor, Malaysia
| |
Collapse
|
40
|
Rocha BM, Filos D, Mendes L, Serbes G, Ulukaya S, Kahya YP, Jakovljevic N, Turukalo TL, Vogiatzis IM, Perantoni E, Kaimakamis E, Natsiavas P, Oliveira A, Jácome C, Marques A, Maglaveras N, Pedro Paiva R, Chouvarda I, de Carvalho P. An open access database for the evaluation of respiratory sound classification algorithms. Physiol Meas 2019; 40:035001. [PMID: 30708353 DOI: 10.1088/1361-6579/ab03ea] [Citation(s) in RCA: 52] [Impact Index Per Article: 10.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
Abstract
OBJECTIVE Over the last few decades, there has been significant interest in the automatic analysis of respiratory sounds. However, currently there are no publicly available large databases with which new algorithms can be evaluated and compared. Further developments in the field are dependent on the creation of such databases. APPROACH This paper describes a public respiratory sound database, which was compiled for an international competition, the first scientific challenge of the IFMBE's International Conference on Biomedical and Health Informatics. The database includes 920 recordings acquired from 126 participants and two sets of annotations. One set contains 6898 annotated respiratory cycles, some including crackles, wheezes, or a combination of both, and some with no adventitious respiratory sounds. In the other set, precise locations of 10 775 events of crackles and wheezes were annotated. MAIN RESULTS The best system that participated in the challenge achieved an average score of 52.5% with the respiratory cycle annotations and an average score of 91.2% with the event annotations. SIGNIFICANCE The creation and public release of this database will be useful to the research community and could bring attention to the respiratory sound classification problem.
Collapse
Affiliation(s)
- Bruno M Rocha
- Department of Informatics Engineering, Centre for Informatics and Systems (CISUC), University of Coimbra, Coimbra, Portugal. Author to whom any correspondence should be addressed
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | |
Collapse
|
41
|
Evaluation of features for classification of wheezes and normal respiratory sounds. PLoS One 2019; 14:e0213659. [PMID: 30861052 PMCID: PMC6414007 DOI: 10.1371/journal.pone.0213659] [Citation(s) in RCA: 21] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2018] [Accepted: 02/26/2019] [Indexed: 12/01/2022] Open
Abstract
Chronic Respiratory Diseases (CRDs), such as Asthma and Chronic Obstructive Pulmonary Disease (COPD), are leading causes of deaths worldwide. Although both Asthma and COPD are not curable, they can be managed by close monitoring of symptoms to prevent worsening of the condition. One key symptom that needs to be monitored is the occurrence of wheezing sounds during breathing since its early identification could prevent serious exacerbations. Since wheezing can happen randomly without warning, a long-term monitoring system with automatic wheeze detection could be extremely helpful to manage these respiratory diseases. This study evaluates the discriminatory ability of different types of feature used in previous related studies, with a total size of 105 individual features, for automatic identification of wheezing sound during breathing. A linear classifier is used to determine the best features for classification by evaluating several performance metrics, including ranksum statistical test, area under the sensitivity-–specificity curve (AUC), F1 score, Matthews Correlation Coefficient (MCC), and relative computation time. Tonality index attained the highest effect size, at 87.95%, and was found to be the feature with the lowest p-value when ranksum significance test was performed. Third MFCC coefficient achieved the highest AUC and average optimum F1 score at 0.8919 and 82.67% respectively, while the highest average optimum MCC was obtained by the first coefficient of a 6th order LPC. The best possible combination of two and three features for wheeze detection is also studied. The study concludes with an analysis of the different trade-offs between accuracy, reliability, and computation requirements of the different features since these will be highly useful for researchers when designing algorithms for automatic wheeze identification.
Collapse
|
42
|
Niu J, Cai M, Shi Y, Ren S, Xu W, Gao W, Luo Z, Reinhardt JM. A Novel Method for Automatic Identification of Breathing State. Sci Rep 2019; 9:103. [PMID: 30643176 PMCID: PMC6331627 DOI: 10.1038/s41598-018-36454-5] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2018] [Accepted: 11/20/2018] [Indexed: 11/20/2022] Open
Abstract
Sputum deposition blocks the airways of patients and leads to blood oxygen desaturation. Medical staff must periodically check the breathing state of intubated patients. This process increases staff workload. In this paper, we describe a system designed to acquire respiratory sounds from intubated subjects, extract the audio features, and classify these sounds to detect the presence of sputum. Our method uses 13 features extracted from the time-frequency spectrum of the respiratory sounds. To test our system, 220 respiratory sound samples were collected. Half of the samples were collected from patients with sputum present, and the remainder were collected from patients with no sputum present. Testing was performed based on ten-fold cross-validation. In the ten-fold cross-validation experiment, the logistic classifier identified breath sounds with sputum present with a sensitivity of 93.36% and a specificity of 93.36%. The feature extraction and classification methods are useful and reliable for sputum detection. This approach differs from waveform research and can provide a better visualization of sputum conditions. The proposed system can be used in the ICU to inform medical staff when sputum is present in a patient's trachea.
Collapse
Affiliation(s)
- Jinglong Niu
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191, China
- Department of Biomedical Engineering, University of Iowa, Iowa City, IA, 52246, United States
| | - Maolin Cai
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191, China
| | - Yan Shi
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191, China.
| | - Shuai Ren
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191, China
| | - Weiqing Xu
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, 100191, China
| | - Wei Gao
- Department of Respiration, Beijing Anzhen Hospital, Capital Medical University, Beijing, 100029, China.
| | - Zujin Luo
- Department of Respiratory and Critical Care Medicine, Beijing Engineering Research Center of Respiratory and Critical Care Medicine, Beijing Institute of Respiratory Medicine, Beijing Chao-Yang Hospital,Capital Medical University, Beijing, 100043, China
| | - Joseph M Reinhardt
- Department of Biomedical Engineering, University of Iowa, Iowa City, IA, 52246, United States
| |
Collapse
|
43
|
Ulukaya S, Serbes G, Kahya YP. Wheeze type classification using non-dyadic wavelet transform based optimal energy ratio technique. Comput Biol Med 2018; 104:175-182. [PMID: 30496939 DOI: 10.1016/j.compbiomed.2018.11.004] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2018] [Revised: 11/03/2018] [Accepted: 11/05/2018] [Indexed: 10/27/2022]
Abstract
BACKGROUND AND OBJECTIVE Wheezes in pulmonary sounds are anomalies which are often associated with obstructive type of lung diseases. The previous works on wheeze-type classification focused mainly on using fixed time-frequency/scale resolution based on Fourier and wavelet transforms. The main contribution of the proposed method, in which the time-scale resolution can be tuned according to the signal of interest, is to discriminate monophonic and polyphonic wheezes with higher accuracy than previously suggested time and time-frequency/scale based methods. METHODS An optimal Rational Dilation Wavelet Transform (RADWT) based peak energy ratio (PER) parameter selection method is proposed to discriminate wheeze types. Previously suggested Quartile Frequency Ratios, Mean Crossing Irregularity, Multiple Signal Classification, Mel-frequency Cepstrum and Dyadic Discrete Wavelet Transform approaches are also applied and the superiority of the proposed method is demonstrated in leave-one-out (LOO) and leave-one-subject-out (LOSO) cross validation schemes with support vector machine (SVM), k nearest neighbor (k-NN) and extreme learning machine (ELM) classifiers. RESULTS The results show that the proposed RADWT based method outperforms the state-of-the-art time, frequency, time-frequency and time-scale domain approaches for all classifiers in both LOO and LOSO cross validation settings. The highest accuracy values are obtained as 86% and 82.9% in LOO and LOSO respectively when the proposed PER features are fed into SVM. CONCLUSIONS It is concluded that time and frequency domain characteristics of wheezes are not steady and hence, tunable time-scale representations are more successful in discriminating polyphonic and monophonic wheezes when compared with conventional fixed resolution representations.
Collapse
Affiliation(s)
- Sezer Ulukaya
- Department of Electrical and Electronics Engineering, Boǧaziçi University, 34342, Istanbul, Turkey; Department of Electrical and Electronics Engineering, Trakya University, 22030, Edirne, Turkey.
| | - Gorkem Serbes
- Department of Biomedical Engineering, Yildiz Technical University, 34220, Istanbul, Turkey.
| | - Yasemin P Kahya
- Department of Electrical and Electronics Engineering, Boǧaziçi University, 34342, Istanbul, Turkey.
| |
Collapse
|
44
|
Characterization and classification of asthmatic wheeze sounds according to severity level using spectral integrated features. Comput Biol Med 2018; 104:52-61. [PMID: 30439599 DOI: 10.1016/j.compbiomed.2018.10.035] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2018] [Revised: 10/31/2018] [Accepted: 10/31/2018] [Indexed: 11/21/2022]
Abstract
OBJECTIVE This study aimed to investigate and classify wheeze sounds of asthmatic patients according to their severity level (mild, moderate and severe) using spectral integrated (SI) features. METHOD Segmented and validated wheeze sounds were obtained from auscultation recordings of the trachea and lower lung base of 55 asthmatic patients during tidal breathing manoeuvres. The segments were multi-labelled into 9 groups based on the auscultation location and/or breath phases. Bandwidths were selected based on the physiology, and a corresponding SI feature was computed for each segment. Univariate and multivariate statistical analyses were then performed to investigate the discriminatory behaviour of the features with respect to the severity levels in the various groups. The asthmatic severity levels in the groups were then classified using the ensemble (ENS), support vector machine (SVM) and k-nearest neighbour (KNN) methods. RESULTS AND CONCLUSION All statistical comparisons exhibited a significant difference (p < 0.05) among the severity levels with few exceptions. In the classification experiments, the ensemble classifier exhibited better performance in terms of sensitivity, specificity and positive predictive value (PPV). The trachea inspiratory group showed the highest classification performance compared with all the other groups. Overall, the best PPV for the mild, moderate and severe samples were 95% (ENS), 88% (ENS) and 90% (SVM), respectively. With respect to location, the tracheal related wheeze sounds were most sensitive and specific predictors of asthma severity levels. In addition, the classification performances of the inspiratory and expiratory related groups were comparable, suggesting that the samples from these locations are equally informative.
Collapse
|
45
|
Abstract
Recent developments in sensor technology and computational analysis methods enable new strategies to measure and interpret lung acoustic signals that originate internally, such as breathing or vocal sounds, or are externally introduced, such as in chest percussion or airway insonification. A better understanding of these sounds has resulted in a new instrumentation that allows for highly accurate as well as portable options for measurement in the hospital, in the clinic, and even at home. This review outlines the instrumentation for acoustic stimulation and measurement of the lungs. We first review the fundamentals of acoustic lung signals and the pathophysiology of the diseases that these signals are used to detect. Then, we focus on different methods of measuring and creating signals that have been used in recent research for pulmonary disease diagnosis. These new methods, combined with signal processing and modeling techniques, lead to a reduction in noise and allow improved feature extraction and signal classification. We conclude by presenting the results of human subject studies taking advantage of both the instrumentation and signal processing tools to accurately diagnose common lung diseases. This paper emphasizes the active areas of research within modern lung acoustics and encourages the standardization of future work in this field.
Collapse
|
46
|
Nodera H, Osaki Y, Yamazaki H, Mori A, Izumi Y, Kaji R. Classification of needle-EMG resting potentials by machine learning. Muscle Nerve 2018; 59:224-228. [PMID: 30353953 DOI: 10.1002/mus.26363] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2018] [Revised: 10/14/2018] [Accepted: 10/16/2018] [Indexed: 11/10/2022]
Abstract
INTRODUCTION The diagnostic importance of audio signal characteristics in needle electromyography (EMG) is well established. Given the recent advent of audio-sound identification by artificial intelligence, we hypothesized that the extraction of characteristic resting EMG signals and application of machine learning algorithms could help classify various EMG discharges. METHODS Data files of 6 classes of resting EMG signals were divided into 2-s segments. Extraction of characteristic features (384 and 4,367 features each) was used to classify the 6 types of discharges using machine learning algorithms. RESULTS Across 841 audio files, the best overall accuracy of 90.4% was observed for the smaller feature set. Among the feature classes, mel-frequency cepstral coefficients (MFCC)-related features were useful in correct classification. CONCLUSIONS We showed that needle EMG resting signals were satisfactorily classifiable by the combination of feature extraction and machine learning, and this can be applied to clinical settings. Muscle Nerve 59:224-228, 2019.
Collapse
Affiliation(s)
- Hiroyuki Nodera
- Department of Neurology, 3-18-15 Kuramotocho, Tokushima City, 770-8503, Japan
| | - Yusuke Osaki
- Department of Neurology, 3-18-15 Kuramotocho, Tokushima City, 770-8503, Japan
| | - Hiroki Yamazaki
- Department of Neurology, 3-18-15 Kuramotocho, Tokushima City, 770-8503, Japan
| | - Atsuko Mori
- Department of Neurology, 3-18-15 Kuramotocho, Tokushima City, 770-8503, Japan
| | - Yuishin Izumi
- Department of Neurology, 3-18-15 Kuramotocho, Tokushima City, 770-8503, Japan
| | - Ryuji Kaji
- Department of Neurology, 3-18-15 Kuramotocho, Tokushima City, 770-8503, Japan
| |
Collapse
|
47
|
|
48
|
Rao A, Chu S, Batlivala N, Zetumer S, Roy S. Improved Detection of Lung Fluid With Standardized Acoustic Stimulation of the Chest. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE-JTEHM 2018; 6:3200107. [PMID: 30310761 PMCID: PMC6168182 DOI: 10.1109/jtehm.2018.2863366] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/23/2018] [Revised: 05/29/2018] [Accepted: 06/04/2018] [Indexed: 11/17/2022]
Abstract
Accumulation of excess air and water in the lungs leads to breakdown of respiratory function and is a common cause of patient hospitalization. Compact and non-invasive methods to detect the changes in lung fluid accumulation can allow physicians to assess patients’ respiratory conditions. In this paper, an acoustic transducer and a digital stethoscope system are proposed as a targeted solution for this clinical need. Alterations in the structure of the lungs lead to measurable changes which can be used to assess lung pathology. We standardize this procedure by sending a controlled signal through the lungs of six healthy subjects and six patients with lung disease. We extract mel-frequency cepstral coefficients and spectroid audio features, commonly used in classification for music retrieval, to characterize subjects as healthy or diseased. Using the \documentclass[12pt]{minimal}
\usepackage{amsmath}
\usepackage{wasysym}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage{upgreek}
\usepackage{mathrsfs}
\setlength{\oddsidemargin}{-69pt}
\begin{document}
}{}$K$
\end{document}-nearest neighbors algorithm, we demonstrate 91.7% accuracy in distinguishing between healthy subjects and patients with lung pathology.
Collapse
Affiliation(s)
- Adam Rao
- Department of Bioengineering and Therapeutic SciencesUniversity of California at San FranciscoSan FranciscoCA94158USA
| | - Simon Chu
- School of MedicineUniversity of California at San FranciscoSan FranciscoCA94143USA
| | | | - Samuel Zetumer
- School of MedicineUniversity of California at San FranciscoSan FranciscoCA94143USA
| | - Shuvo Roy
- Department of Bioengineering and Therapeutic SciencesUniversity of California at San FranciscoSan FranciscoCA94158USA
| |
Collapse
|
49
|
Mondal A, Banerjee P, Tang H. A novel feature extraction technique for pulmonary sound analysis based on EMD. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2018; 159:199-209. [PMID: 29650313 DOI: 10.1016/j.cmpb.2018.03.016] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2017] [Revised: 01/27/2018] [Accepted: 03/20/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND AND OBJECTIVE The stethoscope based auscultation technique is a primary diagnostic tool for chest sound analysis. However, the performance of this method is limited due to its dependency on physicians experience, knowledge and also clarity of the signal. To overcome this problem we need an automated computer-aided diagnostic system that will be competent in noisy environment. In this paper, a novel feature extraction technique is introduced for discriminating various pulmonary dysfunctions in an automated way based on pattern recognition algorithms. METHOD In this work, the disease correlated relevant characteristics of lung sounds signals are identified in terms of statistical distribution parameters: mean, variance, skewness, and kurtosis. These features are extracted from selective morphological components of the mapped signal in the empirical mode decomposition domain. The feature set is fed to the classifier model to differentiate their corresponding classes. RESULTS The significance of features developed are validated by conducting several experiments using supervised and unsupervised classifiers. Furthermore, the discriminating power of the proposed features is compared with three types of baseline features. The experimental result is evaluated by statistical analysis and also validated with physicians inference. CONCLUSIONS It is found that the proposed features extraction technique is superior to the baseline methods in terms of classification accuracy, sensitivity and specificity. The developed method gives better results compared to baseline methods in any circumstance. The proposed method gives a higher accuracy of 94.16, sensitivity of 100 and specificity of 93.75 for an artificial neural network classifier.
Collapse
Affiliation(s)
- Ashok Mondal
- National Institute of Technology, Karnataka, India.
| | | | - Hong Tang
- Dalian University of Technology, Dalian, China.
| |
Collapse
|
50
|
Islam MA, Bandyopadhyaya I, Bhattacharyya P, Saha G. Multichannel lung sound analysis for asthma detection. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2018; 159:111-123. [PMID: 29650306 DOI: 10.1016/j.cmpb.2018.03.002] [Citation(s) in RCA: 28] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/07/2017] [Revised: 02/25/2018] [Accepted: 03/09/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND AND OBJECTIVE Lung sound signals convey valuable information of the lung status. Auscultation is an effective technique to appreciate the condition of the respiratory system using lung sound signals. The prior works on asthma detection from lung sound signals rely on the presence of wheeze. In this paper, we have classified normal and asthmatic subjects using advanced signal processing of posterior lung sound signals, even in the absence of wheeze. METHODS We collected lung sounds of 60 subjects (30 normal and 30 asthma) using a novel 4-channel data acquisition system from four different positions over the posterior chest, as suggested by the pulmonologist. A spectral subband based feature extraction scheme is proposed that works with artificial neural network (ANN) and support vector machine (SVM) classifiers for the multichannel signal. The power spectral density (PSD) is estimated from extracted lung sound cycle using Welch's method, which then decomposed into uniform subbands. A set of statistical features is computed from each subband and applied to ANN and SVM classifiers to classify normal and asthmatic subjects. RESULTS In the first part of this study, the performances of each individual channel and four channels together are evaluated where the combined channel performance is found superior to that of individual channels. Next, the performances of all possible combinations of the channels are investigated and the best classification accuracies of 89.2( ± 3.87)% and 93.3( ± 3.10)% are achieved for 2-channel and 3-channel combinations in ANN and SVM classifiers, respectively. CONCLUSIONS The proposed multichannel asthma detection method where the presence of wheeze in lung sound is not a necessary requirement, outperforms commonly used lung sound classification methods in this field and provides significant relative improvement. The channel combination study gives insight into the contribution of respective lung sound collection areas and their combinations in asthma detection.
Collapse
Affiliation(s)
- Md Ariful Islam
- Department of Electronics and Electrical Communication Engineering, Indian Institute of Technology, Kharagpur, Kharagpur 721302, India.
| | - Irin Bandyopadhyaya
- Department of Electronics and Electrical Communication Engineering, Indian Institute of Technology, Kharagpur, Kharagpur 721302, India.
| | | | - Goutam Saha
- Department of Electronics and Electrical Communication Engineering, Indian Institute of Technology, Kharagpur, Kharagpur 721302, India.
| |
Collapse
|