1
|
Sankaramurthy P, Palaniswamy R, Sellamuthu S, Chelladurai F, Murugadhas A. Lung disease prediction based on CT images using REInf-net and world cup optimization based BI-LSTM classification. NETWORK (BRISTOL, ENGLAND) 2024:1-34. [PMID: 39252464 DOI: 10.1080/0954898x.2024.2392782] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Revised: 06/11/2024] [Accepted: 08/08/2024] [Indexed: 09/11/2024]
Abstract
A major global source of disability as well as mortality is respiratory illness. Though visual evaluation of computed tomography (CT) images and chest radiographs are a primary diagnostic for respiratory illnesses, it is limited in its ability to assess severity and predict patient outcomes due to low specificity and fundamental infectious organisms. In order to address these problems, world cup optimization-based Bi-LSTM classification and lung disease prediction on CT images using REINF-net were employed. To enhance the image quality, the gathered lung CT images are pre-processed using Lucy Richardson and CLAHE algorithms. For the purpose of lung infection segmentation, the pre-processed images are segmented using the REInf-net. The GLRLM method is used to extract features from the segmented images. In order to predict lung disease in CT images, the extracted features are trained using the Bi-LSTM based on world cup optimization. Accuracy, Precision, recall, Error and Specificity for the proposed model are 97.8%, 96.7%, 96.7%, 2.2% and 98.3%. These evaluated values are contrasted with the results of existing methods like WCO-BiLSTM, MLP, CNN and LSTM. Finally, the Lung disease prediction based on CT images using REINF-Net and world cup optimization based BI-LSTM classification performs better than the existing model.
Collapse
Affiliation(s)
- Padmini Sankaramurthy
- Department of Computing Technologies, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| | - Renukadevi Palaniswamy
- Department of Computing Technologies, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| | - Suseela Sellamuthu
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, India
| | - Fancy Chelladurai
- Department of Networking and Communications, School of Computing, Faculty of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| | - Anand Murugadhas
- Department of Networking and Communications, School of Computing, Faculty of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| |
Collapse
|
2
|
Mochizuki H, Hirai K, Furuya H, Niimura F, Suzuki K, Okino T, Ikeda M, Noto H. The analysis of lung sounds in infants and children with a history of wheezing/asthma using an automatic procedure. BMC Pulm Med 2024; 24:394. [PMID: 39143523 PMCID: PMC11323603 DOI: 10.1186/s12890-024-03210-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Accepted: 08/09/2024] [Indexed: 08/16/2024] Open
Abstract
BACKGROUND Lung sound analysis parameters have been reported to be useful biomarkers for evaluating airway condition. We developed an automatic lung sound analysis software program for infants and children based on lung sound spectral curves of frequency and power by leveraging machine learning (ML) technology. METHODS To put this software program into clinical practice, in Study 1, the reliability and reproducibility of the software program using data from younger children were examined. In Study 2, the relationship between lung sound parameters and respiratory flow (L/s) was evaluated using data from older children. In Study 3, we conducted a survey using the ATS-DLD questionnaire to evaluate the clinical usefulness. The survey focused on the history of wheezing and allergies, among healthy 3-year-old infants, and then measured lung sounds. The clinical usefulness was evaluated by comparing the questionnaire results with the results of the new lung sound parameters. RESULTS In Studies 1 and 2, the parameters of the new software program demonstrated excellent reproducibility and reliability, and were not affected by airflow (L/s). In Study 3, infants with a history of wheezing showed lower FAP0 and RPF75p (p < 0.001 and p = 0.025, respectively) and higher PAP0 (p = 0.001) than healthy infants. Furthermore, infants with asthma/asthma-like bronchitis showed lower FAP0 (p = 0.002) and higher PAP0 (p = 0.001) than healthy infants. CONCLUSIONS Lung sound parameters obtained using the ML algorithm were able to accurately assess the respiratory condition of infants. These parameters are useful for the early detection and intervention of childhood asthma.
Collapse
Affiliation(s)
- Hiroyuki Mochizuki
- Department of Pediatrics, Tokai University Hachioji Hospital, Hachioji, Japan.
- Department of Pediatrics, Tokai University School of Medicine, Shimokasuya 143, Isehara, Kanagawa, 259-1193, Japan.
| | - Kota Hirai
- Department of Pediatrics, Tokai University Hachioji Hospital, Hachioji, Japan
- Department of Pediatrics, Tokai University School of Medicine, Shimokasuya 143, Isehara, Kanagawa, 259-1193, Japan
| | - Hiroyuki Furuya
- Department of Basic Clinical Science and Public Health, Tokai University School of Medicine, Isehara, Japan
| | - Fumio Niimura
- Department of Pediatrics, Tokai University Hachioji Hospital, Hachioji, Japan
- Department of Pediatrics, Tokai University School of Medicine, Shimokasuya 143, Isehara, Kanagawa, 259-1193, Japan
| | - Kenta Suzuki
- Murata Manufacturing Co., Ltd, Nagaokakyo, Japan
| | | | - Miki Ikeda
- Murata Manufacturing Co., Ltd, Nagaokakyo, Japan
| | | |
Collapse
|
3
|
Khan R, Khan SU, Saeed U, Koo IS. Auscultation-Based Pulmonary Disease Detection through Parallel Transformation and Deep Learning. Bioengineering (Basel) 2024; 11:586. [PMID: 38927822 PMCID: PMC11200393 DOI: 10.3390/bioengineering11060586] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Revised: 06/05/2024] [Accepted: 06/06/2024] [Indexed: 06/28/2024] Open
Abstract
Respiratory diseases are among the leading causes of death, with many individuals in a population frequently affected by various types of pulmonary disorders. Early diagnosis and patient monitoring (traditionally involving lung auscultation) are essential for the effective management of respiratory diseases. However, the interpretation of lung sounds is a subjective and labor-intensive process that demands considerable medical expertise, and there is a good chance of misclassification. To address this problem, we propose a hybrid deep learning technique that incorporates signal processing techniques. Parallel transformation is applied to adventitious respiratory sounds, transforming lung sound signals into two distinct time-frequency scalograms: the continuous wavelet transform and the mel spectrogram. Furthermore, parallel convolutional autoencoders are employed to extract features from scalograms, and the resulting latent space features are fused into a hybrid feature pool. Finally, leveraging a long short-term memory model, a feature from the latent space is used as input for classifying various types of respiratory diseases. Our work is evaluated using the ICBHI-2017 lung sound dataset. The experimental findings indicate that our proposed method achieves promising predictive performance, with average values for accuracy, sensitivity, specificity, and F1-score of 94.16%, 89.56%, 99.10%, and 89.56%, respectively, for eight-class respiratory diseases; 79.61%, 78.55%, 92.49%, and 78.67%, respectively, for four-class diseases; and 85.61%, 83.44%, 83.44%, and 84.21%, respectively, for binary-class (normal vs. abnormal) lung sounds.
Collapse
Affiliation(s)
- Rehan Khan
- Department of Electrical Electronic and Computer Engineering, University of Ulsan, Ulsan 44610, Republic of Korea; (R.K.); (S.U.K.)
| | - Shafi Ullah Khan
- Department of Electrical Electronic and Computer Engineering, University of Ulsan, Ulsan 44610, Republic of Korea; (R.K.); (S.U.K.)
| | - Umer Saeed
- Research Centre for Intelligent Healthcare, Coventry University, Coventry CV1 5FB, UK;
| | - In-Soo Koo
- Department of Electrical Electronic and Computer Engineering, University of Ulsan, Ulsan 44610, Republic of Korea; (R.K.); (S.U.K.)
| |
Collapse
|
4
|
Sabry AH, I. Dallal Bashi O, Nik Ali N, Mahmood Al Kubaisi Y. Lung disease recognition methods using audio-based analysis with machine learning. Heliyon 2024; 10:e26218. [PMID: 38420389 PMCID: PMC10900411 DOI: 10.1016/j.heliyon.2024.e26218] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Revised: 12/11/2023] [Accepted: 02/08/2024] [Indexed: 03/02/2024] Open
Abstract
The use of computer-based automated approaches and improvements in lung sound recording techniques have made lung sound-based diagnostics even better and devoid of subjectivity errors. Using a computer to evaluate lung sound features more thoroughly with the use of analyzing changes in lung sound behavior, recording measurements, suppressing the presence of noise contaminations, and graphical representations are all made possible by computer-based lung sound analysis. This paper starts with a discussion of the need for this research area, providing an overview of the field and the motivations behind it. Following that, it details the survey methodology used in this work. It presents a discussion on the elements of sound-based lung disease classification using machine learning algorithms. This includes commonly prior considered datasets, feature extraction techniques, pre-processing methods, artifact removal methods, lung-heart sound separation, deep learning algorithms, and wavelet transform of lung audio signals. The study introduces studies that review lung screening including a summary table of these references and discusses the literature gaps in the existing studies. It is concluded that the use of sound-based machine learning in the classification of respiratory diseases has promising results. While we believe this material will prove valuable to physicians and researchers exploring sound-signal-based machine learning, large-scale investigations remain essential to solidify the findings and foster wider adoption within the medical community.
Collapse
Affiliation(s)
- Ahmad H. Sabry
- Department of Medical Instrumentation Engineering Techniques, Shatt Al-Arab University College, Basra, Iraq
| | - Omar I. Dallal Bashi
- Medical Technical Institute, Northern Technical University, 95G2+P34, Mosul, 41002, Iraq
| | - N.H. Nik Ali
- School of Electrical Engineering, College of Engineering, Universiti Teknologi MARA, 40450 Shah Alam, Selangor, Malaysia
| | - Yasir Mahmood Al Kubaisi
- Department of Sustainability Management, Dubai Academic Health Corporation, Dubai, 4545, United Arab Emirates
| |
Collapse
|
5
|
Im S, Kim T, Min C, Kang S, Roh Y, Kim C, Kim M, Kim SH, Shim K, Koh JS, Han S, Lee J, Kim D, Kang D, Seo S. Real-time counting of wheezing events from lung sounds using deep learning algorithms: Implications for disease prediction and early intervention. PLoS One 2023; 18:e0294447. [PMID: 37983213 PMCID: PMC10659186 DOI: 10.1371/journal.pone.0294447] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2023] [Accepted: 10/23/2023] [Indexed: 11/22/2023] Open
Abstract
This pioneering study aims to revolutionize self-symptom management and telemedicine-based remote monitoring through the development of a real-time wheeze counting algorithm. Leveraging a novel approach that includes the detailed labeling of one breathing cycle into three types: break, normal, and wheeze, this study not only identifies abnormal sounds within each breath but also captures comprehensive data on their location, duration, and relationships within entire respiratory cycles, including atypical patterns. This innovative strategy is based on a combination of a one-dimensional convolutional neural network (1D-CNN) and a long short-term memory (LSTM) network model, enabling real-time analysis of respiratory sounds. Notably, it stands out for its capacity to handle continuous data, distinguishing it from conventional lung sound classification algorithms. The study utilizes a substantial dataset consisting of 535 respiration cycles from diverse sources, including the Child Sim Lung Sound Simulator, the EMTprep Open-Source Database, Clinical Patient Records, and the ICBHI 2017 Challenge Database. Achieving a classification accuracy of 90%, the exceptional result metrics encompass the identification of each breath cycle and simultaneous detection of the abnormal sound, enabling the real-time wheeze counting of all respirations. This innovative wheeze counter holds the promise of revolutionizing research on predicting lung diseases based on long-term breathing patterns and offers applicability in clinical and non-clinical settings for on-the-go detection and remote intervention of exacerbated respiratory symptoms.
Collapse
Affiliation(s)
- Sunghoon Im
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Taewi Kim
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | | | - Sanghun Kang
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Yeonwook Roh
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Changhwan Kim
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Minho Kim
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Seung Hyun Kim
- Department of Medical Humanities, Korea University College of Medicine, Seoul, Republic of Korea
| | - KyungMin Shim
- Industry-University Cooperation Foundation, Seogyeong University, Seoul, Republic of Korea
| | - Je-sung Koh
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Seungyong Han
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - JaeWang Lee
- Department of Biomedical Laboratory Science, College of Health Science, Eulji University, Seongnam-si, Gyeonggi-do, Republic of Korea
| | - Dohyeong Kim
- University of Texas at Dallas, Richardson, TX, United States of America
| | - Daeshik Kang
- Department of Mechanical Engineering, Ajou University, Suwon-si, Gyeonggi-do, Republic of Korea
| | - SungChul Seo
- Department of Nano-Chemical, Biological and Environmental Engineering, Seogyeong University, Seoul, Republic of Korea
| |
Collapse
|
6
|
Garcia-Mendez JP, Lal A, Herasevich S, Tekin A, Pinevich Y, Lipatov K, Wang HY, Qamar S, Ayala IN, Khapov I, Gerberi DJ, Diedrich D, Pickering BW, Herasevich V. Machine Learning for Automated Classification of Abnormal Lung Sounds Obtained from Public Databases: A Systematic Review. Bioengineering (Basel) 2023; 10:1155. [PMID: 37892885 PMCID: PMC10604310 DOI: 10.3390/bioengineering10101155] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Revised: 09/15/2023] [Accepted: 09/26/2023] [Indexed: 10/29/2023] Open
Abstract
Pulmonary auscultation is essential for detecting abnormal lung sounds during physical assessments, but its reliability depends on the operator. Machine learning (ML) models offer an alternative by automatically classifying lung sounds. ML models require substantial data, and public databases aim to address this limitation. This systematic review compares characteristics, diagnostic accuracy, concerns, and data sources of existing models in the literature. Papers published from five major databases between 1990 and 2022 were assessed. Quality assessment was accomplished with a modified QUADAS-2 tool. The review encompassed 62 studies utilizing ML models and public-access databases for lung sound classification. Artificial neural networks (ANN) and support vector machines (SVM) were frequently employed in the ML classifiers. The accuracy ranged from 49.43% to 100% for discriminating abnormal sound types and 69.40% to 99.62% for disease class classification. Seventeen public databases were identified, with the ICBHI 2017 database being the most used (66%). The majority of studies exhibited a high risk of bias and concerns related to patient selection and reference standards. Summarizing, ML models can effectively classify abnormal lung sounds using publicly available data sources. Nevertheless, inconsistent reporting and methodologies pose limitations to advancing the field, and therefore, public databases should adhere to standardized recording and labeling procedures.
Collapse
Affiliation(s)
- Juan P. Garcia-Mendez
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Amos Lal
- Department of Medicine, Division of Pulmonary and Critical Care Medicine, Mayo Clinic, Rochester, MN 55905, USA
| | - Svetlana Herasevich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Aysun Tekin
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Yuliya Pinevich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
- Department of Cardiac Anesthesiology and Intensive Care, Republican Clinical Medical Center, 223052 Minsk, Belarus
| | - Kirill Lipatov
- Division of Pulmonary Medicine, Mayo Clinic Health Systems, Essentia Health, Duluth, MN 55805, USA
| | - Hsin-Yi Wang
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
- Department of Anesthesiology, Taipei Veterans General Hospital, National Yang Ming Chiao Tung University, Taipei 11217, Taiwan
- Department of Biomedical Sciences and Engineering, National Central University, Taoyuan 320317, Taiwan
| | - Shahraz Qamar
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Ivan N. Ayala
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Ivan Khapov
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | | | - Daniel Diedrich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Brian W. Pickering
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Division of Critical Care, Mayo Clinic, Rochester, MN 55905, USA (Y.P.); (H.-Y.W.); (I.K.); (V.H.)
| |
Collapse
|
7
|
Sfayyih AH, Sulaiman N, Sabry AH. A review on lung disease recognition by acoustic signal analysis with deep learning networks. JOURNAL OF BIG DATA 2023; 10:101. [PMID: 37333945 PMCID: PMC10259357 DOI: 10.1186/s40537-023-00762-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Accepted: 05/08/2023] [Indexed: 06/20/2023]
Abstract
Recently, assistive explanations for difficulties in the health check area have been made viable thanks in considerable portion to technologies like deep learning and machine learning. Using auditory analysis and medical imaging, they also increase the predictive accuracy for prompt and early disease detection. Medical professionals are thankful for such technological support since it helps them manage further patients because of the shortage of skilled human resources. In addition to serious illnesses like lung cancer and respiratory diseases, the plurality of breathing difficulties is gradually rising and endangering society. Because early prediction and immediate treatment are crucial for respiratory disorders, chest X-rays and respiratory sound audio are proving to be quite helpful together. Compared to related review studies on lung disease classification/detection using deep learning algorithms, only two review studies based on signal analysis for lung disease diagnosis have been conducted in 2011 and 2018. This work provides a review of lung disease recognition with acoustic signal analysis with deep learning networks. We anticipate that physicians and researchers working with sound-signal-based machine learning will find this material beneficial.
Collapse
Affiliation(s)
- Alyaa Hamel Sfayyih
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Universiti Putra Malaysia, 43400 Serdang, Malaysia
| | - Nasri Sulaiman
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Universiti Putra Malaysia, 43400 Serdang, Malaysia
| | - Ahmad H. Sabry
- Department of Computer Engineering, Al-Nahrain University, Al Jadriyah Bridge, 64074 Baghdad, Iraq
| |
Collapse
|
8
|
Sfayyih AH, Sabry AH, Jameel SM, Sulaiman N, Raafat SM, Humaidi AJ, Kubaiaisi YMA. Acoustic-Based Deep Learning Architectures for Lung Disease Diagnosis: A Comprehensive Overview. Diagnostics (Basel) 2023; 13:diagnostics13101748. [PMID: 37238233 DOI: 10.3390/diagnostics13101748] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2023] [Revised: 05/04/2023] [Accepted: 05/11/2023] [Indexed: 05/28/2023] Open
Abstract
Lung auscultation has long been used as a valuable medical tool to assess respiratory health and has gotten a lot of attention in recent years, notably following the coronavirus epidemic. Lung auscultation is used to assess a patient's respiratory role. Modern technological progress has guided the growth of computer-based respiratory speech investigation, a valuable tool for detecting lung abnormalities and diseases. Several recent studies have reviewed this important area, but none are specific to lung sound-based analysis with deep-learning architectures from one side and the provided information was not sufficient for a good understanding of these techniques. This paper gives a complete review of prior deep-learning-based architecture lung sound analysis. Deep-learning-based respiratory sound analysis articles are found in different databases including the Plos, ACM Digital Libraries, Elsevier, PubMed, MDPI, Springer, and IEEE. More than 160 publications were extracted and submitted for assessment. This paper discusses different trends in pathology/lung sound, the common features for classifying lung sounds, several considered datasets, classification methods, signal processing techniques, and some statistical information based on previous study findings. Finally, the assessment concludes with a discussion of potential future improvements and recommendations.
Collapse
Affiliation(s)
- Alyaa Hamel Sfayyih
- Department of Electrical and Electronic Engineering, Faculty of Engineering, University Putra Malaysia, Serdang 43400, Malaysia
| | - Ahmad H Sabry
- Department of Computer Engineering, Al-Nahrain University Al Jadriyah Bridge, Baghdad 64074, Iraq
| | | | - Nasri Sulaiman
- Department of Electrical and Electronic Engineering, Faculty of Engineering, University Putra Malaysia, Serdang 43400, Malaysia
| | - Safanah Mudheher Raafat
- Department of Control and Systems Engineering, University of Technology, Baghdad 10011, Iraq
| | - Amjad J Humaidi
- Department of Control and Systems Engineering, University of Technology, Baghdad 10011, Iraq
| | - Yasir Mahmood Al Kubaiaisi
- Department of Sustainability Management, Dubai Academic Health Corporation, Dubai 4545, United Arab Emirates
| |
Collapse
|
9
|
Irshad RR, Hussain S, Sohail SS, Zamani AS, Madsen DØ, Alattab AA, Ahmed AAA, Norain KAA, Alsaiari OAS. A Novel IoT-Enabled Healthcare Monitoring Framework and Improved Grey Wolf Optimization Algorithm-Based Deep Convolution Neural Network Model for Early Diagnosis of Lung Cancer. SENSORS (BASEL, SWITZERLAND) 2023; 23:2932. [PMID: 36991642 PMCID: PMC10052730 DOI: 10.3390/s23062932] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/08/2023] [Revised: 03/02/2023] [Accepted: 03/03/2023] [Indexed: 06/19/2023]
Abstract
Lung cancer is a high-risk disease that causes mortality worldwide; nevertheless, lung nodules are the main manifestation that can help to diagnose lung cancer at an early stage, lowering the workload of radiologists and boosting the rate of diagnosis. Artificial intelligence-based neural networks are promising technologies for automatically detecting lung nodules employing patient monitoring data acquired from sensor technology through an Internet-of-Things (IoT)-based patient monitoring system. However, the standard neural networks rely on manually acquired features, which reduces the effectiveness of detection. In this paper, we provide a novel IoT-enabled healthcare monitoring platform and an improved grey-wolf optimization (IGWO)-based deep convulution neural network (DCNN) model for lung cancer detection. The Tasmanian Devil Optimization (TDO) algorithm is utilized to select the most pertinent features for diagnosing lung nodules, and the convergence rate of the standard grey wolf optimization (GWO) algorithm is modified, resulting in an improved GWO algorithm. Consequently, an IGWO-based DCNN is trained on the optimal features obtained from the IoT platform, and the findings are saved in the cloud for the doctor's judgment. The model is built on an Android platform with DCNN-enabled Python libraries, and the findings are evaluated against cutting-edge lung cancer detection models.
Collapse
Affiliation(s)
- Reyazur Rashid Irshad
- Department of Computer Science, College of Science and Arts, Najran University, Sharurah 68341, Saudi Arabia
| | - Shahid Hussain
- Department of Computer Science and Engineering, Sejong University, Seoul 30019, Republic of Korea
| | - Shahab Saquib Sohail
- Department of Computer Science and Engineering, School of Engineering Sciences and Technology, Jamia Hamdard, New Delhi 110062, India
| | - Abu Sarwar Zamani
- Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
| | - Dag Øivind Madsen
- USN School of Business, University of South-Eastern Norway, 3511 Hønefoss, Norway
| | - Ahmed Abdu Alattab
- Department of Computer Science, College of Science and Arts, Najran University, Sharurah 68341, Saudi Arabia
- Department of Computer Science, Faculty of Computer Science and Information Systems, Thamar University, Thamar 87246, Yemen
| | | | | | - Omar Ali Saleh Alsaiari
- Department of Computer Science, College of Science and Arts, Najran University, Sharurah 68341, Saudi Arabia
| |
Collapse
|
10
|
Arya N, Mathur A, Saha S, Saha S. Proposal of SVM Utility Kernel for Breast Cancer Survival Estimation. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2023; 20:1372-1383. [PMID: 35994556 DOI: 10.1109/tcbb.2022.3198879] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
The advancement of medical research in the field of cancer prognosis and diagnosis using various modalities has put oncologists under tremendous stress. The complexity and heterogeneity involved in multiple modalities and their significantly varied clinical outcomes make it difficult to analyze the disease and provide the correct treatment. Breast cancer is the major concern among all cancers worldwide, specifically for females. To help oncologists and cancer patients, research for breast cancer survival estimation has been proposed. It ranges from complex deep neural networks to simple and interpretable architectures. We propose a utility kernel for a support vector machine (SVM) in this article. It is a simple yet powerful function, which performs better than other popular machine learning algorithms and deep neural networks in the task of breast cancer survival prediction using the TCGA-BRCA dataset. This study validates the proposed utility kernel using four different modalities (gene expression, copy number variation, clinical, and histopathological tissue images) and their multi-modal combinations. The SVM based on our utility kernel empirically proves its efficacy by achieving the highest value on various performance measures, whereas advanced deep neural networks fail to train on small and highly imbalanced breast cancer data.
Collapse
|
11
|
Zhang M, Li M, Guo L, Liu J. A Low-Cost AI-Empowered Stethoscope and a Lightweight Model for Detecting Cardiac and Respiratory Diseases from Lung and Heart Auscultation Sounds. SENSORS (BASEL, SWITZERLAND) 2023; 23:2591. [PMID: 36904794 PMCID: PMC10007545 DOI: 10.3390/s23052591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/26/2022] [Revised: 02/11/2023] [Accepted: 02/17/2023] [Indexed: 06/18/2023]
Abstract
Cardiac and respiratory diseases are the primary causes of health problems. If we can automate anomalous heart and lung sound diagnosis, we can improve the early detection of disease and enable the screening of a wider population than possible with manual screening. We propose a lightweight yet powerful model for simultaneous lung and heart sound diagnosis, which is deployable in an embedded low-cost device and is valuable in remote areas or developing countries where Internet access may not be available. We trained and tested the proposed model with the ICBHI and the Yaseen datasets. The experimental results showed that our 11-class prediction model could achieve 99.94% accuracy, 99.84% precision, 99.89% specificity, 99.66% sensitivity, and 99.72% F1 score. We designed a digital stethoscope (around USD 5) and connected it to a low-cost, single-board-computer Raspberry Pi Zero 2W (around USD 20), on which our pretrained model can be smoothly run. This AI-empowered digital stethoscope is beneficial for anyone in the medical field, as it can automatically provide diagnostic results and produce digital audio records for further analysis.
Collapse
Affiliation(s)
- Miao Zhang
- School of Mathematics, Shandong University, Jinan 250100, China
- School of Mathematics and Statistics, Shandong University, Weihai 264200, China
| | - Min Li
- School of Mathematics and Statistics, Shandong University, Weihai 264200, China
| | - Liang Guo
- School of Mathematics and Statistics, Shandong University, Weihai 264200, China
- Data Science Institute, Shandong University, Jinan 250100, China
| | - Jianya Liu
- School of Mathematics, Shandong University, Jinan 250100, China
- Data Science Institute, Shandong University, Jinan 250100, China
| |
Collapse
|
12
|
Qiu Y, Lin F, Chen W, Xu M. Pre-training in Medical Data: A Survey. MACHINE INTELLIGENCE RESEARCH 2023. [PMCID: PMC9942039 DOI: 10.1007/s11633-022-1382-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/23/2023]
Abstract
Medical data refers to health-related information associated with regular patient care or as part of a clinical trial program. There are many categories of such data, such as clinical imaging data, bio-signal data, electronic health records (EHR), and multi-modality medical data. With the development of deep neural networks in the last decade, the emerging pre-training paradigm has become dominant in that it has significantly improved machine learning methods’ performance in a data-limited scenario. In recent years, studies of pre-training in the medical domain have achieved significant progress. To summarize these technology advancements, this work provides a comprehensive survey of recent advances for pre-training on several major types of medical data. In this survey, we summarize a large number of related publications and the existing benchmarking in the medical domain. Especially, the survey briefly describes how some pre-training methods are applied to or developed for medical data. From a data-driven perspective, we examine the extensive use of pre-training in many medical scenarios. Moreover, based on the summary of recent pre-training studies, we identify several challenges in this field to provide insights for future studies.
Collapse
Affiliation(s)
- Yixuan Qiu
- The University of Queensland, Brisbane, 4072 Australia
| | - Feng Lin
- The University of Queensland, Brisbane, 4072 Australia
| | - Weitong Chen
- The University of Adelaide, Adelaide, 5005 Australia
| | - Miao Xu
- The University of Queensland, Brisbane, 4072 Australia
| |
Collapse
|
13
|
Campana MG, Delmastro F, Pagani E. Transfer learning for the efficient detection of COVID-19 from smartphone audio data. PERVASIVE AND MOBILE COMPUTING 2023; 89:101754. [PMID: 36741300 PMCID: PMC9884612 DOI: 10.1016/j.pmcj.2023.101754] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Revised: 01/18/2023] [Accepted: 01/25/2023] [Indexed: 06/12/2023]
Abstract
Disease detection from smartphone data represents an open research challenge in mobile health (m-health) systems. COVID-19 and its respiratory symptoms are an important case study in this area and their early detection is a potential real instrument to counteract the pandemic situation. The efficacy of this solution mainly depends on the performances of AI algorithms applied to the collected data and their possible implementation directly on the users' mobile devices. Considering these issues, and the limited amount of available data, in this paper we present the experimental evaluation of 3 different deep learning models, compared also with hand-crafted features, and of two main approaches of transfer learning in the considered scenario: both feature extraction and fine-tuning. Specifically, we considered VGGish, YAMNET, and L3-Net (including 12 different configurations) evaluated through user-independent experiments on 4 different datasets (13,447 samples in total). Results clearly show the advantages of L3-Net in all the experimental settings as it overcomes the other solutions by 12.3% in terms of Precision-Recall AUC as features extractor, and by 10% when the model is fine-tuned. Moreover, we note that to fine-tune only the fully-connected layers of the pre-trained models generally leads to worse performances, with an average drop of 6.6% with respect to feature extraction. Finally, we evaluate the memory footprints of the different models for their possible applications on commercial mobile devices.
Collapse
Affiliation(s)
- Mattia Giovanni Campana
- Institute for Informatics and Telematics of the National Research Council of Italy (IIT-CNR), Pisa, Italy
| | - Franca Delmastro
- Institute for Informatics and Telematics of the National Research Council of Italy (IIT-CNR), Pisa, Italy
| | - Elena Pagani
- Computer Science Department, University of Milano, Milan, Italy
| |
Collapse
|
14
|
Deep Transfer Learning Techniques-Based Automated Classification and Detection of Pulmonary Fibrosis from Chest CT Images. Processes (Basel) 2023. [DOI: 10.3390/pr11020443] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/05/2023] Open
Abstract
Pulmonary Fibrosis (PF) is a non-curable chronic lung disease. Therefore, a quick and accurate PF diagnosis is imperative. In the present study, we aim to compare the performance of the six state-of-the-art Deep Transfer Learning techniques to classify patients accurately and perform abnormality localization in Computer Tomography (CT) scan images. A total of 2299 samples comprising normal and PF-positive CT images were preprocessed. The preprocessed images were split into training (75%), validation (15%), and test data (10%). These transfer learning models were trained and validated by optimizing the hyperparameters, such as the learning rate and the number of epochs. The optimized architectures have been evaluated with different performance metrics to demonstrate the consistency of the optimized model. At epoch 26, using an optimized learning rate of 0.0000625, the ResNet50v2 model achieved the highest training and validation accuracy (training = 99.92%, validation = 99.22%) and minimum loss (training = 0.00428, validation = 0.00683) for CT images. The experimental evaluation on the independent testing data confirms that optimized ResNet50v2 outperformed every other optimized architecture under consideration achieving a perfect score of 1.0 in each of the standard performance measures such as accuracy, precision, recall, F1-score, Mathew Correlation Coefficient (MCC), Area under the Receiver Operating Characteristic (ROC-AUC) curve, and the Area under the Precision recall (AUC_PR) curve. Therefore, we can propose that the optimized ResNet50v2 is a reliable diagnostic model for automatically classifying PF-positive patients using chest CT images.
Collapse
|
15
|
Cinyol F, Baysal U, Köksal D, Babaoğlu E, Ulaşlı SS. Incorporating support vector machine to the classification of respiratory sounds by Convolutional Neural Network. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
16
|
Saldanha J, Chakraborty S, Patil S, Kotecha K, Kumar S, Nayyar A. Data augmentation using Variational Autoencoders for improvement of respiratory disease classification. PLoS One 2022; 17:e0266467. [PMID: 35960763 PMCID: PMC9374267 DOI: 10.1371/journal.pone.0266467] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/21/2021] [Accepted: 03/22/2022] [Indexed: 11/19/2022] Open
Abstract
Computerized auscultation of lung sounds is gaining importance today with the availability of lung sounds and its potential in overcoming the limitations of traditional diagnosis methods for respiratory diseases. The publicly available ICBHI respiratory sounds database is severely imbalanced, making it difficult for a deep learning model to generalize and provide reliable results. This work aims to synthesize respiratory sounds of various categories using variants of Variational Autoencoders like Multilayer Perceptron VAE (MLP-VAE), Convolutional VAE (CVAE) Conditional VAE and compare the influence of augmenting the imbalanced dataset on the performance of various lung sound classification models. We evaluated the quality of the synthetic respiratory sounds’ quality using metrics such as Fréchet Audio Distance (FAD), Cross-Correlation and Mel Cepstral Distortion. Our results showed that MLP-VAE achieved an average FAD of 12.42 over all classes, whereas Convolutional VAE and Conditional CVAE achieved an average FAD of 11.58 and 11.64 for all classes, respectively. A significant improvement in the classification performance metrics was observed upon augmenting the imbalanced dataset for certain minority classes and marginal improvement for the other classes. Hence, our work shows that deep learning-based lung sound classification models are not only a promising solution over traditional methods but can also achieve a significant performance boost upon augmenting an imbalanced training set.
Collapse
Affiliation(s)
- Jane Saldanha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Shaunak Chakraborty
- Dept. of Computer Science, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Shruti Patil
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
- * E-mail:
| | - Ketan Kotecha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Satish Kumar
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, Symbiosis International (Deemed University), Pune, Maharashtra, India
| | - Anand Nayyar
- Graduate School (Computer Science), Faculty of Information Technology, Duy Tan University, Da Nang, Vietnam
| |
Collapse
|
17
|
Neili Z, Sundaraj K. A comparative study of the spectrogram, scalogram, melspectrogram and gammatonegram time-frequency representations for the classification of lung sounds using the ICBHI database based on CNNs. BIOMED ENG-BIOMED TE 2022; 67:367-390. [PMID: 35926850 DOI: 10.1515/bmt-2022-0180] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2022] [Accepted: 06/21/2022] [Indexed: 11/15/2022]
Abstract
In lung sound classification using deep learning, many studies have considered the use of short-time Fourier transform (STFT) as the most commonly used 2D representation of the input data. Consequently, STFT has been widely used as an analytical tool, but other versions of the representation have also been developed. This study aims to evaluate and compare the performance of the spectrogram, scalogram, melspectrogram and gammatonegram representations, and provide comparative information to users regarding the suitability of these time-frequency (TF) techniques in lung sound classification. Lung sound signals used in this study were obtained from the ICBHI 2017 respiratory sound database. These lung sound recordings were converted into images of spectrogram, scalogram, melspectrogram and gammatonegram TF representations respectively. The four types of images were fed separately into the VGG16, ResNet-50 and AlexNet deep-learning architectures. Network performances were analyzed and compared based on accuracy, precision, recall and F1-score. The results of the analysis on the performance of the four representations using these three commonly used CNN deep-learning networks indicate that the generated gammatonegram and scalogram TF images coupled with ResNet-50 achieved maximum classification accuracies.
Collapse
Affiliation(s)
- Zakaria Neili
- Electronics Department, University of Badji Mokhtar Annaba, Annaba, Algeria
| | - Kenneth Sundaraj
- Faculty of Electronics and Computer Engineering, Universiti Teknikal Malaysia Melaka, Melaka, Malaysia
| |
Collapse
|
18
|
Fang L, Wang X. COVID-RDNet: A novel coronavirus pneumonia classification model using the mixed dataset by CT and X-rays images. Biocybern Biomed Eng 2022; 42:977-994. [PMID: 35945982 PMCID: PMC9353669 DOI: 10.1016/j.bbe.2022.07.009] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2022] [Revised: 07/10/2022] [Accepted: 07/31/2022] [Indexed: 12/23/2022]
Abstract
Corona virus disease 2019 (COVID-19) testing relies on traditional screening methods, which require a lot of manpower and material resources. Recently, to effectively reduce the damage caused by radiation and enhance effectiveness, deep learning of classifying COVID-19 negative and positive using the mixed dataset by CT and X-rays images have achieved remarkable research results. However, the details presented on CT and X-ray images have pathological diversity and similarity features, thus increasing the difficulty for physicians to judge specific cases. On this basis, this paper proposes a novel coronavirus pneumonia classification model using the mixed dataset by CT and X-rays images. To solve the problem of feature similarity between lung diseases and COVID-19, the extracted features are enhanced by an adaptive region enhancement algorithm. Besides, the depth network based on the residual blocks and the dense blocks is trained and tested. On the one hand, the residual blocks effectively improve the accuracy of the model and the non-linear COVID-19 features are obtained by cross-layer link. On the other hand, the dense blocks effectively improve the robustness of the model by connecting local and abstract information. On mixed X-ray and CT datasets, the sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), area under curve (AUC), and accuracy can all reach 0.99. On the basis of respecting patient privacy and ethics, the proposed algorithm using the mixed dataset from real cases can effectively assist doctors in performing the accurate COVID-19 negative and positive classification to determine the infection status of patients.
Collapse
Affiliation(s)
- Lingling Fang
- Department of Computing and Information Technology, Liaoning Normal University, Dalian City, Liaoning Province, China
| | - Xin Wang
- Department of Computing and Information Technology, Liaoning Normal University, Dalian City, Liaoning Province, China
| |
Collapse
|
19
|
A Progressively Expanded Database for Automated Lung Sound Analysis: An Update. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12157623] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/01/2023]
Abstract
We previously established an open-access lung sound database, HF_Lung_V1, and developed deep learning models for inhalation, exhalation, continuous adventitious sound (CAS), and discontinuous adventitious sound (DAS) detection. The amount of data used for training contributes to model accuracy. In this study, we collected larger quantities of data to further improve model performance and explored issues of noisy labels and overlapping sounds. HF_Lung_V1 was expanded to HF_Lung_V2 with a 1.43× increase in the number of audio files. Convolutional neural network–bidirectional gated recurrent unit network models were trained separately using the HF_Lung_V1 (V1_Train) and HF_Lung_V2 (V2_Train) training sets. These were tested using the HF_Lung_V1 (V1_Test) and HF_Lung_V2 (V2_Test) test sets, respectively. Segment and event detection performance was evaluated. Label quality was assessed. Overlap ratios were computed between inhalation, exhalation, CAS, and DAS labels. The model trained using V2_Train exhibited improved performance in inhalation, exhalation, CAS, and DAS detection on both V1_Test and V2_Test. Poor CAS detection was attributed to the quality of CAS labels. DAS detection was strongly influenced by the overlapping of DAS with inhalation and exhalation. In conclusion, collecting greater quantities of lung sound data is vital for developing more accurate lung sound analysis models.
Collapse
|
20
|
Pham Thi Viet H, Nguyen Thi Ngoc H, Tran Anh V, Hoang Quang H. Classification of lung sounds using scalogram representation of sound segments and convolutional neural network. J Med Eng Technol 2022; 46:270-279. [PMID: 35212591 DOI: 10.1080/03091902.2022.2040624] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2023]
Abstract
Lung auscultation is one of the most common methods for screening of lung diseases. The increasingly high rate of respiratory diseases leads to the need for robust methods to detect the abnormalities in patients' breathing sounds. Lung sounds analysis stands out as a promising approach to automatic screening of lung diseases, serving as a second opinion for doctors as a stand-alone device for preliminary screening of lung diseases in remote areas. In previous research on lung classification using ICBHI Database on Kaggle, lung audios are converted to spectral images and fed into deep neural networks for training. There are a few studies which uses the scalogram, however they focussed on classification among different lung diseases. The use of scalograms in categorising the sound types are rarely used. In this paper, we combined scalograms and neural networks for classification of lung sound types. Padding methods and augmentation are also considered to evaluate the impacts on classification score. An ensemble learning is incorporated to increase classification accuracy by utilising voting of many models. The model trained and evaluated has shown prominent improvement of this method on classification on the benchmark ICBHI database.
Collapse
Affiliation(s)
| | - Huyen Nguyen Thi Ngoc
- School of Electronics and Telecommunications, Hanoi University of Science and Technology, Hanoi, Vietnam
| | - Vu Tran Anh
- School of Electronics and Telecommunications, Hanoi University of Science and Technology, Hanoi, Vietnam
| | - Huy Hoang Quang
- School of Electronics and Telecommunications, Hanoi University of Science and Technology, Hanoi, Vietnam
| |
Collapse
|
21
|
Shamrat FMJM, Azam S, Karim A, Islam R, Tasnim Z, Ghosh P, De Boer F. LungNet22: A Fine-Tuned Model for Multiclass Classification and Prediction of Lung Disease Using X-ray Images. J Pers Med 2022; 12:jpm12050680. [PMID: 35629103 PMCID: PMC9143659 DOI: 10.3390/jpm12050680] [Citation(s) in RCA: 19] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2022] [Revised: 04/01/2022] [Accepted: 04/20/2022] [Indexed: 12/29/2022] Open
Abstract
In recent years, lung disease has increased manyfold, causing millions of casualties annually. To combat the crisis, an efficient, reliable, and affordable lung disease diagnosis technique has become indispensable. In this study, a multiclass classification of lung disease from frontal chest X-ray imaging using a fine-tuned CNN model is proposed. The classification is conducted on 10 disease classes of the lungs, namely COVID-19, Effusion, Tuberculosis, Pneumonia, Lung Opacity, Mass, Nodule, Pneumothorax, and Pulmonary Fibrosis, along with the Normal class. The dataset is a collective dataset gathered from multiple sources. After pre-processing and balancing the dataset with eight augmentation techniques, a total of 80,000 X-ray images were fed to the model for classification purposes. Initially, eight pre-trained CNN models, AlexNet, GoogLeNet, InceptionV3, MobileNetV2, VGG16, ResNet 50, DenseNet121, and EfficientNetB7, were employed on the dataset. Among these, the VGG16 achieved the highest accuracy at 92.95%. To further improve the classification accuracy, LungNet22 was constructed upon the primary structure of the VGG16 model. An ablation study was used in the work to determine the different hyper-parameters. Using the Adam Optimizer, the proposed model achieved a commendable accuracy of 98.89%. To verify the performance of the model, several performance matrices, including the ROC curve and the AUC values, were computed as well.
Collapse
Affiliation(s)
- F. M. Javed Mehedi Shamrat
- Department of Software Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (F.M.J.M.S.); (Z.T.)
| | - Sami Azam
- College of Engineering, IT and Environment, Charles Darwin University, Casuarina, NT 0909, Australia; (A.K.); (F.D.B.)
- Correspondence:
| | - Asif Karim
- College of Engineering, IT and Environment, Charles Darwin University, Casuarina, NT 0909, Australia; (A.K.); (F.D.B.)
| | - Rakibul Islam
- Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh;
| | - Zarrin Tasnim
- Department of Software Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (F.M.J.M.S.); (Z.T.)
| | - Pronab Ghosh
- Department of Computer Science (CS), Lakehead University, 955 Oliver Rd, Thunder Bay, ON P7B 5E1, Canada;
| | - Friso De Boer
- College of Engineering, IT and Environment, Charles Darwin University, Casuarina, NT 0909, Australia; (A.K.); (F.D.B.)
| |
Collapse
|
22
|
Hassan Naqvi SZ, Choudhry MA. Embedded system design for classification of COPD and pneumonia patients by lung sound analysis. BIOMED ENG-BIOMED TE 2022; 67:201-218. [PMID: 35405045 DOI: 10.1515/bmt-2022-0011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2022] [Accepted: 03/17/2022] [Indexed: 11/15/2022]
Abstract
Chronic obstructive pulmonary disease (COPD) and pneumonia are lethal pulmonary illnesses with equivocal nature of abnormal pulmonic acoustics. Using lung sound signals, the classification of pulmonary abnormalities is a difficult task. A standalone system was conceived for screening COPD and Pneumonia patients through signal processing and machine learning methodologies. The proposed system will assist practitioners and pulmonologists in the accurate classification of disease. In this research work, ICBHI's and self-collected lung sound (LS) databases are used to investigate COPD and pneumonia patient. In this scheme, empirical mode decomposition (EMD), discrete wavelet transform (DWT), and analysis of variance (ANOVA) techniques are employed for segmentation, noise elimination, and feature selection, respectively. To overcome the inherent limitation of ICBHI's LS database, the adaptive synthetic (ADASYN) sampling technique is used to eradicate class imbalance. Lung sound features are used to train fine Gaussian support vector machine (FG-SVM) for classification of COPD, pneumonia, and heathy healthy subjects. This machine learning scheme is implemented on low cost and portable Raspberry pi 3 model B+ (Cortex-A53 (ARMv8) 64-bit SoC @ 1.4 GHz through hardware-supported language. Resultant hardware is capable of screening COPD and pneumonia patients accurately and assist health professionals.
Collapse
Affiliation(s)
- Syed Zohaib Hassan Naqvi
- Department of Electronics Engineering, University of Engineering and Technology Taxila, Taxila, Pakistan
| | - Mohmmad Ahmad Choudhry
- Department of Electrical Engineering, University of Engineering and Technology Taxila, Taxila, Pakistan
| |
Collapse
|
23
|
Abera Tessema B, Nemomssa HD, Lamesgin Simegn G. Acquisition and Classification of Lung Sounds for Improving the Efficacy of Auscultation Diagnosis of Pulmonary Diseases. MEDICAL DEVICES-EVIDENCE AND RESEARCH 2022; 15:89-102. [PMID: 35418786 PMCID: PMC9000552 DOI: 10.2147/mder.s362407] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2022] [Accepted: 03/25/2022] [Indexed: 11/30/2022] Open
Abstract
Purpose Lung diseases are the third leading cause of death worldwide. Stethoscope-based auscultation is the most commonly used, non-invasive, inexpensive, and primary diagnostic approach for assessing lung conditions. However, the manual auscultation-based diagnosis procedure is prone to error, and its accuracy is dependent on the physician’s experience and hearing capacity. Moreover, the stethoscope recording is vulnerable to different noises that can mask the important features of lung sounds which may lead to misdiagnosis. In this paper, a method for the acquisition of lung sound signals and classification of the top 7 lung diseases has been proposed for improving the efficacy of auscultation diagnosis of pulmonary disease. Methods An electronic stethoscope has been constructed for signal acquisition. Lung sound signals were then collected from people with COPD, upper respiratory tract infections (URTI), lower respiratory tract infections (LRTI), pneumonia, bronchiectasis, bronchiolitis, asthma, and healthy people. Lung sounds were analyzed using a wavelet multiresolution analysis. To choose the most relevant features, feature selection using one-way ANOVA was performed. The classification accuracy of various machine learning classifiers was compared, and the Fine Gaussian SVM was chosen for final classification due to its superior performance. Model optimization was accomplished through the application of Bayesian optimization techniques. Results A test classification accuracy of 99%, specificity of 99.2%, and sensitivity of 99.04%, have been achieved for the 7 lung diseases using the optimized Fine Gaussian SVM classifier. Conclusion Our experimental results demonstrate that the proposed method has the potential to be used as a decision support system for the classification of lung diseases, especially in those areas where the expertise and the means are limited.
Collapse
Affiliation(s)
- Biruk Abera Tessema
- School of Biomedical Engineering, Jimma Institute of Technology, Jimma University, Jimma, Ethiopia
- School of Medicine, Haramaya University College of Health and Medical Sciences, Haramaya University, Harar, Ethiopia
| | - Hundessa Daba Nemomssa
- School of Biomedical Engineering, Jimma Institute of Technology, Jimma University, Jimma, Ethiopia
- Correspondence: Hundessa Daba Nemomssa, Tel +251913763777, Email
| | | |
Collapse
|
24
|
Nguyen T, Pernkopf F. Lung Sound Classification Using Co-tuning and Stochastic Normalization. IEEE Trans Biomed Eng 2022; 69:2872-2882. [PMID: 35254969 DOI: 10.1109/tbme.2022.3156293] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Computational methods for lung sound analysis are beneficial for computer-aided diagnosis support, storage and monitoring in critical care. In this paper, we use pre-trained ResNet models as backbone architectures for classification of adventitious lung sounds and respiratory diseases. The learned representation of the pre-trained model is transferred by using vanilla fine-tuning, co-tuning, stochastic normalization and the combination of the co-tuning and stochastic normalization techniques. Furthermore, data augmentation in both time domain and time-frequency domain is used to account for the class imbalance of the ICBHI and our multi-channel lung sound dataset. Additionally, we introduce spectrum correction to account for the variations of the recording device properties on the ICBHI dataset. Empirically, our proposed systems mostly outperform all state-of-the-art lung sound classification systems for the adventitious lung sounds and respiratory diseases of both datasets.
Collapse
|
25
|
Ebbehoj A, Thunbo MØ, Andersen OE, Glindtvad MV, Hulman A. Transfer learning for non-image data in clinical research: A scoping review. PLOS DIGITAL HEALTH 2022; 1:e0000014. [PMID: 36812540 PMCID: PMC9931256 DOI: 10.1371/journal.pdig.0000014] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/05/2021] [Accepted: 12/15/2021] [Indexed: 01/14/2023]
Abstract
BACKGROUND Transfer learning is a form of machine learning where a pre-trained model trained on a specific task is reused as a starting point and tailored to another task in a different dataset. While transfer learning has garnered considerable attention in medical image analysis, its use for clinical non-image data is not well studied. Therefore, the objective of this scoping review was to explore the use of transfer learning for non-image data in the clinical literature. METHODS AND FINDINGS We systematically searched medical databases (PubMed, EMBASE, CINAHL) for peer-reviewed clinical studies that used transfer learning on human non-image data. We included 83 studies in the review. More than half of the studies (63%) were published within 12 months of the search. Transfer learning was most often applied to time series data (61%), followed by tabular data (18%), audio (12%) and text (8%). Thirty-three (40%) studies applied an image-based model to non-image data after transforming data into images (e.g. spectrograms). Twenty-nine (35%) studies did not have any authors with a health-related affiliation. Many studies used publicly available datasets (66%) and models (49%), but fewer shared their code (27%). CONCLUSIONS In this scoping review, we have described current trends in the use of transfer learning for non-image data in the clinical literature. We found that the use of transfer learning has grown rapidly within the last few years. We have identified studies and demonstrated the potential of transfer learning in clinical research in a wide range of medical specialties. More interdisciplinary collaborations and the wider adaption of reproducible research principles are needed to increase the impact of transfer learning in clinical research.
Collapse
Affiliation(s)
- Andreas Ebbehoj
- Department of Endocrinology and Internal Medicine, Aarhus University Hospital, Denmark
- Department of Clinical Medicine, Aarhus University, Denmark
| | | | | | | | - Adam Hulman
- Steno Diabetes Center Aarhus, Aarhus University Hospital, Denmark
- * E-mail:
| |
Collapse
|
26
|
Tariq Z, Shah SK, Lee Y. Feature-Based Fusion Using CNN for Lung and Heart Sound Classification. SENSORS 2022; 22:s22041521. [PMID: 35214424 PMCID: PMC8875944 DOI: 10.3390/s22041521] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/29/2021] [Revised: 02/10/2022] [Accepted: 02/14/2022] [Indexed: 02/04/2023]
Abstract
Lung or heart sound classification is challenging due to the complex nature of audio data, its dynamic properties of time, and frequency domains. It is also very difficult to detect lung or heart conditions with small amounts of data or unbalanced and high noise in data. Furthermore, the quality of data is a considerable pitfall for improving the performance of deep learning. In this paper, we propose a novel feature-based fusion network called FDC-FS for classifying heart and lung sounds. The FDC-FS framework aims to effectively transfer learning from three different deep neural network models built from audio datasets. The innovation of the proposed transfer learning relies on the transformation from audio data to image vectors and from three specific models to one fused model that would be more suitable for deep learning. We used two publicly available datasets for this study, i.e., lung sound data from ICHBI 2017 challenge and heart challenge data. We applied data augmentation techniques, such as noise distortion, pitch shift, and time stretching, dealing with some data issues in these datasets. Importantly, we extracted three unique features from the audio samples, i.e., Spectrogram, MFCC, and Chromagram. Finally, we built a fusion of three optimal convolutional neural network models by feeding the image feature vectors transformed from audio features. We confirmed the superiority of the proposed fusion model compared to the state-of-the-art works. The highest accuracy we achieved with FDC-FS is 99.1% with Spectrogram-based lung sound classification while 97% for Spectrogram and Chromagram based heart sound classification.
Collapse
|
27
|
Petmezas G, Cheimariotis GA, Stefanopoulos L, Rocha B, Paiva RP, Katsaggelos AK, Maglaveras N. Automated Lung Sound Classification Using a Hybrid CNN-LSTM Network and Focal Loss Function. SENSORS (BASEL, SWITZERLAND) 2022; 22:1232. [PMID: 35161977 PMCID: PMC8838187 DOI: 10.3390/s22031232] [Citation(s) in RCA: 21] [Impact Index Per Article: 10.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/07/2022] [Revised: 02/02/2022] [Accepted: 02/03/2022] [Indexed: 11/16/2022]
Abstract
Respiratory diseases constitute one of the leading causes of death worldwide and directly affect the patient's quality of life. Early diagnosis and patient monitoring, which conventionally include lung auscultation, are essential for the efficient management of respiratory diseases. Manual lung sound interpretation is a subjective and time-consuming process that requires high medical expertise. The capabilities that deep learning offers could be exploited in order that robust lung sound classification models can be designed. In this paper, we propose a novel hybrid neural model that implements the focal loss (FL) function to deal with training data imbalance. Features initially extracted from short-time Fourier transform (STFT) spectrograms via a convolutional neural network (CNN) are given as input to a long short-term memory (LSTM) network that memorizes the temporal dependencies between data and classifies four types of lung sounds, including normal, crackles, wheezes, and both crackles and wheezes. The model was trained and tested on the ICBHI 2017 Respiratory Sound Database and achieved state-of-the-art results using three different data splitting strategies-namely, sensitivity 47.37%, specificity 82.46%, score 64.92% and accuracy 73.69% for the official 60/40 split, sensitivity 52.78%, specificity 84.26%, score 68.52% and accuracy 76.39% using interpatient 10-fold cross validation, and sensitivity 60.29% and accuracy 74.57% using leave-one-out cross validation.
Collapse
Affiliation(s)
- Georgios Petmezas
- Laboratory of Computing, Medical Informatics and Biomedical—Imaging Technologies, Medical School, Aristotle University of Thessaloniki, GR 54124 Thessaloniki, Greece; (G.P.); (G.-A.C.); (L.S.)
| | - Grigorios-Aris Cheimariotis
- Laboratory of Computing, Medical Informatics and Biomedical—Imaging Technologies, Medical School, Aristotle University of Thessaloniki, GR 54124 Thessaloniki, Greece; (G.P.); (G.-A.C.); (L.S.)
| | - Leandros Stefanopoulos
- Laboratory of Computing, Medical Informatics and Biomedical—Imaging Technologies, Medical School, Aristotle University of Thessaloniki, GR 54124 Thessaloniki, Greece; (G.P.); (G.-A.C.); (L.S.)
| | - Bruno Rocha
- Centre for Informatics and Systems, Department of Informatics Engineering, University of Coimbra, 3030-290 Coimbra, Portugal; (B.R.); (R.P.P.)
| | - Rui Pedro Paiva
- Centre for Informatics and Systems, Department of Informatics Engineering, University of Coimbra, 3030-290 Coimbra, Portugal; (B.R.); (R.P.P.)
| | - Aggelos K. Katsaggelos
- Department of Electrical and Computer Engineering, Northwestern University, Evanston, IL 60208, USA;
| | - Nicos Maglaveras
- Laboratory of Computing, Medical Informatics and Biomedical—Imaging Technologies, Medical School, Aristotle University of Thessaloniki, GR 54124 Thessaloniki, Greece; (G.P.); (G.-A.C.); (L.S.)
| |
Collapse
|
28
|
Demir F, Demir K, Şengür A. DeepCov19Net: Automated COVID-19 Disease Detection with a Robust and Effective Technique Deep Learning Approach. NEW GENERATION COMPUTING 2022; 40:1053-1075. [PMID: 35035024 PMCID: PMC8753945 DOI: 10.1007/s00354-021-00152-0] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/15/2021] [Accepted: 12/26/2021] [Indexed: 05/17/2023]
Abstract
The new type of coronavirus disease, which has spread from Wuhan, China since the beginning of 2020 called COVID-19, has caused many deaths and cases in most countries and has reached a global pandemic scale. In addition to test kits, imaging techniques with X-rays used in lung patients have been frequently used in the detection of COVID-19 cases. In the proposed method, a novel approach based on a deep learning model named DeepCovNet was utilized to classify chest X-ray images containing COVID-19, normal (healthy), and pneumonia classes. The convolutional-autoencoder model, which had convolutional layers in encoder and decoder blocks, was trained by using the processed chest X-ray images from scratch for deep feature extraction. The distinctive features were selected with a novel and robust algorithm named SDAR from the deep feature set. In the classification stage, an SVM classifier with various kernel functions was used to evaluate the classification performance of the proposed method. Also, hyperparameters of the SVM classifier were optimized with the Bayesian algorithm for increasing classification accuracy. Specificity, sensitivity, precision, and F-score, were also used as performance metrics in addition to accuracy which was used as the main criterion. The proposed method with an accuracy of 99.75 outperformed the other approaches based on deep learning.
Collapse
Affiliation(s)
- Fatih Demir
- Biomedical Department, Vocational School of Technical Sciences, Firat University, Elazig, Turkey
| | - Kürşat Demir
- Mechatronics Engineering Department, Technology Faculty, Firat University, Elazig, Turkey
| | - Abdulkadir Şengür
- Electrical-Electronics Engineering Department, Technology Faculty, Firat University, Elazig, Turkey
| |
Collapse
|
29
|
Fernando T, Sridharan S, Denman S, Ghaemmaghami H, Fookes C. Robust and Interpretable Temporal Convolution Network for Event Detection in Lung Sound Recordings. IEEE J Biomed Health Inform 2022; 26:2898-2908. [DOI: 10.1109/jbhi.2022.3144314] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
30
|
Zulfiqar R, Majeed F, Irfan R, Rauf HT, Benkhelifa E, Belkacem AN. Abnormal Respiratory Sounds Classification Using Deep CNN Through Artificial Noise Addition. Front Med (Lausanne) 2021; 8:714811. [PMID: 34869413 PMCID: PMC8635523 DOI: 10.3389/fmed.2021.714811] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2021] [Accepted: 10/07/2021] [Indexed: 11/29/2022] Open
Abstract
Respiratory sound (RS) attributes and their analyses structure a fundamental piece of pneumonic pathology, and it gives symptomatic data regarding a patient's lung. A couple of decades back, doctors depended on their hearing to distinguish symptomatic signs in lung audios by utilizing the typical stethoscope, which is usually considered a cheap and secure method for examining the patients. Lung disease is the third most ordinary cause of death worldwide, so; it is essential to classify the RS abnormality accurately to overcome the death rate. In this research, we have applied Fourier analysis for the visual inspection of abnormal respiratory sounds. Spectrum analysis was done through Artificial Noise Addition (ANA) in conjunction with different deep convolutional neural networks (CNN) to classify the seven abnormal respiratory sounds—both continuous (CAS) and discontinuous (DAS). The proposed framework contains an adaptive mechanism of adding a similar type of noise to unhealthy respiratory sounds. ANA makes sound features enough reach to be identified more accurately than the respiratory sounds without ANA. The obtained results using the proposed framework are superior to previous techniques since we simultaneously considered the seven different abnormal respiratory sound classes.
Collapse
Affiliation(s)
- Rizwana Zulfiqar
- Faculty of Computing and Information Technology, University of Gujrat, Gujrat, Pakistan
| | - Fiaz Majeed
- Faculty of Computing and Information Technology, University of Gujrat, Gujrat, Pakistan
| | - Rizwana Irfan
- Department of Information Technology, College of Computing and Information Technology at Khulais, University of Jeddah, Jeddah, Saudi Arabia
| | - Hafiz Tayyab Rauf
- Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, Stoke-on-Trent, United Kingdom
| | - Elhadj Benkhelifa
- Cloud Computing and Applications Reseach Lab, Staffordshire University, Stoke-on-Trent, United Kingdom
| | - Abdelkader Nasreddine Belkacem
- Department of Computer and Network Engineering, College of Information Technology, UAE University, Al Ain, United Arab Emirates
| |
Collapse
|
31
|
Mukherjee H, Salam H, Santosh KC. Lung Health Analysis: Adventitious Respiratory Sound Classification Using Filterbank Energies. INT J PATTERN RECOGN 2021. [DOI: 10.1142/s0218001421570081] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Audio-based healthcare technologies are among the most significant applications of pattern recognition and Artificial Intelligence. Lately, a major chunk of the World population has been infected with serious respiratory diseases such as COVID-19. Early recognition of lung health abnormalities can facilitate early intervention, and decrease the mortality rate of the infected population. Research has shown that it is possible to automatically monitor lung health abnormalities through respiratory sounds. In this paper, we propose an approach that employs filter bank energy-based features and Random Forests to classify lung problem types from respiratory sounds. The adventitious sounds, crackles and wheezes appear distinct to the human ear. Moreover, different sounds are characterized by different frequency ranges that are dominant. The proposed approach attempts to distinguish the adventitious sounds (crackles and wheezes) by modeling the human auditory perception of these sounds. Specifically, we propose a respiratory sounds representation technique capable of modeling the dominant frequency range present in such sounds. On a publicly available dataset (ICBHI) of size 6898 cycles spanning over 5[Formula: see text]h, our results can be compared with the state-of-the-art results, in distinguishing two different types of adventitious sounds: crackles and wheezes.
Collapse
Affiliation(s)
- Himadri Mukherjee
- SMART Lab, Department of Computer Science, New York University, Abu Dhabi, UAE
| | - Hanan Salam
- SMART Lab, Department of Computer Science, New York University, Abu Dhabi, UAE
| | - KC Santosh
- KC’s Pattern Analysis & Machine Learning (PAMI), Research Lab – Computer Science, University of South Dakota, Vermillion, SD 57069, USA
| |
Collapse
|
32
|
Do QT, Lipatov K, Wang HY, Pickering BW, Herasevich V. Classification of Respiratory Conditions using Auscultation Sound. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:1942-1945. [PMID: 34891667 DOI: 10.1109/embc46164.2021.9630294] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Management of respiratory conditions relies on timely diagnosis and institution of appropriate management. Computerized analysis and classification of breath sounds has a potential to enhance reliability and accuracy of diagnostic modality while making it suitable for remote monitoring, personalized uses, and self-management uses. In this paper, we describe and compare sound recognition models aimed at automatic diagnostic differentiation of healthy persons vs patients with COPD vs patients with pneumonia using deep learning approaches such as Multi-layer Perceptron Classifier (MLPClassifier) and Convolutional Neural Networks (CNN).Clinical Relevance-Healthcare providers and researchers interested in the field of medical sound analysis, specifically automatic detection/classification of auscultation sound and early diagnosis of respiratory conditions may benefit from this paper.
Collapse
|
33
|
Nguyen T, Pernkopf F. Crackle Detection In Lung Sounds Using Transfer Learning And Multi-Input Convolutional Neural Networks. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:80-83. [PMID: 34891244 DOI: 10.1109/embc46164.2021.9630577] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Large annotated lung sound databases are publicly available and might be used to train algorithms for diagnosis systems. However, it might be a challenge to develop a well-performing algorithm for small non-public data, which have only a few subjects and show differences in recording devices and setup. In this paper, we use transfer learning to tackle the mismatch of the recording setup. This allows us to transfer knowledge from one dataset to another dataset for crackle detection in lung sounds. In particular, a single input convolutional neural network (CNN) model is pre-trained on a source domain using ICBHI 2017, the largest publicly available database of lung sounds. We use log-mel spectrogram features of respiratory cycles of lung sounds. The pre-trained network is used to build a multi-input CNN model, which shares the same network architecture for respiratory cycles and their corresponding respiratory phases. The multi-input model is then fine-tuned on the target domain of our self-collected lung sound database for classifying crackles and normal lung sounds. Our experimental results show significant performance improvements of 9.84% (absolute) in F-score on the target domain using the multi-input CNN model and transfer learning for crackle detection.Clinical relevance- Crackle detection in lung sounds, multi-input convolutional neural networks, transfer learning.
Collapse
|
34
|
Toğaçar M. Disease type detection in lung and colon cancer images using the complement approach of inefficient sets. Comput Biol Med 2021; 137:104827. [PMID: 34560401 DOI: 10.1016/j.compbiomed.2021.104827] [Citation(s) in RCA: 28] [Impact Index Per Article: 9.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2021] [Revised: 08/25/2021] [Accepted: 08/29/2021] [Indexed: 12/19/2022]
Abstract
Lung and colon cancers are deadly diseases that can develop simultaneously in organs and adversely affect human life in some special cases. Although the frequency of simultaneous occurrence of these two types of cancer is unlikely, there is a high probability of metastasis between the two organs if not diagnosed early. Traditionally, specialists have to go through a lengthy and complicated process to examine histopathological images and diagnose cancer cases; yet, it is now possible to achieve this process faster with the available technological possibilities. In this study, artificial intelligence-supported model and optimization methods were used to realize the classification of lung and colon cancers' histopathological images. The used dataset has five classes of histopathological images consisting of two colon cancer classes and three lung cancer classes. In the proposed approach, the image classes were trained from scratch with the DarkNet-19 model, which is one of the deep learning models. In the feature set extracted from the DarkNet-19 model, selection of the inefficient features was performed by using Equilibrium and Manta Ray Foraging optimization algorithms. Then, the set containing the inefficient features was distinguished from the rest of the set features, creating an efficient feature set (complementary rule insets). The efficient features obtained by the two used optimization algorithms were combined and classified with the Support Vector Machine (SVM) method. The overall accuracy rate obtained in the classification process was 99.69%. Based on the outcomes of this study, it has been observed that using the complementary method together with some optimization methods improved the classification performance of the dataset.
Collapse
Affiliation(s)
- Mesut Toğaçar
- Department of Computer Technology, Technical Sciences Vocational School, Fırat UniversityElazig, Turkey.
| |
Collapse
|
35
|
Accurate diagnosis of lung tissues for 2D Raman spectrogram by deep learning based on short-time Fourier transform. Anal Chim Acta 2021; 1179:338821. [PMID: 34535256 DOI: 10.1016/j.aca.2021.338821] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2021] [Revised: 06/29/2021] [Accepted: 06/30/2021] [Indexed: 02/06/2023]
Abstract
Multivariate statistical analysis methods have an important role in spectrochemical analyses to rapidly identify and diagnose cancer and the subtype. However, utilizing these methods to analyze lager amount spectral data is challenging, and poses a major bottleneck toward achieving high accuracy. Here, a new convolutional neural networks (CNN) method based on short-time Fourier transform (STFT) to diagnose lung tissues via Raman spectra readily is proposed. The models yield that the accuracies of the new method are higher than the conventional methods (principal components analysis -linear discriminant analysis and support vector machine) for validation group (95.2% vs 85.5%, 94.4%) and test group (96.5% vs 90.4%, 93.9%) after cross-validation. The results illustrate that the new method which converts one-dimensional Raman data into two-dimensional Raman spectrograms improve the discriminatory ability of lung tissues and can achieve automatically accurate diagnosis of lung tissues.
Collapse
|
36
|
Maurya R, Pathak VK, Burget R, Dutta MK. Automated detection of bioimages using novel deep feature fusion algorithm and effective high-dimensional feature selection approach. Comput Biol Med 2021; 137:104862. [PMID: 34534793 DOI: 10.1016/j.compbiomed.2021.104862] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2021] [Revised: 08/26/2021] [Accepted: 09/07/2021] [Indexed: 11/30/2022]
Abstract
The classification of bioimages plays an important role in several biological studies, such as subcellular localisation, phenotype identification and other types of histopathological examinations. The objective of the present study was to develop a computer-aided bioimage classification method for the classification of bioimages across nine diverse benchmark datasets. A novel algorithm was developed, which systematically fused the features extracted from nine different convolution neural network architectures. A systematic fusion of features boosts the performance of a classifier but at the cost of the high dimensionality of the fused feature set. Therefore, non-discriminatory and redundant features need to be removed from a high-dimensional fused feature set to improve the classification performance and reduce the time complexity. To achieve this aim, a method based on analysis of variance and evolutionary feature selection was developed to select an optimal set of discriminatory features from the fused feature set. The proposed method was evaluated on nine different benchmark datasets. The experimental results showed that the proposed method achieved superior performance, with a significant reduction in the dimensionality of the fused feature set for most bioimage datasets. The performance of the proposed feature selection method was better than that of some of the most recent and classical methods used for feature selection. Thus, the proposed method was desirable because of its superior performance and high compression ratio, which significantly reduced the computational complexity.
Collapse
Affiliation(s)
- Ritesh Maurya
- Centre for Advanced Studies, Dr A.P.J. Abdul Kalam Technical University, Lucknow, India.
| | | | - Radim Burget
- Department of Telecommunications, Faculty of Electrical Engineering and Communication, BRNO University of Technology, Czech Republic.
| | - Malay Kishore Dutta
- Centre for Advanced Studies, Dr A.P.J. Abdul Kalam Technical University, Lucknow, India.
| |
Collapse
|
37
|
Kim Y, Hyon Y, Jung SS, Lee S, Yoo G, Chung C, Ha T. Respiratory sound classification for crackles, wheezes, and rhonchi in the clinical field using deep learning. Sci Rep 2021; 11:17186. [PMID: 34433880 PMCID: PMC8387488 DOI: 10.1038/s41598-021-96724-7] [Citation(s) in RCA: 28] [Impact Index Per Article: 9.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2021] [Accepted: 08/12/2021] [Indexed: 11/09/2022] Open
Abstract
Auscultation has been essential part of the physical examination; this is non-invasive, real-time, and very informative. Detection of abnormal respiratory sounds with a stethoscope is important in diagnosing respiratory diseases and providing first aid. However, accurate interpretation of respiratory sounds requires clinician's considerable expertise, so trainees such as interns and residents sometimes misidentify respiratory sounds. To overcome such limitations, we tried to develop an automated classification of breath sounds. We utilized deep learning convolutional neural network (CNN) to categorize 1918 respiratory sounds (normal, crackles, wheezes, rhonchi) recorded in the clinical setting. We developed the predictive model for respiratory sound classification combining pretrained image feature extractor of series, respiratory sound, and CNN classifier. It detected abnormal sounds with an accuracy of 86.5% and the area under the ROC curve (AUC) of 0.93. It further classified abnormal lung sounds into crackles, wheezes, or rhonchi with an overall accuracy of 85.7% and a mean AUC of 0.92. On the other hand, as a result of respiratory sound classification by different groups showed varying degree in terms of accuracy; the overall accuracies were 60.3% for medical students, 53.4% for interns, 68.8% for residents, and 80.1% for fellows. Our deep learning-based classification would be able to complement the inaccuracies of clinicians' auscultation, and it may aid in the rapid diagnosis and appropriate treatment of respiratory diseases.
Collapse
Affiliation(s)
- Yoonjoo Kim
- Division of Pulmonology and Critical Care Medicine, Department of Internal Medicine, College of Medicine, Chungnam National University, Daejeon, 34134, Republic of Korea
| | - YunKyong Hyon
- Division of Medical Mathematics, National Institute for Mathematical Sciences, Daejeon, 34047, Republic of Korea
| | - Sung Soo Jung
- Division of Pulmonology and Critical Care Medicine, Department of Internal Medicine, College of Medicine, Chungnam National University, Daejeon, 34134, Republic of Korea
| | - Sunju Lee
- Division of Medical Mathematics, National Institute for Mathematical Sciences, Daejeon, 34047, Republic of Korea
| | - Geon Yoo
- Clinical Research Division, National Institute of Food and Drug Safety Evaluation, Cheongju-si, Chungcheongbuk-do, Republic of Korea
| | - Chaeuk Chung
- Division of Pulmonology and Critical Care Medicine, Department of Internal Medicine, College of Medicine, Chungnam National University, Daejeon, 34134, Republic of Korea. .,Infection Control Convergence Research Center, Chungnam National University School of Medicine, Daejeon, 35015, Republic of Korea.
| | - Taeyoung Ha
- Division of Medical Mathematics, National Institute for Mathematical Sciences, Daejeon, 34047, Republic of Korea.
| |
Collapse
|
38
|
Shuvo SB, Ali SN, Swapnil SI, Hasan T, Bhuiyan MIH. A Lightweight CNN Model for Detecting Respiratory Diseases From Lung Auscultation Sounds Using EMD-CWT-Based Hybrid Scalogram. IEEE J Biomed Health Inform 2021; 25:2595-2603. [PMID: 33373309 DOI: 10.1109/jbhi.2020.3048006] [Citation(s) in RCA: 33] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Listening to lung sounds through auscultation is vital in examining the respiratory system for abnormalities. Automated analysis of lung auscultation sounds can be beneficial to the health systems in low-resource settings where there is a lack of skilled physicians. In this work, we propose a lightweight convolutional neural network (CNN) architecture to classify respiratory diseases from individual breath cycles using hybrid scalogram-based features of lung sounds. The proposed feature-set utilizes the empirical mode decomposition (EMD) and the continuous wavelet transform (CWT). The performance of the proposed scheme is studied using a patient independent train-validation-test set from the publicly available ICBHI 2017 lung sound dataset. Employing the proposed framework, weighted accuracy scores of 98.92% for three-class chronic classification and 98.70% for six-class pathological classification are achieved, which outperform well-known and much larger VGG16 in terms of accuracy by absolute margins of 1.10% and 1.11%, respectively. The proposed CNN model also outperforms other contemporary lightweight models while being computationally comparable.
Collapse
|
39
|
Hsu FS, Huang SR, Huang CW, Huang CJ, Cheng YR, Chen CC, Hsiao J, Chen CW, Chen LC, Lai YC, Hsu BF, Lin NJ, Tsai WL, Wu YL, Tseng TL, Tseng CT, Chen YT, Lai F. Benchmarking of eight recurrent neural network variants for breath phase and adventitious sound detection on a self-developed open-access lung sound database-HF_Lung_V1. PLoS One 2021; 16:e0254134. [PMID: 34197556 PMCID: PMC8248710 DOI: 10.1371/journal.pone.0254134] [Citation(s) in RCA: 22] [Impact Index Per Article: 7.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2021] [Accepted: 06/20/2021] [Indexed: 01/15/2023] Open
Abstract
A reliable, remote, and continuous real-time respiratory sound monitor with automated respiratory sound analysis ability is urgently required in many clinical scenarios-such as in monitoring disease progression of coronavirus disease 2019-to replace conventional auscultation with a handheld stethoscope. However, a robust computerized respiratory sound analysis algorithm for breath phase detection and adventitious sound detection at the recording level has not yet been validated in practical applications. In this study, we developed a lung sound database (HF_Lung_V1) comprising 9,765 audio files of lung sounds (duration of 15 s each), 34,095 inhalation labels, 18,349 exhalation labels, 13,883 continuous adventitious sound (CAS) labels (comprising 8,457 wheeze labels, 686 stridor labels, and 4,740 rhonchus labels), and 15,606 discontinuous adventitious sound labels (all crackles). We conducted benchmark tests using long short-term memory (LSTM), gated recurrent unit (GRU), bidirectional LSTM (BiLSTM), bidirectional GRU (BiGRU), convolutional neural network (CNN)-LSTM, CNN-GRU, CNN-BiLSTM, and CNN-BiGRU models for breath phase detection and adventitious sound detection. We also conducted a performance comparison between the LSTM-based and GRU-based models, between unidirectional and bidirectional models, and between models with and without a CNN. The results revealed that these models exhibited adequate performance in lung sound analysis. The GRU-based models outperformed, in terms of F1 scores and areas under the receiver operating characteristic curves, the LSTM-based models in most of the defined tasks. Furthermore, all bidirectional models outperformed their unidirectional counterparts. Finally, the addition of a CNN improved the accuracy of lung sound analysis, especially in the CAS detection tasks.
Collapse
Affiliation(s)
- Fu-Shun Hsu
- Graduate Institute of Biomedical Electronics and Bioinformatics, National Taiwan University, Taipei, Taiwan
- Department of Critical Care Medicine, Far Eastern Memorial Hospital, New Taipei, Taiwan
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
| | | | | | - Chao-Jung Huang
- Joint Research Center for Artificial Intelligence Technology and All Vista Healthcare, National Taiwan University, Taipei, Taiwan
| | - Yuan-Ren Cheng
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
- Department of Life Science, College of Life Science, National Taiwan University, Taipei, Taiwan
- Institute of Biomedical Sciences, Academia Sinica, Taipei, Taiwan
| | | | - Jack Hsiao
- HCC Healthcare Group, New Taipei, Taiwan
| | - Chung-Wei Chen
- Department of Critical Care Medicine, Far Eastern Memorial Hospital, New Taipei, Taiwan
| | - Li-Chin Chen
- Research Center for Information Technology Innovation, Academia Sinica, Taipei, Taiwan
| | - Yen-Chun Lai
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
| | - Bi-Fang Hsu
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
| | - Nian-Jhen Lin
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
- Division of Pulmonary Medicine, Far Eastern Memorial Hospital, New Taipei, Taiwan
| | - Wan-Ling Tsai
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
| | - Yi-Lin Wu
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
| | | | | | - Yi-Tsun Chen
- Heroic Faith Medical Science Co., Ltd., Taipei, Taiwan
| | - Feipei Lai
- Graduate Institute of Biomedical Electronics and Bioinformatics, National Taiwan University, Taipei, Taiwan
| |
Collapse
|
40
|
WANG MONAN, LI DONGHUI, TANG LI. A LUNG IMAGE CLASSIFICATION METHOD: A CLASSIFIER CONSTRUCTED BY COMBINING IMPROVED VGG16 AND GRADIENT BOOSTING DECISION TREE. J MECH MED BIOL 2021. [DOI: 10.1142/s0219519421500421] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Early classification and diagnosis of lung diseases is essential to increase the best chance of patient recovery and survival. Using deep learning to make it possible, the key is how to improve the robustness of the deep learning model and the accuracy of lung image classification. In order to classify the five lung diseases, we used transfer learning to improve and fine-tune the fully connected layer of VGG16, and improve the cross entropy loss function, combined with the gradient boosting decision tree (GBDT), to establish a deep learning model called a classifier. The model was trained using the ChestX-ray14 dataset. On the test set, the classification accuracy of our model for the five lung diseases was 82.43%, 95.37%, 82.11%, 79.81%, 78.13%, which is better than the best published results. The F1 value is 0.456 (95% CI 0.415, 0.496). The robustness of the model exceeds CheXNet and average performance of doctors. This study clarified that the model has strong robustness and effectiveness in classifying five lung diseases.
Collapse
Affiliation(s)
- MONAN WANG
- Harbin University of Science and Technology, Harbin 150080, P. R. China
| | - DONGHUI LI
- Harbin University of Science and Technology, Harbin 150080, P. R. China
| | - LI TANG
- Harbin University of Science and Technology, Harbin 150080, P. R. China
| |
Collapse
|
41
|
Gayathri S, Gopi VP, Palanisamy P. Diabetic retinopathy classification based on multipath CNN and machine learning classifiers. Phys Eng Sci Med 2021; 44:639-653. [PMID: 34033015 DOI: 10.1007/s13246-021-01012-3] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2020] [Accepted: 05/06/2021] [Indexed: 11/28/2022]
Abstract
Eye care professionals generally use fundoscopy to confirm the occurrence of Diabetic Retinopathy (DR) in patients. Early DR detection and accurate DR grading are critical for the care and management of this disease. This work proposes an automated DR grading method in which features can be extracted from the fundus images and categorized based on severity using deep learning and Machine Learning (ML) algorithms. A Multipath Convolutional Neural Network (M-CNN) is used for global and local feature extraction from images. Then, a machine learning classifier is used to categorize the input according to the severity. The proposed model is evaluated across different publicly available databases (IDRiD, Kaggle (for DR detection), and MESSIDOR) and different ML classifiers (Support Vector Machine (SVM), Random Forest, and J48). The metrics selected for model evaluation are the False Positive Rate (FPR), Specificity, Precision, Recall, F1-score, K-score, and Accuracy. The experiments show that the best response is produced by the M-CNN network with the J48 classifier. The classifiers are evaluated across the pre-trained network features and existing DR grading methods. The average accuracy obtained for the proposed work is 99.62% for DR grading. The experiments and evaluation results show that the proposed method works well for accurate DR grading and early disease detection.
Collapse
Affiliation(s)
- S Gayathri
- National Institute of Technology, Tiruchirappalli, Tamil Nadu, India
| | - Varun P Gopi
- National Institute of Technology, Tiruchirappalli, Tamil Nadu, India.
| | - P Palanisamy
- National Institute of Technology, Tiruchirappalli, Tamil Nadu, India
| |
Collapse
|
42
|
Jung SY, Liao CH, Wu YS, Yuan SM, Sun CT. Efficiently Classifying Lung Sounds through Depthwise Separable CNN Models with Fused STFT and MFCC Features. Diagnostics (Basel) 2021; 11:732. [PMID: 33924146 PMCID: PMC8074359 DOI: 10.3390/diagnostics11040732] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2021] [Revised: 04/07/2021] [Accepted: 04/13/2021] [Indexed: 01/18/2023] Open
Abstract
Lung sounds remain vital in clinical diagnosis as they reveal associations with pulmonary pathologies. With COVID-19 spreading across the world, it has become more pressing for medical professionals to better leverage artificial intelligence for faster and more accurate lung auscultation. This research aims to propose a feature engineering process that extracts the dedicated features for the depthwise separable convolution neural network (DS-CNN) to classify lung sounds accurately and efficiently. We extracted a total of three features for the shrunk DS-CNN model: the short-time Fourier-transformed (STFT) feature, the Mel-frequency cepstrum coefficient (MFCC) feature, and the fused features of these two. We observed that while DS-CNN models trained on either the STFT or the MFCC feature achieved an accuracy of 82.27% and 73.02%, respectively, fusing both features led to a higher accuracy of 85.74%. In addition, our method achieved 16 times higher inference speed on an edge device and only 0.45% less accuracy than RespireNet. This finding indicates that the fusion of the STFT and MFCC features and DS-CNN would be a model design for lightweight edge devices to achieve accurate AI-aided detection of lung diseases.
Collapse
Affiliation(s)
- Shing-Yun Jung
- Department of Computer Science, National Chiao Tung University, Hsinchu 300, Taiwan; (C.-H.L.); (Y.-S.W.); (C.-T.S.)
| | - Chia-Hung Liao
- Department of Computer Science, National Chiao Tung University, Hsinchu 300, Taiwan; (C.-H.L.); (Y.-S.W.); (C.-T.S.)
| | - Yu-Sheng Wu
- Department of Computer Science, National Chiao Tung University, Hsinchu 300, Taiwan; (C.-H.L.); (Y.-S.W.); (C.-T.S.)
| | - Shyan-Ming Yuan
- Department of Computer Science, National Chiao Tung University, Hsinchu 300, Taiwan; (C.-H.L.); (Y.-S.W.); (C.-T.S.)
- Department of Computer Science, National Yang Ming Chiao Tung University, Hsinchu 300, Taiwan
| | - Chuen-Tsai Sun
- Department of Computer Science, National Chiao Tung University, Hsinchu 300, Taiwan; (C.-H.L.); (Y.-S.W.); (C.-T.S.)
- Department of Computer Science, National Yang Ming Chiao Tung University, Hsinchu 300, Taiwan
| |
Collapse
|
43
|
Albahli S, Rauf HT, Algosaibi A, Balas VE. AI-driven deep CNN approach for multi-label pathology classification using chest X-Rays. PeerJ Comput Sci 2021; 7:e495. [PMID: 33977135 PMCID: PMC8064140 DOI: 10.7717/peerj-cs.495] [Citation(s) in RCA: 26] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2021] [Accepted: 03/27/2021] [Indexed: 02/05/2023]
Abstract
Artificial intelligence (AI) has played a significant role in image analysis and feature extraction, applied to detect and diagnose a wide range of chest-related diseases. Although several researchers have used current state-of-the-art approaches and have produced impressive chest-related clinical outcomes, specific techniques may not contribute many advantages if one type of disease is detected without the rest being identified. Those who tried to identify multiple chest-related diseases were ineffective due to insufficient data and the available data not being balanced. This research provides a significant contribution to the healthcare industry and the research community by proposing a synthetic data augmentation in three deep Convolutional Neural Networks (CNNs) architectures for the detection of 14 chest-related diseases. The employed models are DenseNet121, InceptionResNetV2, and ResNet152V2; after training and validation, an average ROC-AUC score of 0.80 was obtained competitive as compared to the previous models that were trained for multi-class classification to detect anomalies in x-ray images. This research illustrates how the proposed model practices state-of-the-art deep neural networks to classify 14 chest-related diseases with better accuracy.
Collapse
Affiliation(s)
- Saleh Albahli
- Department of Information Technology, College of Computer Science, Qassim University, Buraydah, Saudi Arabia
| | - Hafiz Tayyab Rauf
- Centre for Smart Systems, AI and Cybersecurity, Staffordshire University, stoke on Trent, United Kingdom
| | | | - Valentina Emilia Balas
- Department of Automation and Applied Informatics, Aurel Vlaicu University of Arad, Arad, Romania
| |
Collapse
|
44
|
Automatic Classification of Adventitious Respiratory Sounds: A (Un)Solved Problem? SENSORS 2020; 21:s21010057. [PMID: 33374363 PMCID: PMC7795327 DOI: 10.3390/s21010057] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/10/2020] [Revised: 12/12/2020] [Accepted: 12/16/2020] [Indexed: 11/29/2022]
Abstract
(1) Background: Patients with respiratory conditions typically exhibit adventitious respiratory sounds (ARS), such as wheezes and crackles. ARS events have variable duration. In this work we studied the influence of event duration on automatic ARS classification, namely, how the creation of the Other class (negative class) affected the classifiers’ performance. (2) Methods: We conducted a set of experiments where we varied the durations of the other events on three tasks: crackle vs. wheeze vs. other (3 Class); crackle vs. other (2 Class Crackles); and wheeze vs. other (2 Class Wheezes). Four classifiers (linear discriminant analysis, support vector machines, boosted trees, and convolutional neural networks) were evaluated on those tasks using an open access respiratory sound database. (3) Results: While on the 3 Class task with fixed durations, the best classifier achieved an accuracy of 96.9%, the same classifier reached an accuracy of 81.8% on the more realistic 3 Class task with variable durations. (4) Conclusion: These results demonstrate the importance of experimental design on the assessment of the performance of automatic ARS classification algorithms. Furthermore, they also indicate, unlike what is stated in the literature, that the automatic classification of ARS is not a solved problem, as the algorithms’ performance decreases substantially under complex evaluation scenarios.
Collapse
|
45
|
Naqvi SZH, Choudhry MA. An Automated System for Classification of Chronic Obstructive Pulmonary Disease and Pneumonia Patients Using Lung Sound Analysis. SENSORS 2020; 20:s20226512. [PMID: 33202613 PMCID: PMC7697014 DOI: 10.3390/s20226512] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/09/2020] [Revised: 11/12/2020] [Accepted: 11/12/2020] [Indexed: 11/16/2022]
Abstract
Chronic obstructive pulmonary disease (COPD) and pneumonia are two of the few fatal lung diseases which share common adventitious lung sounds. Diagnosing the disease from lung sound analysis to design a noninvasive technique for telemedicine is a challenging task. A novel framework is presented to perform a diagnosis of COPD and Pneumonia via application of the signal processing and machine learning approach. This model will help the pulmonologist to accurately detect disease A and B. COPD, normal and pneumonia lung sound (LS) data from the ICBHI respiratory database is used in this research. The performance analysis is evidence of the improved performance of the quadratic discriminate classifier with an accuracy of 99.70% on selected fused features after experimentation. The fusion of time domain, cepstral, and spectral features are employed. Feature selection for fusion is performed through the back-elimination method whereas empirical mode decomposition (EMD) and discrete wavelet transform (DWT)-based techniques are used to denoise and segment the pulmonic signal. Class imbalance is catered with the implementation of the adaptive synthetic (ADASYN) sampling technique.
Collapse
Affiliation(s)
- Syed Zohaib Hassan Naqvi
- Department of Electronics Engineering, University of Engineering and Technology, Taxila 47080, Pakistan
- Correspondence:
| | - Mohammad Ahmad Choudhry
- Department of Electrical Engineering, University of Engineering and Technology, Taxila 47080, Pakistan;
| |
Collapse
|
46
|
Multi-channel lung sound classification with convolutional recurrent neural networks. Comput Biol Med 2020; 122:103831. [PMID: 32658732 DOI: 10.1016/j.compbiomed.2020.103831] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2019] [Revised: 05/18/2020] [Accepted: 05/21/2020] [Indexed: 11/20/2022]
Abstract
In this paper, we present an approach for multi-channel lung sound classification, exploiting spectral, temporal and spatial information. In particular, we propose a frame-wise classification framework to process full breathing cycles of multi-channel lung sound recordings with a convolutional recurrent neural network. With our recently developed 16-channel lung sound recording device, we collect lung sound recordings from lung-healthy subjects and patients with idiopathic pulmonary fibrosis (IPF), within a clinical trial. From the lung sound recordings, we extract spectrogram features and compare different deep neural network architectures for binary classification, i.e. healthy vs. pathological. Our proposed classification framework with the convolutional recurrent neural network outperforms the other networks by achieving an F-score of F1≈92%. Together with our multi-channel lung sound recording device, we present a holistic approach to multi-channel lung sound analysis.
Collapse
|