1
|
Huang L, Ruan S, Xing Y, Feng M. A review of uncertainty quantification in medical image analysis: Probabilistic and non-probabilistic methods. Med Image Anal 2024; 97:103223. [PMID: 38861770 DOI: 10.1016/j.media.2024.103223] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Revised: 03/16/2024] [Accepted: 05/27/2024] [Indexed: 06/13/2024]
Abstract
The comprehensive integration of machine learning healthcare models within clinical practice remains suboptimal, notwithstanding the proliferation of high-performing solutions reported in the literature. A predominant factor hindering widespread adoption pertains to an insufficiency of evidence affirming the reliability of the aforementioned models. Recently, uncertainty quantification methods have been proposed as a potential solution to quantify the reliability of machine learning models and thus increase the interpretability and acceptability of the results. In this review, we offer a comprehensive overview of the prevailing methods proposed to quantify the uncertainty inherent in machine learning models developed for various medical image tasks. Contrary to earlier reviews that exclusively focused on probabilistic methods, this review also explores non-probabilistic approaches, thereby furnishing a more holistic survey of research pertaining to uncertainty quantification for machine learning models. Analysis of medical images with the summary and discussion on medical applications and the corresponding uncertainty evaluation protocols are presented, which focus on the specific challenges of uncertainty in medical image analysis. We also highlight some potential future research work at the end. Generally, this review aims to allow researchers from both clinical and technical backgrounds to gain a quick and yet in-depth understanding of the research in uncertainty quantification for medical image analysis machine learning models.
Collapse
Affiliation(s)
- Ling Huang
- Saw Swee Hock School of Public Health, National University of Singapore, Singapore
| | - Su Ruan
- Quantif, LITIS, University of Rouen Normandy, France.
| | - Yucheng Xing
- Saw Swee Hock School of Public Health, National University of Singapore, Singapore
| | - Mengling Feng
- Saw Swee Hock School of Public Health, National University of Singapore, Singapore; Institute of Data Science, National University of Singapore, Singapore
| |
Collapse
|
2
|
Xu C, Wu J, Zhang F, Freer J, Zhang Z, Cheng Y. A deep image classification model based on prior feature knowledge embedding and application in medical diagnosis. Sci Rep 2024; 14:13244. [PMID: 38853158 PMCID: PMC11163012 DOI: 10.1038/s41598-024-63818-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 06/03/2024] [Indexed: 06/11/2024] Open
Abstract
Aiming at the problem of image classification with insignificant morphological structural features, strong target correlation, and low signal-to-noise ratio, combined with prior feature knowledge embedding, a deep learning method based on ResNet and Radial Basis Probabilistic Neural Network (RBPNN) is proposed model. Taking ResNet50 as a visual modeling network, it uses feature pyramid and self-attention mechanism to extract appearance and semantic features of images at multiple scales, and associate and enhance local and global features. Taking into account the diversity of category features, channel cosine similarity attention and dynamic C-means clustering algorithms are used to select representative sample features in different category of sample subsets to implicitly express prior category feature knowledge, and use them as the kernel centers of radial basis probability neurons (RBPN) to realize the embedding of diverse prior feature knowledge. In the RBPNN pattern aggregation layer, the outputs of RBPN are selectively summed according to the category of the kernel center, that is, the subcategory features are combined into category features, and finally the image classification is implemented based on Softmax. The functional module of the proposed method is designed specifically for image characteristics, which can highlight the significance of local and structural features of the image, form a non-convex decision-making area, and reduce the requirements for the completeness of the sample set. Applying the proposed method to medical image classification, experiments were conducted based on the brain tumor MRI image classification public dataset and the actual cardiac ultrasound image dataset, and the accuracy rate reached 85.82% and 83.92% respectively. Compared with the three mainstream image classification models, the performance indicators of this method have been significantly improved.
Collapse
Affiliation(s)
- Chen Xu
- School of Computer Science, Fudan University, Shanghai, China.
| | - Jiangxing Wu
- School of Computer Science, Fudan University, Shanghai, China
| | - Fan Zhang
- School of Computer Science, Fudan University, Shanghai, China
| | - Jonathan Freer
- School of Computer Science, University of Birmingham, Birmingham, UK
| | - Zhongqun Zhang
- School of Computer Science, University of Birmingham, Birmingham, UK
| | - Yihua Cheng
- School of Computer Science, University of Birmingham, Birmingham, UK
| |
Collapse
|
3
|
Zahari R, Cox J, Obara B. Uncertainty-aware image classification on 3D CT lung. Comput Biol Med 2024; 172:108324. [PMID: 38508053 DOI: 10.1016/j.compbiomed.2024.108324] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Revised: 03/06/2024] [Accepted: 03/14/2024] [Indexed: 03/22/2024]
Abstract
Early detection is crucial for lung cancer to prolong the patient's survival. Existing model architectures used in such systems have shown promising results. However, they lack reliability and robustness in their predictions and the models are typically evaluated on a single dataset, making them overconfident when a new class is present. With the existence of uncertainty, uncertain images can be referred to medical experts for a second opinion. Thus, we propose an uncertainty-aware framework that includes three phases: data preprocessing and model selection and evaluation, uncertainty quantification (UQ), and uncertainty measurement and data referral for the classification of benign and malignant nodules using 3D CT images. To quantify the uncertainty, we employed three approaches; Monte Carlo Dropout (MCD), Deep Ensemble (DE), and Ensemble Monte Carlo Dropout (EMCD). We evaluated eight different deep learning models consisting of ResNet, DenseNet, and the Inception network family, all of which achieved average F1 scores above 0.832, and the highest average value of 0.845 was obtained using InceptionResNetV2. Furthermore, incorporating the UQ demonstrated significant improvement in the overall model performance. Upon evaluation of the uncertainty estimate, MCD outperforms the other UQ models except for the metric, URecall, where DE and EMCD excel, implying that they are better at identifying incorrect predictions with higher uncertainty levels, which is vital in the medical field. Finally, we show that using a threshold for data referral can greatly improve the performance further, increasing the accuracy up to 0.959.
Collapse
Affiliation(s)
- Rahimi Zahari
- School of Computing, Newcastle University, Newcastle upon Tyne, UK
| | - Julie Cox
- County Durham and Darlington NHS Foundation Trust, County Durham, UK
| | - Boguslaw Obara
- School of Computing, Newcastle University, Newcastle upon Tyne, UK; Biosciences Institute, Newcastle University, Newcastle upon Tyne, UK.
| |
Collapse
|
4
|
Abdullah AA, Hassan MM, Mustafa YT. Leveraging Bayesian deep learning and ensemble methods for uncertainty quantification in image classification: A ranking-based approach. Heliyon 2024; 10:e24188. [PMID: 38293520 PMCID: PMC10825337 DOI: 10.1016/j.heliyon.2024.e24188] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2023] [Revised: 12/08/2023] [Accepted: 01/04/2024] [Indexed: 02/01/2024] Open
Abstract
Bayesian deep learning (BDL) has emerged as a powerful technique for quantifying uncertainty in classification tasks, surpassing the effectiveness of traditional models by aligning with the probabilistic nature of real-world data. This alignment allows for informed decision-making by not only identifying the most likely outcome but also quantifying the surrounding uncertainty. Such capabilities hold great significance in fields like medical diagnoses and autonomous driving, where the consequences of misclassification are substantial. To further improve uncertainty quantification, the research community has introduced Bayesian model ensembles, which combines multiple Bayesian models to enhance predictive accuracy and uncertainty quantification. These ensembles have exhibited superior performance compared to individual Bayesian models and even non-Bayesian counterparts. In this study, we propose a novel approach that leverages the power of Bayesian ensembles for enhanced uncertainty quantification. The proposed method exploits the disparity between predicted positive and negative classes and employes it as a ranking metric for model selection. For each instance or sample, the ensemble's output for each class is determined by selecting the top 'k' models based on this ranking. Experimental results on different medical image classifications demonstrate that the proposed method consistently outperforms or achieves comparable performance to conventional Bayesian ensemble. This investigation highlights the practical application of Bayesian ensemble techniques in refining predictive performance and enhancing uncertainty evaluation in image classification tasks.
Collapse
Affiliation(s)
- Abdullah A. Abdullah
- Computer Science Department, Faculty of Science, University of Zakho, Duhok, Iraq
| | - Masoud M. Hassan
- Computer Science Department, Faculty of Science, University of Zakho, Duhok, Iraq
| | - Yaseen T. Mustafa
- Environmental Science Department, Faculty of Science, University of Zakho, Duhok, Iraq
| |
Collapse
|
5
|
Baldeon-Calisto M, Lai-Yuen SK, Puente-Mejia B. StAC-DA: Structure aware cross-modality domain adaptation framework with image and feature-level adaptation for medical image segmentation. Digit Health 2024; 10:20552076241277440. [PMID: 39229464 PMCID: PMC11369866 DOI: 10.1177/20552076241277440] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2024] [Accepted: 08/06/2024] [Indexed: 09/05/2024] Open
Abstract
Objective Convolutional neural networks (CNNs) have achieved state-of-the-art results in various medical image segmentation tasks. However, CNNs often assume that the source and target dataset follow the same probability distribution and when this assumption is not satisfied their performance degrades significantly. This poses a limitation in medical image analysis, where including information from different imaging modalities can bring large clinical benefits. In this work, we present an unsupervised Structure Aware Cross-modality Domain Adaptation (StAC-DA) framework for medical image segmentation. Methods StAC-DA implements an image- and feature-level adaptation in a sequential two-step approach. The first step performs an image-level alignment, where images from the source domain are translated to the target domain in pixel space by implementing a CycleGAN-based model. The latter model includes a structure-aware network that preserves the shape of the anatomical structure during translation. The second step consists of a feature-level alignment. A U-Net network with deep supervision is trained with the transformed source domain images and target domain images in an adversarial manner to produce probable segmentations for the target domain. Results The framework is evaluated on bidirectional cardiac substructure segmentation. StAC-DA outperforms leading unsupervised domain adaptation approaches, being ranked first in the segmentation of the ascending aorta when adapting from Magnetic Resonance Imaging (MRI) to Computed Tomography (CT) domain and from CT to MRI domain. Conclusions The presented framework overcomes the limitations posed by differing distributions in training and testing datasets. Moreover, the experimental results highlight its potential to improve the accuracy of medical image segmentation across diverse imaging modalities.
Collapse
Affiliation(s)
- Maria Baldeon-Calisto
- Departamento de Ingeniería Industrial, Colegio de Ciencias e Ingeniería, Instituto de Innovación en Productividad y Logística CATENA-USFQ, Universidad San Francisco de Quito, Quito, Ecuador
| | - Susana K. Lai-Yuen
- Department of Industrial and Management Systems, University of South Florida, Tampa, FL, USA
| | - Bernardo Puente-Mejia
- Departamento de Ingeniería Industrial, Colegio de Ciencias e Ingeniería, Instituto de Innovación en Productividad y Logística CATENA-USFQ, Universidad San Francisco de Quito, Quito, Ecuador
| |
Collapse
|
6
|
Ozaltin O, Yeniay O, Subasi A. OzNet: A New Deep Learning Approach for Automated Classification of COVID-19 Computed Tomography Scans. BIG DATA 2023; 11:420-436. [PMID: 36927081 DOI: 10.1089/big.2022.0042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Coronavirus disease 2019 (COVID-19) is spreading rapidly around the world. Therefore, the classification of computed tomography (CT) scans alleviates the workload of experts, whose workload increased considerably during the pandemic. Convolutional neural network (CNN) architectures are successful for the classification of medical images. In this study, we have developed a new deep CNN architecture called OzNet. Moreover, we have compared it with pretrained architectures namely AlexNet, DenseNet201, GoogleNet, NASNetMobile, ResNet-50, SqueezeNet, and VGG-16. In addition, we have compared the classification success of three preprocessing methods with raw CT scans. We have not only classified the raw CT scans, but also have performed the classification with three different preprocessing methods, which are discrete wavelet transform (DWT), intensity adjustment, and gray to color red, green, blue image conversion on the data sets. Furthermore, it is known that the architecture's performance increases with the use of DWT preprocessing method rather than using the raw data set. The results are extremely promising with the CNN algorithms using the COVID-19 CT scans processed with the DWT. The proposed DWT-OzNet has achieved a high classification performance of more than 98.8% for each calculated metric.
Collapse
Affiliation(s)
- Oznur Ozaltin
- Department of Statistics, Institute of Science, Hacettepe University, Ankara, Turkey
| | - Ozgur Yeniay
- Department of Statistics, Institute of Science, Hacettepe University, Ankara, Turkey
| | - Abdulhamit Subasi
- Institute of Biomedicine, Faculty of Medicine, University of Turku, Turku, Finland
- Department of Computer Science, College of Engineering, Effat University, Jeddah, Saudi Arabia
| |
Collapse
|
7
|
Ajagbe SA, Adigun MO. Deep learning techniques for detection and prediction of pandemic diseases: a systematic literature review. MULTIMEDIA TOOLS AND APPLICATIONS 2023:1-35. [PMID: 37362693 PMCID: PMC10226029 DOI: 10.1007/s11042-023-15805-z] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/17/2022] [Revised: 03/06/2023] [Accepted: 05/10/2023] [Indexed: 06/28/2023]
Abstract
Deep learning (DL) is becoming a fast-growing field in the medical domain and it helps in the timely detection of any infectious disease (IDs) and is essential to the management of diseases and the prediction of future occurrences. Many scientists and scholars have implemented DL techniques for the detection and prediction of pandemics, IDs and other healthcare-related purposes, these outcomes are with various limitations and research gaps. For the purpose of achieving an accurate, efficient and less complicated DL-based system for the detection and prediction of pandemics, therefore, this study carried out a systematic literature review (SLR) on the detection and prediction of pandemics using DL techniques. The survey is anchored by four objectives and a state-of-the-art review of forty-five papers out of seven hundred and ninety papers retrieved from different scholarly databases was carried out in this study to analyze and evaluate the trend of DL techniques application areas in the detection and prediction of pandemics. This study used various tables and graphs to analyze the extracted related articles from various online scholarly repositories and the analysis showed that DL techniques have a good tool in pandemic detection and prediction. Scopus and Web of Science repositories are given attention in this current because they contain suitable scientific findings in the subject area. Finally, the state-of-the-art review presents forty-four (44) studies of various DL technique performances. The challenges identified from the literature include the low performance of the model due to computational complexities, improper labeling and the absence of a high-quality dataset among others. This survey suggests possible solutions such as the development of improved DL-based techniques or the reduction of the output layer of DL-based architecture for the detection and prediction of pandemic-prone diseases as future considerations.
Collapse
Affiliation(s)
- Sunday Adeola Ajagbe
- Department of Computer & Industrial Production Engineering, First Technical University Ibadan, Ibadan, 200255 Nigeria
- Department of Computer Science, University of Zululand, Kwadlangezwa, 3886 South Africa
| | - Matthew O. Adigun
- Department of Computer Science, University of Zululand, Kwadlangezwa, 3886 South Africa
| |
Collapse
|
8
|
Li G, Togo R, Ogawa T, Haseyama M. Boosting automatic COVID-19 detection performance with self-supervised learning and batch knowledge ensembling. Comput Biol Med 2023; 158:106877. [PMID: 37019015 PMCID: PMC10063457 DOI: 10.1016/j.compbiomed.2023.106877] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Revised: 03/15/2023] [Accepted: 03/30/2023] [Indexed: 04/03/2023]
Abstract
PROBLEM Detecting COVID-19 from chest X-ray (CXR) images has become one of the fastest and easiest methods for detecting COVID-19. However, the existing methods usually use supervised transfer learning from natural images as a pretraining process. These methods do not consider the unique features of COVID-19 and the similar features between COVID-19 and other pneumonia. AIM In this paper, we want to design a novel high-accuracy COVID-19 detection method that uses CXR images, which can consider the unique features of COVID-19 and the similar features between COVID-19 and other pneumonia. METHODS Our method consists of two phases. One is self-supervised learning-based pertaining; the other is batch knowledge ensembling-based fine-tuning. Self-supervised learning-based pretraining can learn distinguished representations from CXR images without manually annotated labels. On the other hand, batch knowledge ensembling-based fine-tuning can utilize category knowledge of images in a batch according to their visual feature similarities to improve detection performance. Unlike our previous implementation, we introduce batch knowledge ensembling into the fine-tuning phase, reducing the memory used in self-supervised learning and improving COVID-19 detection accuracy. RESULTS On two public COVID-19 CXR datasets, namely, a large dataset and an unbalanced dataset, our method exhibited promising COVID-19 detection performance. Our method maintains high detection accuracy even when annotated CXR training images are reduced significantly (e.g., using only 10% of the original dataset). In addition, our method is insensitive to changes in hyperparameters. CONCLUSION The proposed method outperforms other state-of-the-art COVID-19 detection methods in different settings. Our method can reduce the workloads of healthcare providers and radiologists.
Collapse
Affiliation(s)
- Guang Li
- Graduate School of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| | - Ren Togo
- Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| | - Takahiro Ogawa
- Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| | - Miki Haseyama
- Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| |
Collapse
|
9
|
Feng Y, Luo Y, Yang J. Cross-platform privacy-preserving CT image COVID-19 diagnosis based on source-free domain adaptation. Knowl Based Syst 2023; 264:110324. [PMID: 36713615 PMCID: PMC9869622 DOI: 10.1016/j.knosys.2023.110324] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2022] [Revised: 01/05/2023] [Accepted: 01/17/2023] [Indexed: 01/24/2023]
Abstract
In the wake of the Coronavirus disease (COVID-19) pandemic, chest computed tomography (CT) has become an invaluable component in the rapid and accurate detection of COVID-19. CT scans traditionally require manual inspections from medical professionals, which is expensive and tedious. With advancements in machine learning, deep neural networks have been applied to classify CT scans for efficient diagnosis. However, three challenges hinder this application of deep learning: (1) Domain shift across CT platforms and human subjects impedes the performance of neural networks in different hospitals. (2) Unsupervised Domain Adaptation (UDA), the traditional method to overcome domain shift, typically requires access to both source and target data. This is not realistic in COVID-19 diagnosis due to the sensitivity of medical data. The privacy of patients must be protected. (3) Data imbalance may exist between easy/hard samples and between data classes which can overwhelm the training of deep networks, causing degenerate models. To overcome these challenges, we propose a Cross-Platform Privacy-Preserving COVID-19 diagnosis network (CP 3 Net) that integrates domain adaptation, self-supervised learning, imbalanced label learning, and rotation classifier training into one synergistic framework. We also create a new CT benchmark by combining real-world datasets from multiple medical platforms to facilitate the cross-domain evaluation of our method. Through extensive experiments, we demonstrate that CP 3 Net outperforms many popular UDA methods and achieves state-of-the-art results in diagnosing COVID-19 using CT scans.
Collapse
Affiliation(s)
| | - Yuemei Luo
- School of Artificial Intelligence, Nanjing University of Information Science & Technology, China
| | - Jianfei Yang
- School of Electrical and Electronics Engineering, Nanyang Technological University, Singapore,Corresponding author
| |
Collapse
|
10
|
Paul SG, Saha A, Biswas AA, Zulfiker MS, Arefin MS, Rahman MM, Reza AW. Combating Covid-19 using machine learning and deep learning: Applications, challenges, and future perspectives. ARRAY 2023; 17:100271. [PMID: 36530931 PMCID: PMC9737520 DOI: 10.1016/j.array.2022.100271] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2022] [Revised: 12/05/2022] [Accepted: 12/07/2022] [Indexed: 12/14/2022] Open
Abstract
COVID-19, a worldwide pandemic that has affected many people and thousands of individuals have died due to COVID-19, during the last two years. Due to the benefits of Artificial Intelligence (AI) in X-ray image interpretation, sound analysis, diagnosis, patient monitoring, and CT image identification, it has been further researched in the area of medical science during the period of COVID-19. This study has assessed the performance and investigated different machine learning (ML), deep learning (DL), and combinations of various ML, DL, and AI approaches that have been employed in recent studies with diverse data formats to combat the problems that have arisen due to the COVID-19 pandemic. Finally, this study shows the comparison among the stand-alone ML and DL-based research works regarding the COVID-19 issues with the combinations of ML, DL, and AI-based research works. After in-depth analysis and comparison, this study responds to the proposed research questions and presents the future research directions in this context. This review work will guide different research groups to develop viable applications based on ML, DL, and AI models, and will also guide healthcare institutes, researchers, and governments by showing them how these techniques can ease the process of tackling the COVID-19.
Collapse
Affiliation(s)
- Showmick Guha Paul
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Arpa Saha
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Al Amin Biswas
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh,Corresponding author
| | - Md. Sabab Zulfiker
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Mohammad Shamsul Arefin
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh,Department of Computer Science and Engineering, Chittagong University of Engineering and Technology, Chittagong, Bangladesh
| | - Md. Mahfujur Rahman
- Department of Computer Science and Engineering, Daffodil International University, Dhaka, Bangladesh
| | - Ahmed Wasif Reza
- Department of Computer Science and Engineering, East West University, Dhaka, Bangladesh
| |
Collapse
|
11
|
Novel Light Convolutional Neural Network for COVID Detection with Watershed Based Region Growing Segmentation. J Imaging 2023; 9:jimaging9020042. [PMID: 36826961 PMCID: PMC9963211 DOI: 10.3390/jimaging9020042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Revised: 02/06/2023] [Accepted: 02/08/2023] [Indexed: 02/16/2023] Open
Abstract
A rapidly spreading epidemic, COVID-19 had a serious effect on millions and took many lives. Therefore, for individuals with COVID-19, early discovery is essential for halting the infection's progress. To quickly and accurately diagnose COVID-19, imaging modalities, including computed tomography (CT) scans and chest X-ray radiographs, are frequently employed. The potential of artificial intelligence (AI) approaches further explored the creation of automated and precise COVID-19 detection systems. Scientists widely use deep learning techniques to identify coronavirus infection in lung imaging. In our paper, we developed a novel light CNN model architecture with watershed-based region-growing segmentation on Chest X-rays. Both CT scans and X-ray radiographs were employed along with 5-fold cross-validation. Compared to earlier state-of-the-art models, our model is lighter and outperformed the previous methods by achieving a mean accuracy of 98.8% on X-ray images and 98.6% on CT scans, predicting the rate of 0.99% and 0.97% for PPV (Positive predicted Value) and NPV (Negative predicted Value) rate of 0.98% and 0.99%, respectively.
Collapse
|
12
|
Hernández S, López-Córtes X. Evaluating deep learning predictions for COVID-19 from X-ray images using leave-one-out predictive densities. Neural Comput Appl 2023; 35:9819-9830. [PMID: 36778196 PMCID: PMC9900537 DOI: 10.1007/s00521-023-08219-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2022] [Accepted: 01/06/2023] [Indexed: 02/09/2023]
Abstract
Early detection of the COVID-19 virus is an important task for controlling the spread of the pandemic. Imaging techniques such as chest X-ray are relatively inexpensive and accessible, but its interpretation requires expert knowledge to evaluate the disease severity. Several approaches for automatic COVID-19 detection using deep learning techniques have been proposed. While most approaches show high accuracy on the COVID-19 detection task, there is not enough evidence on external evaluation for this technique. Furthermore, data scarcity and sampling biases make difficult to properly evaluate model predictions. In this paper, we propose stochastic gradient Langevin dynamics (SGLD) to take into account the model uncertainty. Four different deep learning architectures are trained using SGLD and compared to their baselines using stochastic gradient descent. The model uncertainties are also evaluated according to their convergence properties and the leave-one-out predictive densities. The proposed approach is able to reduce overconfidence of the baseline estimators while also retaining predictive accuracy for the best-performing cases.
Collapse
Affiliation(s)
- Sergio Hernández
- Departamento de Computación en Industrias. Facultad de Ciencias de la Ingeniería, Universidad Católica del Maule, Av. San Miguel 3605, 100190 Talca, Maule, Chile
| | - Xaviera López-Córtes
- Departamento de Computación en Industrias. Facultad de Ciencias de la Ingeniería, Universidad Católica del Maule, Av. San Miguel 3605, 100190 Talca, Maule, Chile
| |
Collapse
|
13
|
Aldhahi W, Sull S. Uncertain-CAM: Uncertainty-Based Ensemble Machine Voting for Improved COVID-19 CXR Classification and Explainability. Diagnostics (Basel) 2023; 13:441. [PMID: 36766546 PMCID: PMC9914375 DOI: 10.3390/diagnostics13030441] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2022] [Revised: 01/08/2023] [Accepted: 01/20/2023] [Indexed: 01/27/2023] Open
Abstract
The ongoing coronavirus disease 2019 (COVID-19) pandemic has had a significant impact on patients and healthcare systems across the world. Distinguishing non-COVID-19 patients from COVID-19 patients at the lowest possible cost and in the earliest stages of the disease is a major issue. Additionally, the implementation of explainable deep learning decisions is another issue, especially in critical fields such as medicine. The study presents a method to train deep learning models and apply an uncertainty-based ensemble voting policy to achieve 99% accuracy in classifying COVID-19 chest X-rays from normal and pneumonia-related infections. We further present a training scheme that integrates the cyclic cosine annealing approach with cross-validation and uncertainty quantification that is measured using prediction interval coverage probability (PICP) as final ensemble voting weights. We also propose the Uncertain-CAM technique, which improves deep learning explainability and provides a more reliable COVID-19 classification system. We introduce a new image processing technique to measure the explainability based on ground-truth, and we compared it with the widely adopted Grad-CAM method.
Collapse
Affiliation(s)
| | - Sanghoon Sull
- School of Electrical Engineering, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul 02841, Republic of Korea
| |
Collapse
|
14
|
Breast cancer classification by a new approach to assessing deep neural network-based uncertainty quantification methods. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104057] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|
15
|
Dalvi PP, Edla DR, Purushothama BR. Diagnosis of Coronavirus Disease From Chest X-Ray Images Using DenseNet-169 Architecture. SN COMPUTER SCIENCE 2023; 4:214. [PMID: 36811126 PMCID: PMC9936468 DOI: 10.1007/s42979-022-01627-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/27/2022] [Accepted: 12/17/2022] [Indexed: 02/19/2023]
Abstract
The coronavirus disease (COVID-19) is a very contagious and dangerous disease that affects the human respiratory system. Early detection of this disease is very crucial to contain the further spread of the virus. In this paper, we proposed a methodology using DenseNet-169 architecture for diagnosing the disease from chest X-ray images of the patients. We used a pretrained neural network and then utilised the transfer learning method for training on our dataset. We also used Nearest-Neighbour interpolation technique for data preprocessing and Adam Optimizer at the end for optimization. Our methodology achieved 96.37 % accuracy which was better than that obtained using other deep learning models like AlexNet, ResNet-50, VGG-16, and VGG-19.
Collapse
Affiliation(s)
- Pooja Pradeep Dalvi
- Department of Computer Science and Engineering, National Institute of Technology, Goa, India
| | - Damodar Reddy Edla
- Department of Computer Science and Engineering, National Institute of Technology, Goa, India
| | - B. R. Purushothama
- Department of Computer Science and Engineering, National Institute of Technology, Goa, India
| |
Collapse
|
16
|
Chow LS, Tang GS, Solihin MI, Gowdh NM, Ramli N, Rahmat K. Quantitative and Qualitative Analysis of 18 Deep Convolutional Neural Network (CNN) Models with Transfer Learning to Diagnose COVID-19 on Chest X-Ray (CXR) Images. SN COMPUTER SCIENCE 2023; 4:141. [PMID: 36624807 PMCID: PMC9813876 DOI: 10.1007/s42979-022-01545-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/22/2022] [Accepted: 12/03/2022] [Indexed: 01/06/2023]
Abstract
Coronavirus disease 2019 (COVID-19) is a disease caused by a novel strain of coronavirus, severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), severely affecting the lungs. Our study aims to combine both quantitative and qualitative analysis of the convolutional neural network (CNN) model to diagnose COVID-19 on chest X-ray (CXR) images. We investigated 18 state-of-the-art CNN models with transfer learning, which include AlexNet, DarkNet-19, DarkNet-53, DenseNet-201, GoogLeNet, Inception-ResNet-v2, Inception-v3, MobileNet-v2, NasNet-Large, NasNet-Mobile, ResNet-18, ResNet-50, ResNet-101, ShuffleNet, SqueezeNet, VGG-16, VGG-19, and Xception. Their performances were evaluated quantitatively using six assessment metrics: specificity, sensitivity, precision, negative predictive value (NPV), accuracy, and F1-score. The top four models with accuracy higher than 90% are VGG-16, ResNet-101, VGG-19, and SqueezeNet. The accuracy of these top four models is between 90.7% and 94.3%; the F1-score is between 90.8% and 94.3%. The VGG-16 scored the highest accuracy of 94.3% and F1-score of 94.3%. The majority voting with all the 18 CNN models and top 4 models produced an accuracy of 93.0% and 94.0%, respectively. The top four and bottom three models were chosen for the qualitative analysis. A gradient-weighted class activation mapping (Grad-CAM) was used to visualize the significant region of activation for the decision-making of image classification. Two certified radiologists performed blinded subjective voting on the Grad-CAM images in comparison with their diagnosis. The qualitative analysis showed that SqueezeNet is the closest model to the diagnosis of two certified radiologists. It demonstrated a competitively good accuracy of 90.7% and F1-score of 90.8% with 111 times fewer parameters and 7.7 times faster than VGG-16. Therefore, this study recommends both VGG-16 and SqueezeNet as additional tools for the diagnosis of COVID-19.
Collapse
Affiliation(s)
- Li Sze Chow
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Technology and Built Environment, UCSI University, 1, Jalan Puncak Menara Gading, Taman Connaught, Cheras, 56000 Kuala Lumpur, Malaysia
| | - Goon Sheng Tang
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Technology and Built Environment, UCSI University, 1, Jalan Puncak Menara Gading, Taman Connaught, Cheras, 56000 Kuala Lumpur, Malaysia
| | - Mahmud Iwan Solihin
- Department of Mechanical and Mechatronics Engineering, Faculty of Engineering, Technology and Built Environment, UCSI University, 1, Jalan Puncak Menara Gading, Taman Connaught, Cheras, 56000 Kuala Lumpur, Malaysia
| | - Nadia Muhammad Gowdh
- Department of Biomedical Imaging, Faculty of Medicine, University of Malaya, 50603 Kuala Lumpur, Malaysia
| | - Norlisah Ramli
- Department of Biomedical Imaging, Faculty of Medicine, University of Malaya, 50603 Kuala Lumpur, Malaysia
| | - Kartini Rahmat
- Department of Biomedical Imaging, Faculty of Medicine, University of Malaya, 50603 Kuala Lumpur, Malaysia
| |
Collapse
|
17
|
Shibu George G, Raj Mishra P, Sinha P, Ranjan Prusty M. COVID-19 Detection on Chest X-Ray Images Using Homomorphic Transformation and VGG Inspired Deep Convolutional Neural Network. Biocybern Biomed Eng 2022; 43:1-16. [PMID: 36447948 PMCID: PMC9684127 DOI: 10.1016/j.bbe.2022.11.003] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2022] [Revised: 11/01/2022] [Accepted: 11/18/2022] [Indexed: 11/25/2022]
Abstract
COVID-19 had caused the whole world to come to a standstill. The current detection methods are time consuming as well as costly. Using Chest X-rays (CXRs) is a solution to this problem, however, manual examination of CXRs is a cumbersome and difficult process needing specialization in the domain. Most of existing methods used for this application involve the usage of pretrained models such as VGG19, ResNet, DenseNet, Xception, and EfficeintNet which were trained on RGB image datasets. X-rays are fundamentally single channel images, hence using RGB trained model is not appropriate since it increases the operations by involving three channels instead of one. A way of using pretrained model for grayscale images is by replicating the one channel image data to three channel which introduces redundancy and another way is by altering the input layer of pretrained model to take in one channel image data, which comprises the weights in the forward layers that were trained on three channel images which weakens the use of pre-trained weights in a transfer learning approach. A novel approach for identification of COVID-19 using CXRs, Contrast Limited Adaptive Histogram Equalization (CLAHE) along with Homomorphic Transformation Filter which is used to process the pixel data in images and extract features from the CXRs is suggested in this paper. These processed images are then provided as input to a VGG inspired deep Convolutional Neural Network (CNN) model which takes one channel image data as input (grayscale images) to categorize CXRs into three class labels, namely, No-Findings, COVID-19, and Pneumonia. Evaluation of the suggested model is done with the help of two publicly available datasets; one to obtain COVID-19 and No-Finding images and the other to obtain Pneumonia CXRs. The dataset comprises 6750 images in total; 2250 images for each class. Results obtained show that the model has achieved 96.56% for multi-class classification and 98.06% accuracy for binary classification using 5-fold stratified cross validation (CV) method. This result is competitive and up to the mark when compared with the performance shown by existing approaches for COVID-19 classification.
Collapse
Affiliation(s)
- Gerosh Shibu George
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu 600127, India
| | - Pratyush Raj Mishra
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu 600127, India
| | - Panav Sinha
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu 600127, India
| | - Manas Ranjan Prusty
- Centre for Cyber Physical Systems, School of Computer Science and Engineering, Vellore Institute of Technology, Chennai, Tamil Nadu 600127, India
| |
Collapse
|
18
|
Pramanik R, Dey S, Malakar S, Mirjalili S, Sarkar R. TOPSIS aided ensemble of CNN models for screening COVID-19 in chest X-ray images. Sci Rep 2022; 12:15409. [PMID: 36104401 PMCID: PMC9471038 DOI: 10.1038/s41598-022-18463-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2021] [Accepted: 08/12/2022] [Indexed: 11/23/2022] Open
Abstract
The novel coronavirus (COVID-19), has undoubtedly imprinted our lives with its deadly impact. Early testing with isolation of the individual is the best possible way to curb the spread of this deadly virus. Computer aided diagnosis (CAD) provides an alternative and cheap option for screening of the said virus. In this paper, we propose a convolution neural network (CNN)-based CAD method for COVID-19 and pneumonia detection from chest X-ray images. We consider three input types for three identical base classifiers. To capture maximum possible complementary features, we consider the original RGB image, Red channel image and the original image stacked with Robert's edge information. After that we develop an ensemble strategy based on the technique for order preference by similarity to an ideal solution (TOPSIS) to aggregate the outcomes of base classifiers. The overall framework, called TOPCONet, is very light in comparison with standard CNN models in terms of the number of trainable parameters required. TOPCONet achieves state-of-the-art results when evaluated on the three publicly available datasets: (1) IEEE COVID-19 dataset + Kaggle Pneumonia Dataset, (2) Kaggle Radiography dataset and (3) COVIDx.
Collapse
|