1
|
Altan G, Narli SS. DeepCOVIDNet-CXR: deep learning strategies for identifying COVID-19 on enhanced chest X-rays. BIOMED ENG-BIOMED TE 2024:bmt-2021-0272. [PMID: 39370946 DOI: 10.1515/bmt-2021-0272] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2021] [Accepted: 09/10/2024] [Indexed: 10/08/2024]
Abstract
OBJECTIVES COVID-19 is one of the recent major epidemics, which accelerates its mortality and prevalence worldwide. Most literature on chest X-ray-based COVID-19 analysis has focused on multi-case classification (COVID-19, pneumonia, and normal) by the advantages of Deep Learning. However, the limited number of chest X-rays with COVID-19 is a prominent deficiency for clinical relevance. This study aims at evaluating COVID-19 identification performances using adaptive histogram equalization (AHE) to feed the ConvNet architectures with reliable lung anatomy of airways. METHODS We experimented with balanced small- and large-scale COVID-19 databases using left lung, right lung, and complete chest X-rays with various AHE parameters. On multiple strategies, we applied transfer learning on four ConvNet architectures (MobileNet, DarkNet19, VGG16, and AlexNet). RESULTS Whereas DarkNet19 reached the highest multi-case identification performance with an accuracy rate of 98.26 % on the small-scale dataset, VGG16 achieved the best generalization performance with an accuracy rate of 95.04 % on the large-scale dataset. CONCLUSIONS Our study is one of the pioneering approaches that analyses 3615 COVID-19 cases and specifies the most responsible AHE parameters for ConvNet architectures in the multi-case classification.
Collapse
Affiliation(s)
- Gokhan Altan
- Computer Engineering Department, Iskenderun Technical University, Hatay, Türkiye
| | | |
Collapse
|
2
|
Song M, Wang J, Yu Z, Wang J, Yang L, Lu Y, Li B, Wang X, Wang X, Huang Q, Li Z, Kanellakis NI, Liu J, Wang J, Wang B, Yang J. PneumoLLM: Harnessing the power of large language model for pneumoconiosis diagnosis. Med Image Anal 2024; 97:103248. [PMID: 38941859 DOI: 10.1016/j.media.2024.103248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2023] [Revised: 06/17/2024] [Accepted: 06/17/2024] [Indexed: 06/30/2024]
Abstract
The conventional pretraining-and-finetuning paradigm, while effective for common diseases with ample data, faces challenges in diagnosing data-scarce occupational diseases like pneumoconiosis. Recently, large language models (LLMs) have exhibits unprecedented ability when conducting multiple tasks in dialogue, bringing opportunities to diagnosis. A common strategy might involve using adapter layers for vision-language alignment and diagnosis in a dialogic manner. Yet, this approach often requires optimization of extensive learnable parameters in the text branch and the dialogue head, potentially diminishing the LLMs' efficacy, especially with limited training data. In our work, we innovate by eliminating the text branch and substituting the dialogue head with a classification head. This approach presents a more effective method for harnessing LLMs in diagnosis with fewer learnable parameters. Furthermore, to balance the retention of detailed image information with progression towards accurate diagnosis, we introduce the contextual multi-token engine. This engine is specialized in adaptively generating diagnostic tokens. Additionally, we propose the information emitter module, which unidirectionally emits information from image tokens to diagnosis tokens. Comprehensive experiments validate the superiority of our methods.
Collapse
Affiliation(s)
- Meiyue Song
- Institute of Basic Medical Sciences Chinese Academy of Medical Sciences, School of Basic Medicine Peking Union Medical College, Beijing, 100005, China; State Key Laboratory of Respiratory Health and Multimorbidity, Beijing, 100005, China
| | - Jiarui Wang
- School of Automation, Northwestern Polytechnical University, Shaanxi, Xi'an 710072, China
| | - Zhihua Yu
- Jinneng Holding Coal Industry Group Co. Ltd Occupational Disease Precaution Clinic, Shanxi, 037001, China
| | - Jiaxin Wang
- School of Medicine, Tsinghua University, Beijing, 100084, China
| | - Le Yang
- School of Electronics and Control Engineering, Chang'an University, Shaanxi, Xi'an 710064, China
| | - Yuting Lu
- School of Automation, Northwestern Polytechnical University, Shaanxi, Xi'an 710072, China
| | - Baicun Li
- Center of Respiratory Medicine, China-Japan Friendship Hospital, National Center for Respiratory Medicine, Institute of Respiratory Medicine, Chinese Academy of Medical Sciences, National Clinical Research Center for Respiratory Diseases, Beijing, 100020, China
| | - Xue Wang
- Department of Respiratory, the Second Affiliated Hospital of Harbin Medical University, Harbin, Heilongjiang, 150086, China; Internal Medicine, Harbin Medical University, Harbin, Heilongjiang, 150081, China
| | - Xiaoxu Wang
- School of Automation, Northwestern Polytechnical University, Shaanxi, Xi'an 710072, China
| | - Qinghua Huang
- School of Artificial Intelligence, OPtics and ElectroNics (iOPEN), Northwestern Polytechnical University, Xi'an 710072, China
| | - Zhijun Li
- Translational Research Center, Shanghai YangZhi Rehabilitation Hospital (Shanghai Sunshine Rehabilitation Center), Shanghai 201619, China; School of Mechanical Engineering, Tongji University, Shanghai 201804, China
| | - Nikolaos I Kanellakis
- Laboratory of Pleural and Lung Cancer Translational Research, CAMS Oxford Institute, Nuffield Department of Medicine, University of Oxford, Oxford, UK; Oxford Centre for Respiratory Medicine, Churchill Hospital, Oxford University Hospitals NHS Foundation Trust, Oxford, UK; National Institute for Health Research Oxford Biomedical Research Centre, University of Oxford, Oxford, UK
| | - Jiangfeng Liu
- Institute of Basic Medical Sciences Chinese Academy of Medical Sciences, School of Basic Medicine Peking Union Medical College, Beijing, 100005, China; Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, 100144, China; State Key Laboratory of Common Mechanism Research for Major Diseases, Beijing, 100005, China.
| | - Jing Wang
- Institute of Basic Medical Sciences Chinese Academy of Medical Sciences, School of Basic Medicine Peking Union Medical College, Beijing, 100005, China; State Key Laboratory of Respiratory Health and Multimorbidity, Beijing, 100005, China.
| | - Binglu Wang
- School of Automation, Northwestern Polytechnical University, Shaanxi, Xi'an 710072, China.
| | - Juntao Yang
- Institute of Basic Medical Sciences Chinese Academy of Medical Sciences, School of Basic Medicine Peking Union Medical College, Beijing, 100005, China; Plastic Surgery Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, 100144, China; State Key Laboratory of Common Mechanism Research for Major Diseases, Beijing, 100005, China
| |
Collapse
|
3
|
Sun J, Chen K, He Z, Ren S, He X, Liu X, Peng C. Medical image analysis using improved SAM-Med2D: segmentation and classification perspectives. BMC Med Imaging 2024; 24:241. [PMID: 39285324 PMCID: PMC11403950 DOI: 10.1186/s12880-024-01401-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Accepted: 08/14/2024] [Indexed: 09/22/2024] Open
Abstract
Recently emerged SAM-Med2D represents a state-of-the-art advancement in medical image segmentation. Through fine-tuning the Large Visual Model, Segment Anything Model (SAM), on extensive medical datasets, it has achieved impressive results in cross-modal medical image segmentation. However, its reliance on interactive prompts may restrict its applicability under specific conditions. To address this limitation, we introduce SAM-AutoMed, which achieves automatic segmentation of medical images by replacing the original prompt encoder with an improved MobileNet v3 backbone. The performance on multiple datasets surpasses both SAM and SAM-Med2D. Current enhancements on the Large Visual Model SAM lack applications in the field of medical image classification. Therefore, we introduce SAM-MedCls, which combines the encoder of SAM-Med2D with our designed attention modules to construct an end-to-end medical image classification model. It performs well on datasets of various modalities, even achieving state-of-the-art results, indicating its potential to become a universal model for medical image classification.
Collapse
Affiliation(s)
- Jiakang Sun
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China
| | - Ke Chen
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China
| | - Zhiyi He
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China
| | - Siyuan Ren
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China
| | - Xinyang He
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China
| | - Xu Liu
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China
| | - Cheng Peng
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, 610213, Sichuan, China.
- School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, 101499, China.
| |
Collapse
|
4
|
Torrik A, Zarif M. Machine learning assisted sorting of active microswimmers. J Chem Phys 2024; 161:094907. [PMID: 39225539 DOI: 10.1063/5.0216862] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2024] [Accepted: 08/19/2024] [Indexed: 09/04/2024] Open
Abstract
Active matter systems, being in a non-equilibrium state, exhibit complex behaviors, such as self-organization, giving rise to emergent phenomena. There are many examples of active particles with biological origins, including bacteria and spermatozoa, or with artificial origins, such as self-propelled swimmers and Janus particles. The ability to manipulate active particles is vital for their effective application, e.g., separating motile spermatozoa from nonmotile and dead ones, to increase fertilization chance. In this study, we proposed a mechanism-an apparatus-to sort and demix active particles based on their motility values (Péclet number). Initially, using Brownian simulations, we demonstrated the feasibility of sorting self-propelled particles. Following this, we employed machine learning methods, supplemented with data from comprehensive simulations that we conducted for this study, to model the complex behavior of active particles. This enabled us to sort them based on their Péclet number. Finally, we evaluated the performance of the developed models and showed their effectiveness in demixing and sorting the active particles. Our findings can find applications in various fields, including physics, biology, and biomedical science, where the sorting and manipulation of active particles play a pivotal role.
Collapse
Affiliation(s)
- Abdolhalim Torrik
- Department of Physical and Computational Chemistry, Shahid Beheshti University, Tehran 19839-9411, Iran
| | - Mahdi Zarif
- Department of Physical and Computational Chemistry, Shahid Beheshti University, Tehran 19839-9411, Iran
| |
Collapse
|
5
|
Rajaraman S, Xue Z, Antani S. Editorial on Special Issue "Artificial Intelligence in Image-Based Screening, Diagnostics, and Clinical Care". Diagnostics (Basel) 2024; 14:1984. [PMID: 39272768 PMCID: PMC11393920 DOI: 10.3390/diagnostics14171984] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2024] [Accepted: 09/04/2024] [Indexed: 09/15/2024] Open
Abstract
In an era of rapid advancements in artificial intelligence (AI) technologies, particularly in medical imaging and natural language processing, strategic efforts to leverage AI's capabilities in analyzing complex medical data and integrating it into clinical workflows have emerged as a key driver of innovation in healthcare [...].
Collapse
Affiliation(s)
- Sivaramakrishnan Rajaraman
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894, USA
| | - Zhiyun Xue
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894, USA
| | - Sameer Antani
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894, USA
| |
Collapse
|
6
|
Nath A, Hashim Z, Shukla S, Poduvattil PA, Neyaz Z, Mishra R, Singh M, Misra N, Shukla A. A multicentre study to evaluate the diagnostic performance of a novel CAD software, DecXpert, for radiological diagnosis of tuberculosis in the northern Indian population. Sci Rep 2024; 14:20711. [PMID: 39237689 PMCID: PMC11377743 DOI: 10.1038/s41598-024-71346-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2024] [Accepted: 08/27/2024] [Indexed: 09/07/2024] Open
Abstract
Tuberculosis (TB) is the leading cause of mortality among infectious diseases globally. Effectively managing TB requires early identification of individuals with TB disease. Resource-constrained settings often lack skilled professionals for interpreting chest X-rays (CXRs) used in TB diagnosis. To address this challenge, we developed "DecXpert" a novel Computer-Aided Detection (CAD) software solution based on deep neural networks for early TB diagnosis from CXRs, aiming to detect subtle abnormalities that may be overlooked by human interpretation alone. This study was conducted on the largest cohort size to date, where the performance of a CAD software (DecXpert version 1.4) was validated against the gold standard molecular diagnostic technique, GeneXpert MTB/RIF, analyzing data from 4363 individuals across 12 primary health care centers and one tertiary hospital in North India. DecXpert demonstrated 88% sensitivity (95% CI 0.85-0.93) and 85% specificity (95% CI 0.82-0.91) for active TB detection. Incorporating demographics, DecXpert achieved an area under the curve of 0.91 (95% CI 0.88-0.94), indicating robust diagnostic performance. Our findings establish DecXpert's potential as an accurate, efficient AI solution for early identification of active TB cases. Deployed as a screening tool in resource-limited settings, DecXpert could enable early identification of individuals with TB disease and facilitate effective TB management where skilled radiological interpretation is limited.
Collapse
Affiliation(s)
- Alok Nath
- Department of Pulmonary Medicine, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India
| | - Zia Hashim
- Department of Pulmonary Medicine, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India
| | - Saumya Shukla
- Department of Pulmonary Medicine, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India
| | - Prasanth Areekkara Poduvattil
- Department of Pulmonary Medicine, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India
| | - Zafar Neyaz
- Department of Radiodiagnosis, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India
| | - Richa Mishra
- Department of Microbiology, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India
| | - Manika Singh
- Baker Heart and Diabetes Institute, Melbourne, VIC, 3004, Australia
| | - Nikhil Misra
- Department of Electrical Engineering, Indian Institute of Technology Kanpur, Kalyanpur, Kanpur, Uttar Pradesh, 208016, India
| | - Ankit Shukla
- Department of Pulmonary Medicine, Sanjay Gandhi Post Graduate Institute of Medical Sciences, Raebareli Road, Lucknow, Uttar Pradesh, 226014, India.
- Faculty of Medicine, The University of Queensland, Translational Research Institute, 37 Kent Street, Brisbane, QLD, 4102, Australia.
| |
Collapse
|
7
|
Kukker A, Sharma R, Pandey G, Faseehuddin M. Fuzzy lattices assisted EJAYA Q-learning for automated pulmonary diseases classification. Biomed Phys Eng Express 2024; 10:065001. [PMID: 39178885 DOI: 10.1088/2057-1976/ad72f8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2024] [Accepted: 08/23/2024] [Indexed: 08/26/2024]
Abstract
This work proposes a novel technique called Enhanced JAYA (EJAYA) assisted Q-Learning for the classification of pulmonary diseases, such as pneumonia and tuberculosis (TB) sub-classes using chest x-ray images. The work introduces Fuzzy lattices formation to handle real time (non-linear and non-stationary) data based feature extraction using Schrödinger equation. Features based adaptive classification is made possible through the Q-learning algorithm wherein optimal Q-values selection is done via EJAYA optimization algorithm. Fuzzy lattice is formed using x-ray image pixels and lattice Kinetic Energy (K.E.) is calculated using the Schrödinger equation. Feature vector lattices having highest K.E. have been used as an input features for the classifier. The classifier has been employed for pneumonia classification (normal, mild and severe) and Tuberculosis detection (presence or absence). A total of 3000 images have been used for pneumonia classification yielding an accuracy, sensitivity, specificity, precision and F-scores of 97.90%, 98.43%, 97.25%, 97.78% and 98.10%, respectively. For Tuberculosis 600 samples have been used. The achived accuracy, sensitivity, specificity, precision and F-score are 95.50%, 96.39%, 94.40% 95.52% and 95.95%, respectively. Computational time are 40.96 and 39.98 s for pneumonia and TB classification. Classifier learning rate (training accuracy) for pneumonia classes (normal, mild and severe) are 97.907%, 95.375% and 96.391%, respectively and for tuberculosis (present and absent) are 96.928% and 95.905%, respectively. The results have been compared with contemporary classification techniques which shows superiority of the proposed approach in terms of accuracy and speed of classification. The technique could serve as a fast and accurate tool for automated pneumonia and tuberculosis classification.
Collapse
Affiliation(s)
- Amit Kukker
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, SRM Institute of Science and Technology, Delhi-NCR Campus, Modinagar, Ghaziabad, U.P., 201204, India
| | - Rajneesh Sharma
- ICE Division, Netaji Subhas University of Technology, Delhi, 110078, India
| | - Gaurav Pandey
- ICE Division, Netaji Subhas University of Technology, Delhi, 110078, India
| | - Mohammad Faseehuddin
- Department of Electronics and Communication Engineering, Thapar Institute of Engineering and Technology, Patiala, Punjab, 147004, India
| |
Collapse
|
8
|
Yang L, Lu S, Zhou L. The Implications of Artificial Intelligence on Infection Prevention and Control: Current Progress and Future Perspectives. China CDC Wkly 2024; 6:901-904. [PMID: 39233995 PMCID: PMC11369059 DOI: 10.46234/ccdcw2024.192] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2024] [Accepted: 07/16/2024] [Indexed: 09/06/2024] Open
Affiliation(s)
- Lin Yang
- School of Nursing, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region, China
| | - Shuya Lu
- School of Nursing, The Hong Kong Polytechnic University, Hong Kong Special Administrative Region, China
| | - Lei Zhou
- Chinese Center for Disease Control and Prevention, Beijing, China
| |
Collapse
|
9
|
Patterson BK, Guevara-Coto J, Mora J, Francisco EB, Yogendra R, Mora-Rodríguez RA, Beaty C, Lemaster G, Kaplan DO G, Katz A, Bellanti JA. Long COVID diagnostic with differentiation from chronic lyme disease using machine learning and cytokine hubs. Sci Rep 2024; 14:19743. [PMID: 39187577 PMCID: PMC11347643 DOI: 10.1038/s41598-024-70929-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2024] [Accepted: 08/22/2024] [Indexed: 08/28/2024] Open
Abstract
The absence of a long COVID (LC) or post-acute sequelae of COVID-19 (PASC) diagnostic has profound implications for research and potential therapeutics given the lack of specificity with symptom-based identification of LC and the overlap of symptoms with other chronic inflammatory conditions. Here, we report a machine-learning approach to LC/PASC diagnosis on 347 individuals using cytokine hubs that are also capable of differentiating LC from chronic lyme disease (CLD). We derived decision tree, random forest, and gradient-boosting machine (GBM) classifiers and compared their diagnostic capabilities on a dataset partitioned into training (178 individuals) and evaluation (45 individuals) sets. The GBM model generated 89% sensitivity and 96% specificity for LC with no evidence of overfitting. We tested the GBM on an additional random dataset (106 LC/PASC and 18 Lyme), resulting in high sensitivity (97%) and specificity (90%) for LC. We constructed a Lyme Index confirmatory algorithm to discriminate LC and CLD.
Collapse
Affiliation(s)
- Bruce K Patterson
- IncellDx Inc, 30920 Huntwood Ave, San Carlos, Hayward, CA, 94544, USA.
| | - Jose Guevara-Coto
- IncellDx Inc, 30920 Huntwood Ave, San Carlos, Hayward, CA, 94544, USA
| | - Javier Mora
- Lab of Tumor Chemosensitivity, Faculty of Microbiology, CIET/CICICA, Universidad de Costa Rica, San José, Costa Rica
| | - Edgar B Francisco
- IncellDx Inc, 30920 Huntwood Ave, San Carlos, Hayward, CA, 94544, USA
| | | | - Rodrigo A Mora-Rodríguez
- Lab of Tumor Chemosensitivity, Faculty of Microbiology, CIET/CICICA, Universidad de Costa Rica, San José, Costa Rica
| | - Christopher Beaty
- IncellDx Inc, 30920 Huntwood Ave, San Carlos, Hayward, CA, 94544, USA
| | - Gwyneth Lemaster
- IncellDx Inc, 30920 Huntwood Ave, San Carlos, Hayward, CA, 94544, USA
| | - Gary Kaplan DO
- Department of Community and Family Medicine, Georgetown University School of Medicine, Washington, DC, USA
| | - Amiram Katz
- Neurology Specialist Affiliated With Norwalk Hospital, Orange, CT, USA
| | - Joseph A Bellanti
- Departments of Pediatrics and Microbiology-Immunology, and the International Center for Interdisciplinary Studies of Immunology, Georgetown University Medical Center, Washington, DC, USA
| |
Collapse
|
10
|
Shang Z, Chauhan V, Devi K, Patil S. Artificial Intelligence, the Digital Surgeon: Unravelling Its Emerging Footprint in Healthcare - The Narrative Review. J Multidiscip Healthc 2024; 17:4011-4022. [PMID: 39165254 PMCID: PMC11333562 DOI: 10.2147/jmdh.s482757] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2024] [Accepted: 08/09/2024] [Indexed: 08/22/2024] Open
Abstract
Background Artificial Intelligence (AI) holds transformative potential for the healthcare industry, offering innovative solutions for diagnosis, treatment planning, and improving patient outcomes. As AI continues to be integrated into healthcare systems, it promises advancements across various domains. This review explores the diverse applications of AI in healthcare, along with the challenges and limitations that need to be addressed. The aim is to provide a comprehensive overview of AI's impact on healthcare and to identify areas for further development and focus. Main Applications The review discusses the broad range of AI applications in healthcare. In medical imaging and diagnostics, AI enhances the accuracy and efficiency of diagnostic processes, aiding in early disease detection. AI-powered clinical decision support systems assist healthcare professionals in patient management and decision-making. Predictive analytics using AI enables the prediction of patient outcomes and identification of potential health risks. AI-driven robotic systems have revolutionized surgical procedures, improving precision and outcomes. Virtual assistants and chatbots enhance patient interaction and support, providing timely information and assistance. In the pharmaceutical industry, AI accelerates drug discovery and development by identifying potential drug candidates and predicting their efficacy. Additionally, AI improves administrative efficiency and operational workflows in healthcare, streamlining processes and reducing costs. AI-powered remote monitoring and telehealth solutions expand access to healthcare, particularly in underserved areas. Challenges and Limitations Despite the significant promise of AI in healthcare, several challenges persist. Ensuring the reliability and consistency of AI-driven outcomes is crucial. Privacy and security concerns must be navigated carefully, particularly in handling sensitive patient data. Ethical considerations, including bias and fairness in AI algorithms, need to be addressed to prevent unintended consequences. Overcoming these challenges is critical for the ethical and successful integration of AI in healthcare. Conclusion The integration of AI into healthcare is advancing rapidly, offering substantial benefits in improving patient care and operational efficiency. However, addressing the associated challenges is essential to fully realize the transformative potential of AI in healthcare. Future efforts should focus on enhancing the reliability, transparency, and ethical standards of AI technologies to ensure they contribute positively to global health outcomes.
Collapse
Affiliation(s)
- Zifang Shang
- Guangdong Engineering Technological Research Centre of Clinical Molecular Diagnosis and Antibody Drugs, Meizhou People’s Hospital (Huangtang Hospital), Meizhou Academy of Medical Sciences, Meizhou, People’s Republic of China
| | - Varun Chauhan
- Multi-Disciplinary Research Unit, Government Institute of Medical Sciences, Greater Noida, India
| | - Kirti Devi
- Department of Medicine, Government Institute of Medical Sciences, Greater Noida, India
| | - Sandip Patil
- Department Haematology and Oncology, Shenzhen Children’s Hospital, Shenzhen, People’s Republic of China
| |
Collapse
|
11
|
Sannasi Chakravarthy SR, Bharanidharan N, Vinothini C, Vinoth Kumar V, Mahesh TR, Guluwadi S. Adaptive Mish activation and ranger optimizer-based SEA-ResNet50 model with explainable AI for multiclass classification of COVID-19 chest X-ray images. BMC Med Imaging 2024; 24:206. [PMID: 39123118 PMCID: PMC11313131 DOI: 10.1186/s12880-024-01394-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2024] [Accepted: 08/06/2024] [Indexed: 08/12/2024] Open
Abstract
A recent global health crisis, COVID-19 is a significant global health crisis that has profoundly affected lifestyles. The detection of such diseases from similar thoracic anomalies using medical images is a challenging task. Thus, the requirement of an end-to-end automated system is vastly necessary in clinical treatments. In this way, the work proposes a Squeeze-and-Excitation Attention-based ResNet50 (SEA-ResNet50) model for detecting COVID-19 utilizing chest X-ray data. Here, the idea lies in improving the residual units of ResNet50 using the squeeze-and-excitation attention mechanism. For further enhancement, the Ranger optimizer and adaptive Mish activation function are employed to improve the feature learning of the SEA-ResNet50 model. For evaluation, two publicly available COVID-19 radiographic datasets are utilized. The chest X-ray input images are augmented during experimentation for robust evaluation against four output classes namely normal, pneumonia, lung opacity, and COVID-19. Then a comparative study is done for the SEA-ResNet50 model against VGG-16, Xception, ResNet18, ResNet50, and DenseNet121 architectures. The proposed framework of SEA-ResNet50 together with the Ranger optimizer and adaptive Mish activation provided maximum classification accuracies of 98.38% (multiclass) and 99.29% (binary classification) as compared with the existing CNN architectures. The proposed method achieved the highest Kappa validation scores of 0.975 (multiclass) and 0.98 (binary classification) over others. Furthermore, the visualization of the saliency maps of the abnormal regions is represented using the explainable artificial intelligence (XAI) model, thereby enhancing interpretability in disease diagnosis.
Collapse
Affiliation(s)
- S R Sannasi Chakravarthy
- Department of Electronics and Communication Engineering, Bannari Amman Institute of Technology, Sathyamangalam, India
| | - N Bharanidharan
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, 632014, India
| | - C Vinothini
- Department of Computer Science and Engineering, Dayananda Sagar College of Engineering, Bangalore, India
| | - Venkatesan Vinoth Kumar
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, 632014, India
| | - T R Mahesh
- Department of Computer Science and Engineering, JAIN (Deemed-to-Be University), Bengaluru, 562112, India
| | - Suresh Guluwadi
- Adama Science and Technology University, Adama, 302120, Ethiopia.
| |
Collapse
|
12
|
Alshemaimri BK. Novel Deep CNNs Explore Regions, Boundaries, and Residual Learning for COVID-19 Infection Analysis in Lung CT. Tomography 2024; 10:1205-1221. [PMID: 39195726 PMCID: PMC11359787 DOI: 10.3390/tomography10080091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2024] [Revised: 07/06/2024] [Accepted: 07/17/2024] [Indexed: 08/29/2024] Open
Abstract
COVID-19 poses a global health crisis, necessitating precise diagnostic methods for timely containment. However, accurately delineating COVID-19-affected regions in lung CT scans is challenging due to contrast variations and significant texture diversity. In this regard, this study introduces a novel two-stage classification and segmentation CNN approach for COVID-19 lung radiological pattern analysis. A novel Residual-BRNet is developed to integrate boundary and regional operations with residual learning, capturing key COVID-19 radiological homogeneous regions, texture variations, and structural contrast patterns in the classification stage. Subsequently, infectious CT images undergo lesion segmentation using the newly proposed RESeg segmentation CNN in the second stage. The RESeg leverages both average and max-pooling implementations to simultaneously learn region homogeneity and boundary-related patterns. Furthermore, novel pixel attention (PA) blocks are integrated into RESeg to effectively address mildly COVID-19-infected regions. The evaluation of the proposed Residual-BRNet CNN in the classification stage demonstrates promising performance metrics, achieving an accuracy of 97.97%, F1-score of 98.01%, sensitivity of 98.42%, and MCC of 96.81%. Meanwhile, PA-RESeg in the segmentation phase achieves an optimal segmentation performance with an IoU score of 98.43% and a dice similarity score of 95.96% of the lesion region. The framework's effectiveness in detecting and segmenting COVID-19 lesions highlights its potential for clinical applications.
Collapse
Affiliation(s)
- Bader Khalid Alshemaimri
- Software Engineering Department, College of Computing and Information Sciences, King Saud University, Riyadh 11671, Saudi Arabia
| |
Collapse
|
13
|
Zhang Y, Kohne J, Wittrup E, Najarian K. Three-Stage Framework for Accurate Pediatric Chest X-ray Diagnosis Using Self-Supervision and Transfer Learning on Small Datasets. Diagnostics (Basel) 2024; 14:1634. [PMID: 39125510 PMCID: PMC11312211 DOI: 10.3390/diagnostics14151634] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2024] [Revised: 07/19/2024] [Accepted: 07/25/2024] [Indexed: 08/12/2024] Open
Abstract
Pediatric respiratory disease diagnosis and subsequent treatment require accurate and interpretable analysis. A chest X-ray is the most cost-effective and rapid method for identifying and monitoring various thoracic diseases in children. Recent developments in self-supervised and transfer learning have shown their potential in medical imaging, including chest X-ray areas. In this article, we propose a three-stage framework with knowledge transfer from adult chest X-rays to aid the diagnosis and interpretation of pediatric thorax diseases. We conducted comprehensive experiments with different pre-training and fine-tuning strategies to develop transformer or convolutional neural network models and then evaluate them qualitatively and quantitatively. The ViT-Base/16 model, fine-tuned with the CheXpert dataset, a large chest X-ray dataset, emerged as the most effective, achieving a mean AUC of 0.761 (95% CI: 0.759-0.763) across six disease categories and demonstrating a high sensitivity (average 0.639) and specificity (average 0.683), which are indicative of its strong discriminative ability. The baseline models, ViT-Small/16 and ViT-Base/16, when directly trained on the Pediatric CXR dataset, only achieved mean AUC scores of 0.646 (95% CI: 0.641-0.651) and 0.654 (95% CI: 0.648-0.660), respectively. Qualitatively, our model excels in localizing diseased regions, outperforming models pre-trained on ImageNet and other fine-tuning approaches, thus providing superior explanations. The source code is available online and the data can be obtained from PhysioNet.
Collapse
Affiliation(s)
- Yufeng Zhang
- Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI 48109, USA (K.N.)
| | - Joseph Kohne
- Department of Pediatrics, University of Michigan, Ann Arbor, MI 48103, USA
| | - Emily Wittrup
- Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI 48109, USA (K.N.)
| | - Kayvan Najarian
- Department of Computational Medicine and Bioinformatics, University of Michigan, Ann Arbor, MI 48109, USA (K.N.)
- Michigan Institute for Data Science (MIDAS), University of Michigan, Ann Arbor, MI 48109, USA
- Department of Emergency Medicine, University of Michigan, Ann Arbor, MI 48109, USA
- Max Harry Weil Institute for Critical Care Research and Innovation, University of Michigan, Ann Arbor, MI 48109, USA
| |
Collapse
|
14
|
Xie X, Liu Z, Wang Y, Fu H, Liu M, Zhang Y, Xu J. Puppet Dynasty Recognition System Based on MobileNetV2. ENTROPY (BASEL, SWITZERLAND) 2024; 26:645. [PMID: 39202115 PMCID: PMC11353730 DOI: 10.3390/e26080645] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/07/2024] [Revised: 07/23/2024] [Accepted: 07/24/2024] [Indexed: 09/03/2024]
Abstract
Traditional image classification usually relies on manual feature extraction; however, with the rapid development of artificial intelligence and intelligent vision technology, deep learning models such as CNNs can automatically extract key features from input images to achieve efficient classification. This study focuses on the application of lightweight separable convolutional neural networks in domain-specific image classification tasks. In this paper, we discuss how to use the SSDLite object detection algorithm combined with the MobileNetV2 lightweight convolutional architecture for puppet dynasty recognition from images-a novel and challenging task. By constructing a system that combines object detection and image classification, we aimed to solve the problem of automatic puppet dynasty recognition to reduce manual intervention and improve recognition efficiency and accuracy. We hope that this will have significant implications in the fields of cultural protection and art history research.
Collapse
Affiliation(s)
- Xiaona Xie
- Art College, Northeastern University, No. 11 Lane, Wenhua Road, Heping District, Shenyang 110819, China;
| | - Zeqian Liu
- College of Sciences, Northeastern University, No. 11 Lane, Wenhua Road, Heping District, Shenyang 110819, China; (M.L.); (Y.Z.)
| | - Yuanshuai Wang
- Institute of Artificial Intelligence, Beihang University, No. 37 Xueyuan Road, Haidian District, Beijing 100191, China;
| | - Haoyue Fu
- College of Sciences, Northeastern University, No. 11 Lane, Wenhua Road, Heping District, Shenyang 110819, China; (M.L.); (Y.Z.)
| | - Mengqi Liu
- College of Sciences, Northeastern University, No. 11 Lane, Wenhua Road, Heping District, Shenyang 110819, China; (M.L.); (Y.Z.)
| | - Yingqin Zhang
- College of Sciences, Northeastern University, No. 11 Lane, Wenhua Road, Heping District, Shenyang 110819, China; (M.L.); (Y.Z.)
| | - Jinbo Xu
- School of Mechanical Engineering and Automation, Northeastern University, No. 11 Lane, Wenhua Road, Heping District, Shenyang 110819, China;
| |
Collapse
|
15
|
Siddiqi R, Javaid S. Deep Learning for Pneumonia Detection in Chest X-ray Images: A Comprehensive Survey. J Imaging 2024; 10:176. [PMID: 39194965 DOI: 10.3390/jimaging10080176] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2024] [Revised: 07/15/2024] [Accepted: 07/19/2024] [Indexed: 08/29/2024] Open
Abstract
This paper addresses the significant problem of identifying the relevant background and contextual literature related to deep learning (DL) as an evolving technology in order to provide a comprehensive analysis of the application of DL to the specific problem of pneumonia detection via chest X-ray (CXR) imaging, which is the most common and cost-effective imaging technique available worldwide for pneumonia diagnosis. This paper in particular addresses the key period associated with COVID-19, 2020-2023, to explain, analyze, and systematically evaluate the limitations of approaches and determine their relative levels of effectiveness. The context in which DL is applied as both an aid to and an automated substitute for existing expert radiography professionals, who often have limited availability, is elaborated in detail. The rationale for the undertaken research is provided, along with a justification of the resources adopted and their relevance. This explanatory text and the subsequent analyses are intended to provide sufficient detail of the problem being addressed, existing solutions, and the limitations of these, ranging in detail from the specific to the more general. Indeed, our analysis and evaluation agree with the generally held view that the use of transformers, specifically, vision transformers (ViTs), is the most promising technique for obtaining further effective results in the area of pneumonia detection using CXR images. However, ViTs require extensive further research to address several limitations, specifically the following: biased CXR datasets, data and code availability, the ease with which a model can be explained, systematic methods of accurate model comparison, the notion of class imbalance in CXR datasets, and the possibility of adversarial attacks, the latter of which remains an area of fundamental research.
Collapse
Affiliation(s)
- Raheel Siddiqi
- Computer Science Department, Karachi Campus, Bahria University, Karachi 73500, Pakistan
| | - Sameena Javaid
- Computer Science Department, Karachi Campus, Bahria University, Karachi 73500, Pakistan
| |
Collapse
|
16
|
Park J, Nguyen T, Park S, Hill B, Shadgan B, Gandjbakhche A. Two-Stream Convolutional Neural Networks for Breathing Pattern Classification: Real-Time Monitoring of Respiratory Disease Patients. Bioengineering (Basel) 2024; 11:709. [PMID: 39061791 PMCID: PMC11273486 DOI: 10.3390/bioengineering11070709] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 06/26/2024] [Accepted: 07/11/2024] [Indexed: 07/28/2024] Open
Abstract
A two-stream convolutional neural network (TCNN) for breathing pattern classification has been devised for the continuous monitoring of patients with infectious respiratory diseases. The TCNN consists of a convolutional neural network (CNN)-based autoencoder and classifier. The encoder of the autoencoder generates deep compressed feature maps, which contain the most important information constituting data. These maps are concatenated with feature maps generated by the classifier to classify breathing patterns. The TCNN, single-stream CNN (SCNN), and state-of-the-art classification models were applied to classify four breathing patterns: normal, slow, rapid, and breath holding. The input data consisted of chest tissue hemodynamic responses measured using a wearable near-infrared spectroscopy device on 14 healthy adult participants. Among the classification models evaluated, random forest had the lowest classification accuracy at 88.49%, while the TCNN achieved the highest classification accuracy at 94.63%. In addition, the proposed TCNN performed 2.6% better in terms of classification accuracy than an SCNN (without an autoencoder). Moreover, the TCNN mitigates the issue of declining learning performance with increasing network depth, as observed in the SCNN model. These results prove the robustness of the TCNN in classifying breathing patterns despite using a significantly smaller number of parameters and computations compared to state-of-the-art classification models.
Collapse
Affiliation(s)
- Jinho Park
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.); (B.H.)
| | - Thien Nguyen
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.); (B.H.)
| | - Soongho Park
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.); (B.H.)
- National Heart, Lung and Blood Institute, National Institutes of Health, 10 Center Dr., Bethesda, MD 20892, USA
| | - Brian Hill
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.); (B.H.)
| | - Babak Shadgan
- Implantable Biosensing Laboratory, International Collaboration on Repair Discoveries, Vancouver, BC V5Z 1M9, Canada;
- Department of Pathology & Laboratory Medicine, University of British Columbia, Vancouver, BC V6T 1Z7, Canada
| | - Amir Gandjbakhche
- Eunice Kennedy Shriver National Institute of Child Health and Human Development, National Institutes of Health, 49 Convent Dr., Bethesda, MD 20894, USA; (J.P.); (T.N.); (S.P.); (B.H.)
| |
Collapse
|
17
|
Kadhim YA, Guzel MS, Mishra A. A Novel Hybrid Machine Learning-Based System Using Deep Learning Techniques and Meta-Heuristic Algorithms for Various Medical Datatypes Classification. Diagnostics (Basel) 2024; 14:1469. [PMID: 39061605 PMCID: PMC11275302 DOI: 10.3390/diagnostics14141469] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Revised: 06/27/2024] [Accepted: 07/02/2024] [Indexed: 07/28/2024] Open
Abstract
Medicine is one of the fields where the advancement of computer science is making significant progress. Some diseases require an immediate diagnosis in order to improve patient outcomes. The usage of computers in medicine improves precision and accelerates data processing and diagnosis. In order to categorize biological images, hybrid machine learning, a combination of various deep learning approaches, was utilized, and a meta-heuristic algorithm was provided in this research. In addition, two different medical datasets were introduced, one covering the magnetic resonance imaging (MRI) of brain tumors and the other dealing with chest X-rays (CXRs) of COVID-19. These datasets were introduced to the combination network that contained deep learning techniques, which were based on a convolutional neural network (CNN) or autoencoder, to extract features and combine them with the next step of the meta-heuristic algorithm in order to select optimal features using the particle swarm optimization (PSO) algorithm. This combination sought to reduce the dimensionality of the datasets while maintaining the original performance of the data. This is considered an innovative method and ensures highly accurate classification results across various medical datasets. Several classifiers were employed to predict the diseases. The COVID-19 dataset found that the highest accuracy was 99.76% using the combination of CNN-PSO-SVM. In comparison, the brain tumor dataset obtained 99.51% accuracy, the highest accuracy derived using the combination method of autoencoder-PSO-KNN.
Collapse
Affiliation(s)
- Yezi Ali Kadhim
- College of Engineering, University of Baghdad, Jadriyah, Baghdad 10071, Iraq;
- Department of Modeling and Design of Engineering Systems (MODES), Atilim University, Ankara 06830, Turkey
- Department of Electrical and Electronics Engineering, Atilim University, Incek, Ankara 06830, Turkey
| | - Mehmet Serdar Guzel
- Department of Computer Engineering, Ankara University, Yenimahalle, Ankara 06100, Turkey;
| | - Alok Mishra
- Faculty of Engineering, Norwegian University of Science and Technology, 7034 Trondheim, Norway
- Department of Software Engineering, Atilim University, Incek, Ankara 06830, Turkey
| |
Collapse
|
18
|
Li Y, Xin Y, Li X, Zhang Y, Liu C, Cao Z, Du S, Wang L. Omni-dimensional dynamic convolution feature coordinate attention network for pneumonia classification. Vis Comput Ind Biomed Art 2024; 7:17. [PMID: 38976189 PMCID: PMC11231110 DOI: 10.1186/s42492-024-00168-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2024] [Accepted: 06/22/2024] [Indexed: 07/09/2024] Open
Abstract
Pneumonia is a serious disease that can be fatal, particularly among children and the elderly. The accuracy of pneumonia diagnosis can be improved by combining artificial-intelligence technology with X-ray imaging. This study proposes X-ODFCANet, which addresses the issues of low accuracy and excessive parameters in existing deep-learning-based pneumonia-classification methods. This network incorporates a feature coordination attention module and an omni-dimensional dynamic convolution (ODConv) module, leveraging the residual module for feature extraction from X-ray images. The feature coordination attention module utilizes two one-dimensional feature encoding processes to aggregate feature information from different spatial directions. Additionally, the ODConv module extracts and fuses feature information in four dimensions: the spatial dimension of the convolution kernel, input and output channel quantities, and convolution kernel quantity. The experimental results demonstrate that the proposed method can effectively improve the accuracy of pneumonia classification, which is 3.77% higher than that of ResNet18. The model parameters are 4.45M, which was reduced by approximately 2.5 times. The code is available at https://github.com/limuni/X-ODFCANET .
Collapse
Affiliation(s)
- Yufei Li
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Yufei Xin
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Xinni Li
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Yinrui Zhang
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Cheng Liu
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Zhengwen Cao
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Shaoyi Du
- Department of Ultrasound, the Second Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shaanxi Province, 710004, China.
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, Shaanxi Province, 710049, China.
| | - Lin Wang
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China.
| |
Collapse
|
19
|
Ko J, Park S, Woo HG. Optimization of vision transformer-based detection of lung diseases from chest X-ray images. BMC Med Inform Decis Mak 2024; 24:191. [PMID: 38978027 PMCID: PMC11232177 DOI: 10.1186/s12911-024-02591-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2023] [Accepted: 06/27/2024] [Indexed: 07/10/2024] Open
Abstract
BACKGROUND Recent advances in Vision Transformer (ViT)-based deep learning have significantly improved the accuracy of lung disease prediction from chest X-ray images. However, limited research exists on comparing the effectiveness of different optimizers for lung disease prediction within ViT models. This study aims to systematically evaluate and compare the performance of various optimization methods for ViT-based models in predicting lung diseases from chest X-ray images. METHODS This study utilized a chest X-ray image dataset comprising 19,003 images containing both normal cases and six lung diseases: COVID-19, Viral Pneumonia, Bacterial Pneumonia, Middle East Respiratory Syndrome (MERS), Severe Acute Respiratory Syndrome (SARS), and Tuberculosis. Each ViT model (ViT, FastViT, and CrossViT) was individually trained with each optimization method (Adam, AdamW, NAdam, RAdam, SGDW, and Momentum) to assess their performance in lung disease prediction. RESULTS When tested with ViT on the dataset with balanced-sample sized classes, RAdam demonstrated superior accuracy compared to other optimizers, achieving 95.87%. In the dataset with imbalanced sample size, FastViT with NAdam achieved the best performance with an accuracy of 97.63%. CONCLUSIONS We provide comprehensive optimization strategies for developing ViT-based model architectures, which can enhance the performance of these models for lung disease prediction from chest X-ray images.
Collapse
Affiliation(s)
- Jinsol Ko
- Department of Physiology, Ajou University School of Medicine, Suwon, Republic of Korea
- Department of Biomedical Science, Graduate School, Ajou University, Suwon, Republic of Korea
| | - Soyeon Park
- Ajou University School of Medicine, Suwon, Republic of Korea
| | - Hyun Goo Woo
- Department of Physiology, Ajou University School of Medicine, Suwon, Republic of Korea.
- Department of Biomedical Science, Graduate School, Ajou University, Suwon, Republic of Korea.
| |
Collapse
|
20
|
Hao Z, Wang Y, Li F, Ding G, Gao Y. mmWave-RM: A Respiration Monitoring and Pattern Classification System Based on mmWave Radar. SENSORS (BASEL, SWITZERLAND) 2024; 24:4315. [PMID: 39001094 PMCID: PMC11243972 DOI: 10.3390/s24134315] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Revised: 06/11/2024] [Accepted: 07/01/2024] [Indexed: 07/16/2024]
Abstract
Breathing is one of the body's most basic functions and abnormal breathing can indicate underlying cardiopulmonary problems. Monitoring respiratory abnormalities can help with early detection and reduce the risk of cardiopulmonary diseases. In this study, a 77 GHz frequency-modulated continuous wave (FMCW) millimetre-wave (mmWave) radar was used to detect different types of respiratory signals from the human body in a non-contact manner for respiratory monitoring (RM). To solve the problem of noise interference in the daily environment on the recognition of different breathing patterns, the system utilised breathing signals captured by the millimetre-wave radar. Firstly, we filtered out most of the static noise using a signal superposition method and designed an elliptical filter to obtain a more accurate image of the breathing waveforms between 0.1 Hz and 0.5 Hz. Secondly, combined with the histogram of oriented gradient (HOG) feature extraction algorithm, K-nearest neighbours (KNN), convolutional neural network (CNN), and HOG support vector machine (G-SVM) were used to classify four breathing modes, namely, normal breathing, slow and deep breathing, quick breathing, and meningitic breathing. The overall accuracy reached up to 94.75%. Therefore, this study effectively supports daily medical monitoring.
Collapse
Affiliation(s)
- Zhanjun Hao
- College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China; (Y.W.); (F.L.); (G.D.); (Y.G.)
- Gansu Province Internet of Things Engineering Research Center, Lanzhou 730070, China
| | - Yue Wang
- College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China; (Y.W.); (F.L.); (G.D.); (Y.G.)
| | - Fenfang Li
- College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China; (Y.W.); (F.L.); (G.D.); (Y.G.)
| | - Guozhen Ding
- College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China; (Y.W.); (F.L.); (G.D.); (Y.G.)
| | - Yifei Gao
- College of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China; (Y.W.); (F.L.); (G.D.); (Y.G.)
| |
Collapse
|
21
|
Fernández-Miranda PM, Fraguela EM, de Linera-Alperi MÁ, Cobo M, Del Barrio AP, González DR, Vega JA, Iglesias LL. A retrospective study of deep learning generalization across two centers and multiple models of X-ray devices using COVID-19 chest-X rays. Sci Rep 2024; 14:14657. [PMID: 38918499 PMCID: PMC11199585 DOI: 10.1038/s41598-024-64941-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 06/14/2024] [Indexed: 06/27/2024] Open
Abstract
Generalization of deep learning (DL) algorithms is critical for the secure implementation of computer-aided diagnosis systems in clinical practice. However, broad generalization remains to be a challenge in machine learning. This research aims to identify and study potential factors that can affect the internal validation and generalization of DL networks, namely the institution where the images come from, the image processing applied by the X-ray device, and the type of response function of the X-ray device. For these purposes, a pre-trained convolutional neural network (CNN) (VGG16) was trained three times for classifying COVID-19 and control chest radiographs with the same hyperparameters, but using different combinations of data acquired in two institutions by three different X-ray device manufacturers. Regarding internal validation, the addition of images from an external institution to the training set did not modify the algorithm's internal performance, however, the inclusion of images acquired by a device from a different manufacturer decreased the performance up to 8% (p < 0.05). In contrast, generalization across institutions and X-ray devices with the same type of response function was achieved. Nonetheless, generalization was not observed across devices with different types of response function. This factor was the key impediment to achieving broad generalization in our research, followed by the device's image-processing and the inter-institutional differences, which both reduced generalization performance to 18.9% (p < 0.05), and 9.8% (p < 0.05), respectively. Finally, clustering analysis with features extracted by the CNN was performed, revealing a substantial dependence of feature values extracted by the pre-trained CNN on the X-ray device which acquired the images.
Collapse
Affiliation(s)
- Pablo Menéndez Fernández-Miranda
- Departamento de Radiología, Hospital Universitario Rey Juan Carlos, Calle Gladiolo, s/n, 28933, Móstoles, Spain
- Departamento de Tecnologías de La Información, Universidad CEU San Pablo, Calle Julián Romea, 22, 28003, Madrid, Spain
| | - Enrique Marqués Fraguela
- Departamento de Radiofísica y Protección Radiológica, Hospital Universitario Marqués de Valdecilla, Avenida de Valdecilla s/n, Santander, Spain
| | - Marta Álvarez de Linera-Alperi
- Departamento de Otorrinolaringología, Clínica Universidad de Navarra, Calle del Marquesado de Santa Marta, 1, 28027, Madrid, Spain
| | - Miriam Cobo
- Advanced Computing and E-Science Research Group, Institute of Physics of Cantabria, Grupo de Computación y E-Ciencia, CSIC-UC, IFCA-CSIC, Avenida de los Castros s/n, 39005, Santander, Spain.
| | - Amaia Pérez Del Barrio
- Servicio de Radiología, Complejo Hospitalario de Navarra, C. de Irunlarrea, 3, 31008, Pamplona, Spain
| | - David Rodríguez González
- Advanced Computing and E-Science Research Group, Institute of Physics of Cantabria, Grupo de Computación y E-Ciencia, CSIC-UC, IFCA-CSIC, Avenida de los Castros s/n, 39005, Santander, Spain
| | - José A Vega
- Departamento de Morfología y Biología Celular, Grupo SINPOS, Universidad de Oviedo, Avenida Julián Clavería, 6, 33006, Oviedo, Spain
- Facultad de Ciencias de La Salud, Universidad Autónoma de Chile, Avenida Pedro de Valdivía, 425, 751 1185, Providencia-Santiago de Chile, Chile
| | - Lara Lloret Iglesias
- Advanced Computing and E-Science Research Group, Institute of Physics of Cantabria, Grupo de Computación y E-Ciencia, CSIC-UC, IFCA-CSIC, Avenida de los Castros s/n, 39005, Santander, Spain
| |
Collapse
|
22
|
Talib MA, Afadar Y, Nasir Q, Nassif AB, Hijazi H, Hasasneh A. A tree-based explainable AI model for early detection of Covid-19 using physiological data. BMC Med Inform Decis Mak 2024; 24:179. [PMID: 38915001 PMCID: PMC11194929 DOI: 10.1186/s12911-024-02576-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 06/13/2024] [Indexed: 06/26/2024] Open
Abstract
With the outbreak of COVID-19 in 2020, countries worldwide faced significant concerns and challenges. Various studies have emerged utilizing Artificial Intelligence (AI) and Data Science techniques for disease detection. Although COVID-19 cases have declined, there are still cases and deaths around the world. Therefore, early detection of COVID-19 before the onset of symptoms has become crucial in reducing its extensive impact. Fortunately, wearable devices such as smartwatches have proven to be valuable sources of physiological data, including Heart Rate (HR) and sleep quality, enabling the detection of inflammatory diseases. In this study, we utilize an already-existing dataset that includes individual step counts and heart rate data to predict the probability of COVID-19 infection before the onset of symptoms. We train three main model architectures: the Gradient Boosting classifier (GB), CatBoost trees, and TabNet classifier to analyze the physiological data and compare their respective performances. We also add an interpretability layer to our best-performing model, which clarifies prediction results and allows a detailed assessment of effectiveness. Moreover, we created a private dataset by gathering physiological data from Fitbit devices to guarantee reliability and avoid bias.The identical set of models was then applied to this private dataset using the same pre-trained models, and the results were documented. Using the CatBoost tree-based method, our best-performing model outperformed previous studies with an accuracy rate of 85% on the publicly available dataset. Furthermore, this identical pre-trained CatBoost model produced an accuracy of 81% when applied to the private dataset. You will find the source code in the link: https://github.com/OpenUAE-LAB/Covid-19-detection-using-Wearable-data.git .
Collapse
Affiliation(s)
- Manar Abu Talib
- Department of Computer Science, College of Computing and Informatics, University of Sharjah, P.O. Box 27272, Sharjah, UAE.
| | - Yaman Afadar
- Department of Computer Engineering, College of Computing and Informatics, University of Sharjah, Sharjah, UAE
| | - Qassim Nasir
- Department of Computer Engineering, College of Computing and Informatics, University of Sharjah, Sharjah, UAE
| | - Ali Bou Nassif
- Department of Computer Engineering, College of Computing and Informatics, University of Sharjah, Sharjah, UAE
| | - Haytham Hijazi
- Centre for Informatics and Systems of the University of Coimbra (CISUC), University of Coimbra, Coimbra, P-3030-290, Portugal
- Intelligent Systems Department, Ahliya University, Bethlehem, P-150-199, Palestine
| | - Ahmad Hasasneh
- Department of Natural, Engineering and Technology Sciences, Faculty of Graduate Studies, Arab American University, P.O. Box 240, Ramallah, Palestine
| |
Collapse
|
23
|
Wang S, Ren J, Guo X. A high-accuracy lightweight network model for X-ray image diagnosis: A case study of COVID detection. PLoS One 2024; 19:e0303049. [PMID: 38889106 PMCID: PMC11185471 DOI: 10.1371/journal.pone.0303049] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 04/15/2024] [Indexed: 06/20/2024] Open
Abstract
The Coronavirus Disease 2019(COVID-19) has caused widespread and significant harm globally. In order to address the urgent demand for a rapid and reliable diagnostic approach to mitigate transmission, the application of deep learning stands as a viable solution. The impracticality of many existing models is attributed to excessively large parameters, significantly limiting their utility. Additionally, the classification accuracy of the model with few parameters falls short of desirable levels. Motivated by this observation, the present study employs the lightweight network MobileNetV3 as the underlying architecture. This paper incorporates the dense block to capture intricate spatial information in images, as well as the transition layer designed to reduce the size and channel number of the feature map. Furthermore, this paper employs label smoothing loss to address the inter-class similarity effects and uses class weighting to tackle the problem of data imbalance. Additionally, this study applies the pruning technique to eliminate unnecessary structures and further reduce the number of parameters. As a result, this improved model achieves an impressive 98.71% accuracy on an openly accessible database, while utilizing only 5.94 million parameters. Compared to the previous method, this maximum improvement reaches 5.41%. Moreover, this research successfully reduces the parameter count by up to 24 times, showcasing the efficacy of our approach. This demonstrates the significant benefits in regions with limited availability of medical resources.
Collapse
Affiliation(s)
- Shujuan Wang
- College of Mathematics and Information Science, Zhengzhou University of Light Industry, Zhengzhou, China
| | - Jialin Ren
- College of Mathematics and Information Science, Zhengzhou University of Light Industry, Zhengzhou, China
| | - Xiaoli Guo
- College of Mathematics and Information Science, Zhengzhou University of Light Industry, Zhengzhou, China
| |
Collapse
|
24
|
Nair SSK, David LR, Shariff A, Maskari SA, Mawali AA, Weis S, Fouad T, Ozsahin DU, Alshuweihi A, Obaideen A, Elshami W. CovMediScanX: A medical imaging solution for COVID-19 diagnosis from chest X-ray images. J Med Imaging Radiat Sci 2024; 55:272-280. [PMID: 38594085 DOI: 10.1016/j.jmir.2024.03.046] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2023] [Revised: 02/17/2024] [Accepted: 03/19/2024] [Indexed: 04/11/2024]
Abstract
INTRODUCTION Radiologists have extensively employed the interpretation of chest X-rays (CXR) to identify visual markers indicative of COVID-19 infection, offering an alternative approach for the screening of infected individuals. This research article presents CovMediScanX, a deep learning-based framework designed for a rapid and automated diagnosis of COVID-19 from CXR scan images. METHODS The proposed approach encompasses gathering and preprocessing CXR image datasets, training deep learning-based custom-made Convolutional Neural Network (CNN), pre-trained and hybrid transfer learning models, identifying the highest-performing model based on key evaluation metrics, and embedding this model into a web interface called CovMediScanX, designed for radiologists to detect the COVID-19 status in new CXR images. RESULTS The custom-made CNN model obtained a remarkable testing accuracy of 94.32% outperforming other models. CovMediScanX, employing the custom-made CNN underwent evaluation with an independent dataset also. The images in the independent dataset are sourced from a scanning machine that is entirely different from those used for the training dataset, highlighting a clear distinction of datasets in their origins. The evaluation outcome highlighted the framework's capability to accurately detect COVID-19 cases, showcasing encouraging results with a precision of 73% and a recall of 84% for positive cases. However, the model requires further enhancement, particularly in improving its detection of normal cases, as evidenced by lower precision and recall rates. CONCLUSION The research proposes CovMediScanX framework that demonstrates promising potential in automatically identifying COVID-19 cases from CXR images. While the model's overall performance on independent data needs improvement, it is evident that addressing bias through the inclusion of diverse data sources during training could further enhance accuracy and reliability.
Collapse
Affiliation(s)
| | - Leena R David
- Department of Medical Diagnostic Imaging, College of Health Sciences, University of Sharjah, United Arab Emirates.
| | - Abdulwahid Shariff
- Department of Postgraduate Studies, University of Dar es Salaam, Tanzania
| | - Saqar Al Maskari
- Department of Computing and Electronics Engineering, Middle East College, Sultanate of Oman
| | - Adhra Al Mawali
- Quality Assurance and Planning, German University of Technology (GUtech), Sultanate of Oman
| | - Sammy Weis
- University Hospital, Sharjah, United Arab Emirates
| | - Taha Fouad
- University Hospital, Sharjah, United Arab Emirates
| | - Dilber Uzun Ozsahin
- Department of Medical Diagnostic Imaging, College of Health Sciences, University of Sharjah, United Arab Emirates
| | | | | | - Wiam Elshami
- Department of Medical Diagnostic Imaging, College of Health Sciences, University of Sharjah, United Arab Emirates
| |
Collapse
|
25
|
Qiu Y, Liu Y, Li S, Xu J. MiniSeg: An Extremely Minimum Network Based on Lightweight Multiscale Learning for Efficient COVID-19 Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:8570-8584. [PMID: 37015641 DOI: 10.1109/tnnls.2022.3230821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
The rapid spread of the new pandemic, i.e., coronavirus disease 2019 (COVID-19), has severely threatened global health. Deep-learning-based computer-aided screening, e.g., COVID-19 infected area segmentation from computed tomography (CT) image, has attracted much attention by serving as an adjunct to increase the accuracy of COVID-19 screening and clinical diagnosis. Although lesion segmentation is a hot topic, traditional deep learning methods are usually data-hungry with millions of parameters, easy to overfit under limited available COVID-19 training data. On the other hand, fast training/testing and low computational cost are also necessary for quick deployment and development of COVID-19 screening systems, but traditional methods are usually computationally intensive. To address the above two problems, we propose MiniSeg, a lightweight model for efficient COVID-19 segmentation from CT images. Our efforts start with the design of an attentive hierarchical spatial pyramid (AHSP) module for lightweight, efficient, effective multiscale learning that is essential for image segmentation. Then, we build a two-path (TP) encoder for deep feature extraction, where one path uses AHSP modules for learning multiscale contextual features and the other is a shallow convolutional path for capturing fine details. The two paths interact with each other for learning effective representations. Based on the extracted features, a simple decoder is added for COVID-19 segmentation. For comparing MiniSeg to previous methods, we build a comprehensive COVID-19 segmentation benchmark. Extensive experiments demonstrate that the proposed MiniSeg achieves better accuracy because its only 83k parameters make it less prone to overfitting. Its high efficiency also makes it easy to deploy and develop. The code has been released at https://github.com/yun-liu/MiniSeg.
Collapse
|
26
|
Parikh V, Tariq A, Patel B, Banerjee I. Comparative Analysis of Fusion Strategies for Imaging and Non-imaging Data - Use-case of Hospital Discharge Prediction. AMIA JOINT SUMMITS ON TRANSLATIONAL SCIENCE PROCEEDINGS. AMIA JOINT SUMMITS ON TRANSLATIONAL SCIENCE 2024; 2024:652-661. [PMID: 38827051 PMCID: PMC11141810] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Subscribe] [Scholar Register] [Indexed: 06/04/2024]
Abstract
Accurate prediction of future clinical events such as discharge from hospital can not only improve hospital resource management but also provide an indicator of a patient's clinical condition. Within the scope of this work, we perform a comparative analysis of deep learning based fusion strategies against traditional single source models for prediction of discharge from hospital by fusing information encoded in two diverse but relevant data modalities, i.e., chest X-ray images and tabular electronic health records (EHR). We evaluate multiple fusion strategies including late, early and joint fusion in terms of their efficacy for target prediction compared to EHR-only and Image-only predictive models. Results indicated the importance of merging information from two modalities for prediction as fusion models tended to outperform single modality models and indicate that the joint fusion scheme was the most effective for target prediction. Joint fusion model merges the two modalities through a branched neural network that is jointly trained in an end-to-end fashion to extract target-relevant information from both modalities.
Collapse
|
27
|
Ying X, Peng H, Xie J. Big data analysis for Covid-19 in hospital information systems. PLoS One 2024; 19:e0294481. [PMID: 38776299 PMCID: PMC11111070 DOI: 10.1371/journal.pone.0294481] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2023] [Accepted: 10/31/2023] [Indexed: 05/24/2024] Open
Abstract
The COVID-19 pandemic has triggered a global public health crisis, affecting hundreds of countries. With the increasing number of infected cases, developing automated COVID-19 identification tools based on CT images can effectively assist clinical diagnosis and reduce the tedious workload of image interpretation. To expand the dataset for machine learning methods, it is necessary to aggregate cases from different medical systems to learn robust and generalizable models. This paper proposes a novel deep learning joint framework that can effectively handle heterogeneous datasets with distribution discrepancies for accurate COVID-19 identification. We address the cross-site domain shift by redesigning the COVID-Net's network architecture and learning strategy, and independent feature normalization in latent space to improve prediction accuracy and learning efficiency. Additionally, we propose using a contrastive training objective to enhance the domain invariance of semantic embeddings and boost classification performance on each dataset. We develop and evaluate our method with two large-scale public COVID-19 diagnosis datasets containing CT images. Extensive experiments show that our method consistently improves the performance both datasets, outperforming the original COVID-Net trained on each dataset by 13.27% and 15.15% in AUC respectively, also exceeding existing state-of-the-art multi-site learning methods.
Collapse
Affiliation(s)
- Xinpa Ying
- Hospital of Chengdu University of TCM, Chengdu, Sichuan, China
| | - Haiyang Peng
- Hospital of Chengdu University of TCM, Chengdu, Sichuan, China
| | - Jun Xie
- Hospital of Chengdu University of TCM, Chengdu, Sichuan, China
| |
Collapse
|
28
|
Harkness R, Frangi AF, Zucker K, Ravikumar N. Multi-centre benchmarking of deep learning models for COVID-19 detection in chest x-rays. FRONTIERS IN RADIOLOGY 2024; 4:1386906. [PMID: 38836218 PMCID: PMC11148230 DOI: 10.3389/fradi.2024.1386906] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 05/07/2024] [Indexed: 06/06/2024]
Abstract
Introduction This study is a retrospective evaluation of the performance of deep learning models that were developed for the detection of COVID-19 from chest x-rays, undertaken with the goal of assessing the suitability of such systems as clinical decision support tools. Methods Models were trained on the National COVID-19 Chest Imaging Database (NCCID), a UK-wide multi-centre dataset from 26 different NHS hospitals and evaluated on independent multi-national clinical datasets. The evaluation considers clinical and technical contributors to model error and potential model bias. Model predictions are examined for spurious feature correlations using techniques for explainable prediction. Results Models performed adequately on NHS populations, with performance comparable to radiologists, but generalised poorly to international populations. Models performed better in males than females, and performance varied across age groups. Alarmingly, models routinely failed when applied to complex clinical cases with confounding pathologies and when applied to radiologist defined "mild" cases. Discussion This comprehensive benchmarking study examines the pitfalls in current practices that have led to impractical model development. Key findings highlight the need for clinician involvement at all stages of model development, from data curation and label definition, to model evaluation, to ensure that all clinical factors and disease features are appropriately considered during model design. This is imperative to ensure automated approaches developed for disease detection are fit-for-purpose in a clinical setting.
Collapse
Affiliation(s)
- Rachael Harkness
- School of Computing, University of Leeds, Leeds, United Kingdom
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, Leeds, United Kingdom
| | - Alejandro F Frangi
- Division of Informatics, Imaging, and Data Sciences, School of Health Sciences, University of Manchester, Manchester, United Kingdom
- Department of Computer Science, School of Engineering, University of Manchester, Manchester, United Kingdom
| | - Kieran Zucker
- Leeds Institute of Medical Research, School of Medicine, University of Leeds, Leeds, United Kingdom
| | - Nishant Ravikumar
- School of Computing, University of Leeds, Leeds, United Kingdom
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, Leeds, United Kingdom
| |
Collapse
|
29
|
Farghaly O, Deshpande P. Texture-Based Classification to Overcome Uncertainty between COVID-19 and Viral Pneumonia Using Machine Learning and Deep Learning Techniques. Diagnostics (Basel) 2024; 14:1017. [PMID: 38786315 PMCID: PMC11119936 DOI: 10.3390/diagnostics14101017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Revised: 05/11/2024] [Accepted: 05/13/2024] [Indexed: 05/25/2024] Open
Abstract
The SARS-CoV-2 virus, responsible for COVID-19, often manifests symptoms akin to viral pneumonia, complicating early detection and potentially leading to severe COVID pneumonia and long-term effects. Particularly affecting young individuals, the elderly, and those with weakened immune systems, the accurate classification of COVID-19 poses challenges, especially with highly dimensional image data. Past studies have faced limitations due to simplistic algorithms and small, biased datasets, yielding inaccurate results. In response, our study introduces a novel classification model that integrates advanced texture feature extraction methods, including GLCM, GLDM, and wavelet transform, within a deep learning framework. This innovative approach enables the effective classification of chest X-ray images into normal, COVID-19, and viral pneumonia categories, overcoming the limitations encountered in previous studies. Leveraging the unique textures inherent to each dataset class, our model achieves superior classification performance, even amidst the complexity and diversity of the data. Moreover, we present comprehensive numerical findings demonstrating the superiority of our approach over traditional methods. The numerical results highlight the accuracy (random forest (RF): 0.85; SVM (support vector machine): 0.70; deep learning neural network (DLNN): 0.92), recall (RF: 0.85, SVM: 0.74, DLNN: 0.93), precision (RF: 0.86, SVM: 0.71, DLNN: 0.87), and F1-Score (RF: 0.86, SVM: 0.72, DLNN: 0.89) of our proposed model. Our study represents a significant advancement in AI-based diagnostic systems for COVID-19 and pneumonia, promising improved patient outcomes and healthcare management strategies.
Collapse
Affiliation(s)
- Omar Farghaly
- Data-Intensive Computing Distributed Systems Laboratory, Department of Electrical and Computer Engineering, Marquette University, Milwaukee, WI 53233, USA
| | - Priya Deshpande
- Data-Intensive Computing Distributed Systems Laboratory, Department of Electrical and Computer Engineering, Marquette University, Milwaukee, WI 53233, USA
| |
Collapse
|
30
|
Bennour A, Ben Aoun N, Khalaf OI, Ghabban F, Wong WK, Algburi S. Contribution to pulmonary diseases diagnostic from X-ray images using innovative deep learning models. Heliyon 2024; 10:e30308. [PMID: 38707425 PMCID: PMC11068804 DOI: 10.1016/j.heliyon.2024.e30308] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 04/09/2024] [Accepted: 04/23/2024] [Indexed: 05/07/2024] Open
Abstract
Pulmonary disease identification and characterization are among the most intriguing research topics of recent years since they require an accurate and prompt diagnosis. Although pulmonary radiography has helped in lung disease diagnosis, the interpretation of the radiographic image has always been a major concern for doctors and radiologists to reduce diagnosis errors. Due to their success in image classification and segmentation tasks, cutting-edge artificial intelligence techniques like machine learning (ML) and deep learning (DL) are widely encouraged to be applied in the field of diagnosing lung disorders and identifying them using medical images, particularly radiographic ones. For this end, the researchers are concurring to build systems based on these techniques in particular deep learning ones. In this paper, we proposed three deep-learning models that were trained to identify the presence of certain lung diseases using thoracic radiography. The first model, named "CovCXR-Net", identifies the COVID-19 disease (two cases: COVID-19 or normal). The second model, named "MDCXR3-Net", identifies the COVID-19 and pneumonia diseases (three cases: COVID-19, pneumonia, or normal), and the last model, named "MDCXR4-Net", is destined to identify the COVID-19, pneumonia and the pulmonary opacity diseases (4 cases: COVID-19, pneumonia, pulmonary opacity or normal). These models have proven their superiority in comparison with the state-of-the-art models and reached an accuracy of 99,09 %, 97.74 %, and 90,37 % respectively with three benchmarks.
Collapse
Affiliation(s)
- Akram Bennour
- LAMIS Laboratiry, Echahid Cheikh Larbi Tebessi University, Tebessa, Algeria
| | - Najib Ben Aoun
- College of Computer Science and Information Technology, Al-Baha University, Al Baha, Saudi Arabia
- REGIM-Lab: Research Groups in Intelligent Machines, National School of Engineers of Sfax (ENIS), University of Sfax, Tunisia
| | - Osamah Ibrahim Khalaf
- Department of Solar, Al-Nahrain Research Center for Renewable Energy, Al-Nahrain University, Jadriya, Baghdad, Iraq
| | - Fahad Ghabban
- College of Computer Science and Engineering, Taibah University, Medina, Saudi Arabia
| | | | - Sameer Algburi
- Al-Kitab University, College of Engineering Techniques, Kirkuk, Iraq
| |
Collapse
|
31
|
Wang L, Wang Q, Wang X, Ma Y, Zhang L, Liu M. Triplet-constrained deep hashing for chest X-ray image retrieval in COVID-19 assessment. Neural Netw 2024; 173:106182. [PMID: 38387203 DOI: 10.1016/j.neunet.2024.106182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Revised: 01/15/2024] [Accepted: 02/15/2024] [Indexed: 02/24/2024]
Abstract
Radiology images of the chest, such as computer tomography scans and X-rays, have been prominently used in computer-aided COVID-19 analysis. Learning-based radiology image retrieval has attracted increasing attention recently, which generally involves image feature extraction and finding matches in extensive image databases based on query images. Many deep hashing methods have been developed for chest radiology image search due to the high efficiency of retrieval using hash codes. However, they often overlook the complex triple associations between images; that is, images belonging to the same category tend to share similar characteristics and vice versa. To this end, we develop a triplet-constrained deep hashing (TCDH) framework for chest radiology image retrieval to facilitate automated analysis of COVID-19. The TCDH consists of two phases, including (a) feature extraction and (b) image retrieval. For feature extraction, we have introduced a triplet constraint and an image reconstruction task to enhance discriminative ability of learned features, and these features are then converted into binary hash codes to capture semantic information. Specifically, the triplet constraint is designed to pull closer samples within the same category and push apart samples from different categories. Additionally, an auxiliary image reconstruction task is employed during feature extraction to help effectively capture anatomical structures of images. For image retrieval, we utilize learned hash codes to conduct searches for medical images. Extensive experiments on 30,386 chest X-ray images demonstrate the superiority of the proposed method over several state-of-the-art approaches in automated image search. The code is now available online.
Collapse
Affiliation(s)
- Linmin Wang
- School of Mathematics Science, Liaocheng University, Liaocheng, Shandong, 252000, China
| | - Qianqian Wang
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA
| | - Xiaochuan Wang
- School of Mathematics Science, Liaocheng University, Liaocheng, Shandong, 252000, China
| | - Yunling Ma
- School of Mathematics Science, Liaocheng University, Liaocheng, Shandong, 252000, China
| | - Limei Zhang
- School of Computer Science and Technology, Shandong Jianzhu University, Jinan, Shandong, 250101, China.
| | - Mingxia Liu
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA.
| |
Collapse
|
32
|
Lee YM, Stretton B, Tan S, Gupta A, Kovoor J, Bacchi S, Lim W, Chan WO. Captive markets and medical artificial intelligence. J Med Imaging Radiat Oncol 2024; 68:278-281. [PMID: 38563301 DOI: 10.1111/1754-9485.13648] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Accepted: 03/21/2024] [Indexed: 04/04/2024]
Affiliation(s)
- Yong Min Lee
- Ophthalmology Department, Royal Adelaide Hospital, Adelaide, South Australia, Australia
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
| | - Brandon Stretton
- Ophthalmology Department, Royal Adelaide Hospital, Adelaide, South Australia, Australia
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
| | - Sheryn Tan
- Ophthalmology Department, Royal Adelaide Hospital, Adelaide, South Australia, Australia
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
| | - Aashray Gupta
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
- Gold Coast University Hospital, Gold Coast, Queensland, Australia
| | - Joshua Kovoor
- Ophthalmology Department, Royal Adelaide Hospital, Adelaide, South Australia, Australia
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
| | - Stephen Bacchi
- Ophthalmology Department, Royal Adelaide Hospital, Adelaide, South Australia, Australia
- Flinders University, Adelaide, South Australia, Australia
| | - Wanyin Lim
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
- Department of Radiology, Royal Adelaide Hospital, Adelaide, South Australia, Australia
| | - Weng Onn Chan
- Ophthalmology Department, Royal Adelaide Hospital, Adelaide, South Australia, Australia
- Faculty of Health and Medical Sciences, The University of Adelaide, Adelaide, South Australia, Australia
| |
Collapse
|
33
|
Zhang Z, Chen B, Luo Y. A Deep Ensemble Dynamic Learning Network for Corona Virus Disease 2019 Diagnosis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:3912-3926. [PMID: 36054386 DOI: 10.1109/tnnls.2022.3201198] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Corona virus disease 2019 is an extremely fatal pandemic around the world. Intelligently recognizing X-ray chest radiography images for automatically identifying corona virus disease 2019 from other types of pneumonia and normal cases provides clinicians with tremendous conveniences in diagnosis process. In this article, a deep ensemble dynamic learning network is proposed. After a chain of image preprocessing steps and the division of image dataset, convolution blocks and the final average pooling layer are pretrained as a feature extractor. For classifying the extracted feature samples, two-stage bagging dynamic learning network is trained based on neural dynamic learning and bagging algorithms, which diagnoses the presence and types of pneumonia successively. Experimental results manifest that using the proposed deep ensemble dynamic learning network obtains 98.7179% diagnosis accuracy, which indicates more excellent diagnosis effect than existing state-of-the-art models on the open image dataset. Such accurate diagnosis effects provide convincing evidences for further detections and treatments.
Collapse
|
34
|
Pan CT, Kumar R, Wen ZH, Wang CH, Chang CY, Shiue YL. Improving Respiratory Infection Diagnosis with Deep Learning and Combinatorial Fusion: A Two-Stage Approach Using Chest X-ray Imaging. Diagnostics (Basel) 2024; 14:500. [PMID: 38472972 DOI: 10.3390/diagnostics14050500] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2023] [Revised: 02/16/2024] [Accepted: 02/18/2024] [Indexed: 03/14/2024] Open
Abstract
The challenges of respiratory infections persist as a global health crisis, placing substantial stress on healthcare infrastructures and necessitating ongoing investigation into efficacious treatment modalities. The persistent challenge of respiratory infections, including COVID-19, underscores the critical need for enhanced diagnostic methodologies to support early treatment interventions. This study introduces an innovative two-stage data analytics framework that leverages deep learning algorithms through a strategic combinatorial fusion technique, aimed at refining the accuracy of early-stage diagnosis of such infections. Utilizing a comprehensive dataset compiled from publicly available lung X-ray images, the research employs advanced pre-trained deep learning models to navigate the complexities of disease classification, addressing inherent data imbalances through methodical validation processes. The core contribution of this work lies in its novel application of combinatorial fusion, integrating select models to significantly elevate diagnostic precision. This approach not only showcases the adaptability and strength of deep learning in navigating the intricacies of medical imaging but also marks a significant step forward in the utilization of artificial intelligence to improve outcomes in healthcare diagnostics. The study's findings illuminate the path toward leveraging technological advancements in enhancing diagnostic accuracies, ultimately contributing to the timely and effective treatment of respiratory diseases.
Collapse
Affiliation(s)
- Cheng-Tang Pan
- Department of Mechanical and Electro-Mechanical Engineering, National Sun Yat-sen University, Kaohsiung 804, Taiwan
- Institute of Precision Medicine, National Sun Yat-sen University, Kaohsiung 804, Taiwan
- Taiwan Instrument Research Institute, National Applied Research Laboratories, Hsinchu 300, Taiwan
- Institute of Advanced Semiconductor Packaging and Testing, College of Semiconductor and Advanced Technology Research, National Sun Yat-sen University, Kaohsiung 804, Taiwan
| | - Rahul Kumar
- Department of Mechanical and Electro-Mechanical Engineering, National Sun Yat-sen University, Kaohsiung 804, Taiwan
| | - Zhi-Hong Wen
- Department of Marine Biotechnology and Research, National Sun Yat-sen University, Kaohsiung 804, Taiwan
| | - Chih-Hsuan Wang
- Division of Nephrology and Metabolism, Department of Internal Medicine, Kaohsiung Armed Forces General Hospital, Kaohsiung 804, Taiwan
- Institute of Medical Science and Technology, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| | - Chun-Yung Chang
- Division of Nephrology and Metabolism, Department of Internal Medicine, Kaohsiung Armed Forces General Hospital, Kaohsiung 804, Taiwan
- Institute of Medical Science and Technology, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| | - Yow-Ling Shiue
- Institute of Precision Medicine, National Sun Yat-sen University, Kaohsiung 804, Taiwan
- Institute of Biomedical Sciences, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| |
Collapse
|
35
|
Hassan J, Saeed SM, Deka L, Uddin MJ, Das DB. Applications of Machine Learning (ML) and Mathematical Modeling (MM) in Healthcare with Special Focus on Cancer Prognosis and Anticancer Therapy: Current Status and Challenges. Pharmaceutics 2024; 16:260. [PMID: 38399314 PMCID: PMC10892549 DOI: 10.3390/pharmaceutics16020260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2023] [Revised: 01/29/2024] [Accepted: 02/07/2024] [Indexed: 02/25/2024] Open
Abstract
The use of data-driven high-throughput analytical techniques, which has given rise to computational oncology, is undisputed. The widespread use of machine learning (ML) and mathematical modeling (MM)-based techniques is widely acknowledged. These two approaches have fueled the advancement in cancer research and eventually led to the uptake of telemedicine in cancer care. For diagnostic, prognostic, and treatment purposes concerning different types of cancer research, vast databases of varied information with manifold dimensions are required, and indeed, all this information can only be managed by an automated system developed utilizing ML and MM. In addition, MM is being used to probe the relationship between the pharmacokinetics and pharmacodynamics (PK/PD interactions) of anti-cancer substances to improve cancer treatment, and also to refine the quality of existing treatment models by being incorporated at all steps of research and development related to cancer and in routine patient care. This review will serve as a consolidation of the advancement and benefits of ML and MM techniques with a special focus on the area of cancer prognosis and anticancer therapy, leading to the identification of challenges (data quantity, ethical consideration, and data privacy) which are yet to be fully addressed in current studies.
Collapse
Affiliation(s)
- Jasmin Hassan
- Drug Delivery & Therapeutics Lab, Dhaka 1212, Bangladesh; (J.H.); (S.M.S.)
| | | | - Lipika Deka
- Faculty of Computing, Engineering and Media, De Montfort University, Leicester LE1 9BH, UK;
| | - Md Jasim Uddin
- Department of Pharmaceutical Technology, Faculty of Pharmacy, Universiti Malaya, Kuala Lumpur 50603, Malaysia
| | - Diganta B. Das
- Department of Chemical Engineering, Loughborough University, Loughborough LE11 3TU, UK
| |
Collapse
|
36
|
Chatterjee S, Saad F, Sarasaen C, Ghosh S, Krug V, Khatun R, Mishra R, Desai N, Radeva P, Rose G, Stober S, Speck O, Nürnberger A. Exploration of Interpretability Techniques for Deep COVID-19 Classification Using Chest X-ray Images. J Imaging 2024; 10:45. [PMID: 38392093 PMCID: PMC10889835 DOI: 10.3390/jimaging10020045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2024] [Revised: 01/24/2024] [Accepted: 02/05/2024] [Indexed: 02/24/2024] Open
Abstract
The outbreak of COVID-19 has shocked the entire world with its fairly rapid spread, and has challenged different sectors. One of the most effective ways to limit its spread is the early and accurate diagnosing of infected patients. Medical imaging, such as X-ray and computed tomography (CT), combined with the potential of artificial intelligence (AI), plays an essential role in supporting medical personnel in the diagnosis process. Thus, in this article, five different deep learning models (ResNet18, ResNet34, InceptionV3, InceptionResNetV2, and DenseNet161) and their ensemble, using majority voting, have been used to classify COVID-19, pneumoniæ and healthy subjects using chest X-ray images. Multilabel classification was performed to predict multiple pathologies for each patient, if present. Firstly, the interpretability of each of the networks was thoroughly studied using local interpretability methods-occlusion, saliency, input X gradient, guided backpropagation, integrated gradients, and DeepLIFT-and using a global technique-neuron activation profiles. The mean micro F1 score of the models for COVID-19 classifications ranged from 0.66 to 0.875, and was 0.89 for the ensemble of the network models. The qualitative results showed that the ResNets were the most interpretable models. This research demonstrates the importance of using interpretability methods to compare different models before making a decision regarding the best performing model.
Collapse
Affiliation(s)
- Soumick Chatterjee
- Data and Knowledge Engineering Group, Otto von Guericke University, 39106 Magdeburg, Germany
- Faculty of Computer Science, Otto von Guericke University, 39106 Magdeburg, Germany
- Genomics Research Centre, Human Technopole, 20157 Milan, Italy
| | - Fatima Saad
- Institute for Medical Engineering, Otto von Guericke University, 39106 Magdeburg, Germany
- Research Campus STIMULATE, Otto von Guericke University, 39106 Magdeburg, Germany
| | - Chompunuch Sarasaen
- Institute for Medical Engineering, Otto von Guericke University, 39106 Magdeburg, Germany
- Research Campus STIMULATE, Otto von Guericke University, 39106 Magdeburg, Germany
- Biomedical Magnetic Resonance, Otto von Guericke University, 39106 Magdeburg, Germany
| | - Suhita Ghosh
- Faculty of Computer Science, Otto von Guericke University, 39106 Magdeburg, Germany
- Artificial Intelligence Lab, Otto von Guericke University, 39106 Magdeburg, Germany
| | - Valerie Krug
- Faculty of Computer Science, Otto von Guericke University, 39106 Magdeburg, Germany
- Artificial Intelligence Lab, Otto von Guericke University, 39106 Magdeburg, Germany
| | - Rupali Khatun
- Department of Mathematics and Computer Science, University of Barcelona, 08028 Barcelona, Spain
- Translational Radiobiology, Department of Radiation Oncology, Universitätsklinikum Erlangen, 91054 Erlangen, Germany
| | | | | | - Petia Radeva
- Department of Mathematics and Computer Science, University of Barcelona, 08028 Barcelona, Spain
- Computer Vision Centre, 08193 Cerdanyola, Spain
| | - Georg Rose
- Institute for Medical Engineering, Otto von Guericke University, 39106 Magdeburg, Germany
- Research Campus STIMULATE, Otto von Guericke University, 39106 Magdeburg, Germany
- Centre for Behavioural Brain Sciences, 39106 Magdeburg, Germany
| | - Sebastian Stober
- Faculty of Computer Science, Otto von Guericke University, 39106 Magdeburg, Germany
- Artificial Intelligence Lab, Otto von Guericke University, 39106 Magdeburg, Germany
| | - Oliver Speck
- Research Campus STIMULATE, Otto von Guericke University, 39106 Magdeburg, Germany
- Biomedical Magnetic Resonance, Otto von Guericke University, 39106 Magdeburg, Germany
- Centre for Behavioural Brain Sciences, 39106 Magdeburg, Germany
- German Centre for Neurodegenerative Diseases, 39106 Magdeburg, Germany
| | - Andreas Nürnberger
- Data and Knowledge Engineering Group, Otto von Guericke University, 39106 Magdeburg, Germany
- Faculty of Computer Science, Otto von Guericke University, 39106 Magdeburg, Germany
- Centre for Behavioural Brain Sciences, 39106 Magdeburg, Germany
| |
Collapse
|
37
|
Kumar S, Kumar H, Kumar G, Singh SP, Bijalwan A, Diwakar M. A methodical exploration of imaging modalities from dataset to detection through machine learning paradigms in prominent lung disease diagnosis: a review. BMC Med Imaging 2024; 24:30. [PMID: 38302883 PMCID: PMC10832080 DOI: 10.1186/s12880-024-01192-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2023] [Accepted: 01/03/2024] [Indexed: 02/03/2024] Open
Abstract
BACKGROUND Lung diseases, both infectious and non-infectious, are the most prevalent cause of mortality overall in the world. Medical research has identified pneumonia, lung cancer, and Corona Virus Disease 2019 (COVID-19) as prominent lung diseases prioritized over others. Imaging modalities, including X-rays, computer tomography (CT) scans, magnetic resonance imaging (MRIs), positron emission tomography (PET) scans, and others, are primarily employed in medical assessments because they provide computed data that can be utilized as input datasets for computer-assisted diagnostic systems. Imaging datasets are used to develop and evaluate machine learning (ML) methods to analyze and predict prominent lung diseases. OBJECTIVE This review analyzes ML paradigms, imaging modalities' utilization, and recent developments for prominent lung diseases. Furthermore, the research also explores various datasets available publically that are being used for prominent lung diseases. METHODS The well-known databases of academic studies that have been subjected to peer review, namely ScienceDirect, arXiv, IEEE Xplore, MDPI, and many more, were used for the search of relevant articles. Applied keywords and combinations used to search procedures with primary considerations for review, such as pneumonia, lung cancer, COVID-19, various imaging modalities, ML, convolutional neural networks (CNNs), transfer learning, and ensemble learning. RESULTS This research finding indicates that X-ray datasets are preferred for detecting pneumonia, while CT scan datasets are predominantly favored for detecting lung cancer. Furthermore, in COVID-19 detection, X-ray datasets are prioritized over CT scan datasets. The analysis reveals that X-rays and CT scans have surpassed all other imaging techniques. It has been observed that using CNNs yields a high degree of accuracy and practicability in identifying prominent lung diseases. Transfer learning and ensemble learning are complementary techniques to CNNs to facilitate analysis. Furthermore, accuracy is the most favored metric for assessment.
Collapse
Affiliation(s)
- Sunil Kumar
- Department of Computer Engineering, J. C. Bose University of Science and Technology, YMCA, Faridabad, India
- Department of Information Technology, School of Engineering and Technology (UIET), CSJM University, Kanpur, India
| | - Harish Kumar
- Department of Computer Engineering, J. C. Bose University of Science and Technology, YMCA, Faridabad, India
| | - Gyanendra Kumar
- Department of Computer and Communication Engineering, Manipal University Jaipur, Jaipur, India
| | | | - Anchit Bijalwan
- Faculty of Electrical and Computer Engineering, Arba Minch University, Arba Minch, Ethiopia.
| | - Manoj Diwakar
- Department of Computer Science and Engineering, Graphic Era Deemed to Be University, Dehradun, India
| |
Collapse
|
38
|
Shiri I, Razeghi B, Ferdowsi S, Salimi Y, Gündüz D, Teodoro D, Voloshynovskiy S, Zaidi H. PRIMIS: Privacy-preserving medical image sharing via deep sparsifying transform learning with obfuscation. J Biomed Inform 2024; 150:104583. [PMID: 38191010 DOI: 10.1016/j.jbi.2024.104583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2023] [Revised: 11/19/2023] [Accepted: 01/02/2024] [Indexed: 01/10/2024]
Abstract
OBJECTIVE The primary objective of our study is to address the challenge of confidentially sharing medical images across different centers. This is often a critical necessity in both clinical and research environments, yet restrictions typically exist due to privacy concerns. Our aim is to design a privacy-preserving data-sharing mechanism that allows medical images to be stored as encoded and obfuscated representations in the public domain without revealing any useful or recoverable content from the images. In tandem, we aim to provide authorized users with compact private keys that could be used to reconstruct the corresponding images. METHOD Our approach involves utilizing a neural auto-encoder. The convolutional filter outputs are passed through sparsifying transformations to produce multiple compact codes. Each code is responsible for reconstructing different attributes of the image. The key privacy-preserving element in this process is obfuscation through the use of specific pseudo-random noise. When applied to the codes, it becomes computationally infeasible for an attacker to guess the correct representation for all the codes, thereby preserving the privacy of the images. RESULTS The proposed framework was implemented and evaluated using chest X-ray images for different medical image analysis tasks, including classification, segmentation, and texture analysis. Additionally, we thoroughly assessed the robustness of our method against various attacks using both supervised and unsupervised algorithms. CONCLUSION This study provides a novel, optimized, and privacy-assured data-sharing mechanism for medical images, enabling multi-party sharing in a secure manner. While we have demonstrated its effectiveness with chest X-ray images, the mechanism can be utilized in other medical images modalities as well.
Collapse
Affiliation(s)
- Isaac Shiri
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland; Department of Cardiology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Behrooz Razeghi
- Department of Computer Science, University of Geneva, Switzerland; Idiap Research Institute, Switzerland
| | - Sohrab Ferdowsi
- Department of Radiology and Medical Informatics, University of Geneva, Switzerland
| | - Yazdan Salimi
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland
| | - Deniz Gündüz
- Department of Electrical and Electronic Engineering, Imperial College London, UK
| | - Douglas Teodoro
- Department of Radiology and Medical Informatics, University of Geneva, Switzerland
| | | | - Habib Zaidi
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland; Department of Nuclear Medicine and Molecular Imaging, University of Groningen, The Netherlands; Department of Nuclear Medicine, University of Southern Denmark, Denmark; University Research and Innovation Center, Óbuda University, Budapest, Hungary.
| |
Collapse
|
39
|
Chen S, Ren S, Wang G, Huang M, Xue C. Interpretable CNN-Multilevel Attention Transformer for Rapid Recognition of Pneumonia From Chest X-Ray Images. IEEE J Biomed Health Inform 2024; 28:753-764. [PMID: 37027681 DOI: 10.1109/jbhi.2023.3247949] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/24/2023]
Abstract
Chest imaging plays an essential role in diagnosing and predicting patients with COVID-19 with evidence of worsening respiratory status. Many deep learning-based approaches for pneumonia recognition have been developed to enable computer-aided diagnosis. However, the long training and inference time makes them inflexible, and the lack of interpretability reduces their credibility in clinical medical practice. This paper aims to develop a pneumonia recognition framework with interpretability, which can understand the complex relationship between lung features and related diseases in chest X-ray (CXR) images to provide high-speed analytics support for medical practice. To reduce the computational complexity to accelerate the recognition process, a novel multi-level self-attention mechanism within Transformer has been proposed to accelerate convergence and emphasize the task-related feature regions. Moreover, a practical CXR image data augmentation has been adopted to address the scarcity of medical image data problems to boost the model's performance. The effectiveness of the proposed method has been demonstrated on the classic COVID-19 recognition task using the widespread pneumonia CXR image dataset. In addition, abundant ablation experiments validate the effectiveness and necessity of all of the components of the proposed method.
Collapse
|
40
|
Haque SBU, Zafar A. Robust Medical Diagnosis: A Novel Two-Phase Deep Learning Framework for Adversarial Proof Disease Detection in Radiology Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:308-338. [PMID: 38343214 PMCID: PMC11266337 DOI: 10.1007/s10278-023-00916-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 09/23/2023] [Accepted: 10/08/2023] [Indexed: 03/02/2024]
Abstract
In the realm of medical diagnostics, the utilization of deep learning techniques, notably in the context of radiology images, has emerged as a transformative force. The significance of artificial intelligence (AI), specifically machine learning (ML) and deep learning (DL), lies in their capacity to rapidly and accurately diagnose diseases from radiology images. This capability has been particularly vital during the COVID-19 pandemic, where rapid and precise diagnosis played a pivotal role in managing the spread of the virus. DL models, trained on vast datasets of radiology images, have showcased remarkable proficiency in distinguishing between normal and COVID-19-affected cases, offering a ray of hope amidst the crisis. However, as with any technological advancement, vulnerabilities emerge. Deep learning-based diagnostic models, although proficient, are not immune to adversarial attacks. These attacks, characterized by carefully crafted perturbations to input data, can potentially disrupt the models' decision-making processes. In the medical context, such vulnerabilities could have dire consequences, leading to misdiagnoses and compromised patient care. To address this, we propose a two-phase defense framework that combines advanced adversarial learning and adversarial image filtering techniques. We use a modified adversarial learning algorithm to enhance the model's resilience against adversarial examples during the training phase. During the inference phase, we apply JPEG compression to mitigate perturbations that cause misclassification. We evaluate our approach on three models based on ResNet-50, VGG-16, and Inception-V3. These models perform exceptionally in classifying radiology images (X-ray and CT) of lung regions into normal, pneumonia, and COVID-19 pneumonia categories. We then assess the vulnerability of these models to three targeted adversarial attacks: fast gradient sign method (FGSM), projected gradient descent (PGD), and basic iterative method (BIM). The results show a significant drop in model performance after the attacks. However, our defense framework greatly improves the models' resistance to adversarial attacks, maintaining high accuracy on adversarial examples. Importantly, our framework ensures the reliability of the models in diagnosing COVID-19 from clean images.
Collapse
Affiliation(s)
- Sheikh Burhan Ul Haque
- Department of Computer Science, Aligarh Muslim University, Uttar Pradesh, Aligarh, 202002, India.
| | - Aasim Zafar
- Department of Computer Science, Aligarh Muslim University, Uttar Pradesh, Aligarh, 202002, India
| |
Collapse
|
41
|
Pannipulath Venugopal V, Babu Saheer L, Maktabdar Oghaz M. COVID-19 lateral flow test image classification using deep CNN and StyleGAN2. Front Artif Intell 2024; 6:1235204. [PMID: 38348096 PMCID: PMC10860423 DOI: 10.3389/frai.2023.1235204] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Accepted: 12/28/2023] [Indexed: 02/15/2024] Open
Abstract
Introduction Artificial intelligence (AI) in healthcare can enhance clinical workflows and diagnoses, particularly in large-scale operations like COVID-19 mass testing. This study presents a deep Convolutional Neural Network (CNN) model for automated COVID-19 RATD image classification. Methods To address the absence of a RATD image dataset, we crowdsourced 900 real-world images focusing on positive and negative cases. Rigorous data augmentation and StyleGAN2-ADA generated simulated images to overcome dataset limitations and class imbalances. Results The best CNN model achieved a 93% validation accuracy. Test accuracies were 88% for simulated datasets and 82% for real datasets. Augmenting simulated images during training did not significantly improve real-world test image performance but enhanced simulated test image performance. Discussion The findings of this study highlight the potential of the developed model in expediting COVID-19 testing processes and facilitating large-scale testing and tracking systems. The study also underscores the challenges in designing and developing such models, emphasizing the importance of addressing dataset limitations and class imbalances. Conclusion This research contributes to the deployment of large-scale testing and tracking systems, offering insights into the potential applications of AI in mitigating outbreaks similar to COVID-19. Future work could focus on refining the model and exploring its adaptability to other healthcare scenarios.
Collapse
Affiliation(s)
| | - Lakshmi Babu Saheer
- School of Computing and Information Science, Anglia Ruskin University, Cambridge, United Kingdom
| | | |
Collapse
|
42
|
Fili V, Savelonas M. Self-attention-driven retrieval of chest CT images for COVID-19 assessment. Biomed Phys Eng Express 2024; 10:025013. [PMID: 38224614 DOI: 10.1088/2057-1976/ad1e76] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2023] [Accepted: 01/15/2024] [Indexed: 01/17/2024]
Abstract
Numerous methods have been developed for computer-aided diagnosis (CAD) of coronavirus disease-19 (COVID-19), based on chest computed tomography (CT) images. The majority of these methods are based on deep neural networks and often act as "black boxes" that cannot easily gain the trust of medical community, whereas their result is uniformly influenced by all image regions. This work introduces a novel, self-attention-driven method for content-based image retrieval (CBIR) of chest CT images. The proposed method analyzes a query CT image and returns a classification result, as well as a list of classified images, ranked according to similarity with the query. Each CT image is accompanied by a heatmap, which is derived by gradient-weighted class activation mapping (Grad-CAM) and represents the contribution of lung tissue and lesions to COVID-19 pathology. Beyond visualization, Grad-CAM weights are employed in a self-attention mechanism, in order to strengthen the influence of the most COVID-19-related image regions on the retrieval result. Experiments on two publicly available datasets demonstrate that the binary classification accuracy obtained by means of DenseNet-201 is 81.3% and 96.4%, for COVID-CT and SARS-CoV-2 datasets, respectively, with a false negative rate which is less than 3% in both datasets. In addition, the Grad-CAM-guided CBIR framework slightly outperforms the plain CBIR in most cases, with respect to nearest neighbour (NN) and first four (FF). The proposed method could serve as a computational tool for a more transparent decision-making process that could be trusted by the medical community. In addition, the employed self-attention mechanism increases the obtained retrieval performance.
Collapse
Affiliation(s)
- Victoria Fili
- Department of Computer Science and Biomedical Informatics, University of Thessaly, Papasiopoulou 2-4, Lamia, 35131, Greece
| | - Michalis Savelonas
- Department of Computer Science and Biomedical Informatics, University of Thessaly, Papasiopoulou 2-4, Lamia, 35131, Greece
| |
Collapse
|
43
|
Anand A, Krithivasan S, Roy K. RoMIA: a framework for creating Robust Medical Imaging AI models for chest radiographs. FRONTIERS IN RADIOLOGY 2024; 3:1274273. [PMID: 38260820 PMCID: PMC10800823 DOI: 10.3389/fradi.2023.1274273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Accepted: 12/18/2023] [Indexed: 01/24/2024]
Abstract
Artificial Intelligence (AI) methods, particularly Deep Neural Networks (DNNs), have shown great promise in a range of medical imaging tasks. However, the susceptibility of DNNs to producing erroneous outputs under the presence of input noise and variations is of great concern and one of the largest challenges to their adoption in medical settings. Towards addressing this challenge, we explore the robustness of DNNs trained for chest radiograph classification under a range of perturbations reflective of clinical settings. We propose RoMIA, a framework for the creation of Robust Medical Imaging AI models. RoMIA adds three key steps to the model training and deployment flow: (i) Noise-added training, wherein a part of the training data is synthetically transformed to represent common noise sources, (ii) Fine-tuning with input mixing, in which the model is refined with inputs formed by mixing data from the original training set with a small number of images from a different source, and (iii) DCT-based denoising, which removes a fraction of high-frequency components of each image before applying the model to classify it. We applied RoMIA to create six different robust models for classifying chest radiographs using the CheXpert dataset. We evaluated the models on the CheXphoto dataset, which consists of naturally and synthetically perturbed images intended to evaluate robustness. Models produced by RoMIA show 3%-5% improvement in robust accuracy, which corresponds to an average reduction of 22.6% in misclassifications. These results suggest that RoMIA can be a useful step towards enabling the adoption of AI models in medical imaging applications.
Collapse
Affiliation(s)
- Aditi Anand
- School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN, United States
| | | | | |
Collapse
|
44
|
Shao J, Wu F, Zhang J. Selective knowledge sharing for privacy-preserving federated distillation without a good teacher. Nat Commun 2024; 15:349. [PMID: 38191466 PMCID: PMC10774276 DOI: 10.1038/s41467-023-44383-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2023] [Accepted: 12/12/2023] [Indexed: 01/10/2024] Open
Abstract
While federated learning (FL) is promising for efficient collaborative learning without revealing local data, it remains vulnerable to white-box privacy attacks, suffers from high communication overhead, and struggles to adapt to heterogeneous models. Federated distillation (FD) emerges as an alternative paradigm to tackle these challenges, which transfers knowledge among clients instead of model parameters. Nevertheless, challenges arise due to variations in local data distributions and the absence of a well-trained teacher model, which leads to misleading and ambiguous knowledge sharing that significantly degrades model performance. To address these issues, this paper proposes a selective knowledge sharing mechanism for FD, termed Selective-FD, to identify accurate and precise knowledge from local and ensemble predictions, respectively. Empirical studies, backed by theoretical insights, demonstrate that our approach enhances the generalization capabilities of the FD framework and consistently outperforms baseline methods. We anticipate our study to enable a privacy-preserving, communication-efficient, and heterogeneity-adaptive federated training framework.
Collapse
Affiliation(s)
- Jiawei Shao
- Hong Kong University of Science and Technology, Hong Kong, China
| | | | - Jun Zhang
- Hong Kong University of Science and Technology, Hong Kong, China.
| |
Collapse
|
45
|
Saha S, Nandi D. SVM-RLF-DNN: A DNN with reliefF and SVM for automatic identification of COVID from chest X-ray and CT images. Digit Health 2024; 10:20552076241257045. [PMID: 38812845 PMCID: PMC11135098 DOI: 10.1177/20552076241257045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2024] [Accepted: 05/08/2024] [Indexed: 05/31/2024] Open
Abstract
Aim To develop an advanced determination technology for detecting COVID-19 patterns from chest X-ray and CT-scan films with distinct applications of deep learning and machine learning methods. Methods and Materials The newly enhanced proposed hybrid classification network (SVM-RLF-DNN) comprises of three phases: feature extraction, selection and classification. The in-depth features are extracted from a series of 3×3 convolution, 2×2 max polling operations followed by a flattened and fully connected layer of the deep neural network (DNN). ReLU activation function and Adam optimizer are used in the model. The ReliefF is an improved feature selection algorithm of Relief that uses Manhattan distance instead of Euclidean distance. Based on the significance of the feature, the ReliefF assigns weight to each extracted feature received from a fully connected layer. The weight to each feature is the average of k closest hits and misses in each class for a neighbouring instance pair in multiclass problems. The ReliefF eliminates lower-weight features by setting the node value to zero. The higher weights of the features are kept to obtain the feature selection. At the last layer of the neural network, the multiclass Support Vector Machine (SVM) is used to classify the patterns of COVID-19, viral pneumonia and healthy cases. The three classes with three binary SVM classifiers use linear kernel function for each binary SVM following a one-versus-all approach. The hinge loss function and L2-norm regularization are selected for more stable results. The proposed method is assessed on publicly available chest X-ray and CT-scan image databases from Kaggle and GitHub. The performance of the proposed classification model has comparable training, validation, and test accuracy, as well as sensitivity, specificity, and confusion matrix for quantitative evaluation on five-fold cross-validation. Results Our proposed network has achieved test accuracy of 98.48% and 95.34% on 2-class X-rays and CT. More importantly, the proposed model's test accuracy, sensitivity, and specificity are 87.9%, 86.32%, and 90.25% for 3-class classification (COVID-19, Pneumonia, Normal) on chest X-rays. The proposed model provides the test accuracy, sensitivity, and specificity of 95.34%, 94.12%, and 96.15% for 2-class classification (COVID-19, Non-COVID) on chest CT. Conclusion Our proposed classification network experimental results indicate competitiveness with existing neural networks. The proposed neural network assists clinicians in determining and surveilling the disease.
Collapse
Affiliation(s)
- Sanjib Saha
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, India
- Department of Computer Science and Engineering, Dr. B. C. Roy Engineering College, Durgapur, India
| | - Debashis Nandi
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, India
| |
Collapse
|
46
|
Zahedi Nasab R, Mohseni H, Montazeri M, Ghasemian F, Amin S. AFEX-Net: Adaptive feature extraction convolutional neural network for classifying computerized tomography images. Digit Health 2024; 10:20552076241232882. [PMID: 38406769 PMCID: PMC10894540 DOI: 10.1177/20552076241232882] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Accepted: 01/23/2024] [Indexed: 02/27/2024] Open
Abstract
Purpose Deep convolutional neural networks are favored methods that are widely used in medical image processing due to their demonstrated performance in this area. Recently, the emergence of new lung diseases, such as COVID-19, and the possibility of early detection of their symptoms from chest computerized tomography images has attracted many researchers to classify diseases by training deep convolutional neural networks on lung computerized tomography images. The trained networks are expected to distinguish between different lung indications in various diseases, especially at the early stages. The purpose of this study is to introduce and assess an efficient deep convolutional neural network, called AFEX-Net, that can classify different lung diseases from chest computerized tomography images. Methods We designed a lightweight convolutional neural network called AFEX-Net with adaptive feature extraction layers, adaptive pooling layers, and adaptive activation functions. We trained and tested AFEX-Net on a dataset of more than 10,000 chest computerized tomography slices from different lung diseases (CC dataset), using an effective pre-processing method to remove bias. We also applied AFEX-Net to the public COVID-CTset dataset to assess its generalizability. The study was mainly conducted based on data collected over approximately six months during the pandemic outbreak in Afzalipour Hospital, Iran, which is the largest hospital in Southeast Iran. Results AFEX-Net achieved high accuracy and fast training on both datasets, outperforming several state-of-the-art convolutional neural networks. It has an accuracy of 99.7 % and 98.8 % on the CC and COVID-CTset datasets, respectively, with a learning speed that is 3 times faster compared to similar methods due to its lightweight structure. AFEX-Net was able to extract distinguishing features and classify chest computerized tomography images, especially at the early stages of lung diseases. Conclusion The AFEX-Net is a high-performing convolutional neural network for classifying lung diseases from chest CT images. It is efficient, adaptable, and compatible with input data, making it a reliable tool for early detection and diagnosis of lung diseases.
Collapse
Affiliation(s)
- Roxana Zahedi Nasab
- Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran
| | - Hadis Mohseni
- Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran
| | - Mahdieh Montazeri
- Health Information Sciences Department, Kerman University of Medical Sciences, Kerman, Iran
| | - Fahimeh Ghasemian
- Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran
| | - Sobhan Amin
- Kazerun Branch, Islamic Azad University, Kazerun, Iran
| |
Collapse
|
47
|
Kondamuri SR, Thadikemalla VSG, Suryanarayana G, Karthik C, Reddy VS, Sahithi VB, Anitha Y, Yogitha V, Valli PR. Chest CT Image based Lung Disease Classification - A Review. Curr Med Imaging 2024; 20:1-14. [PMID: 38389342 DOI: 10.2174/0115734056248176230923143105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Revised: 07/22/2023] [Accepted: 08/22/2023] [Indexed: 02/24/2024]
Abstract
Computed tomography (CT) scans are widely used to diagnose lung conditions due to their ability to provide a detailed overview of the body's respiratory system. Despite its popularity, visual examination of CT scan images can lead to misinterpretations that impede a timely diagnosis. Utilizing technology to evaluate images for disease detection is also a challenge. As a result, there is a significant demand for more advanced systems that can accurately classify lung diseases from CT scan images. In this work, we provide an extensive analysis of different approaches and their performances that can help young researchers to build more advanced systems. First, we briefly introduce diagnosis and treatment procedures for various lung diseases. Then, a brief description of existing methods used for the classification of lung diseases is presented. Later, an overview of the general procedures for lung disease classification using machine learning (ML) is provided. Furthermore, an overview of recent progress in ML-based classification of lung diseases is provided. Finally, existing challenges in ML techniques are presented. It is concluded that deep learning techniques have revolutionized the early identification of lung disorders. We expect that this work will equip medical professionals with the awareness they require in order to recognize and classify certain medical disorders.
Collapse
Affiliation(s)
- Shri Ramtej Kondamuri
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| | | | - Gunnam Suryanarayana
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| | - Chandran Karthik
- Department of Robotics and Automation, Jyothi Engineering College, Thrissur, Kerala 679531, India
| | - Vanga Siva Reddy
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| | - V Bhuvana Sahithi
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| | - Y Anitha
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| | - V Yogitha
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| | - P Reshma Valli
- Department of ECE, Velagapudi Ramakrishna Siddhartha Engineering College, Vijayawada, Andhra Pradesh, 520007, India
| |
Collapse
|
48
|
Antar S, Abd El-Sattar HKH, Abdel-Rahman MH, F M Ghaleb F. COVID-19 infection segmentation using hybrid deep learning and image processing techniques. Sci Rep 2023; 13:22737. [PMID: 38123587 PMCID: PMC10733411 DOI: 10.1038/s41598-023-49337-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2023] [Accepted: 12/07/2023] [Indexed: 12/23/2023] Open
Abstract
The coronavirus disease 2019 (COVID-19) epidemic has become a worldwide problem that continues to affect people's lives daily, and the early diagnosis of COVID-19 has a critical importance on the treatment of infected patients for medical and healthcare organizations. To detect COVID-19 infections, medical imaging techniques, including computed tomography (CT) scan images and X-ray images, are considered some of the helpful medical tests that healthcare providers carry out. However, in addition to the difficulty of segmenting contaminated areas from CT scan images, these approaches also offer limited accuracy for identifying the virus. Accordingly, this paper addresses the effectiveness of using deep learning (DL) and image processing techniques, which serve to expand the dataset without the need for any augmentation strategies, and it also presents a novel approach for detecting COVID-19 virus infections in lung images, particularly the infection prediction issue. In our proposed method, to reveal the infection, the input images are first preprocessed using a threshold then resized to 128 × 128. After that, a density heat map tool is used for coloring the resized lung images. The three channels (red, green, and blue) are then separated from the colored image and are further preprocessed through image inverse and histogram equalization, and are subsequently fed, in independent directions, into three separate U-Nets with the same architecture for segmentation. Finally, the segmentation results are combined and run through a convolution layer one by one to get the detection. Several evaluation metrics using the CT scan dataset were used to measure the performance of the proposed approach in comparison with other state-of-the-art techniques in terms of accuracy, sensitivity, precision, and the dice coefficient. The experimental results of the proposed approach reached 99.71%, 0.83, 0.87, and 0.85, respectively. These results show that coloring the CT scan images dataset and then dividing each image into its RGB image channels can enhance the COVID-19 detection, and it also increases the U-Net power in the segmentation when merging the channel segmentation results. In comparison to other existing segmentation techniques employing bigger 512 × 512 images, this study is one of the few that can rapidly and correctly detect the COVID-19 virus with high accuracy on smaller 128 × 128 images using the metrics of accuracy, sensitivity, precision, and dice coefficient.
Collapse
Affiliation(s)
- Samar Antar
- Computer Science Division, Department of Mathematics, Faculty of Science, Ain Shams University, Abbassia, Cairo, 11566, Egypt
| | | | - Mohammad H Abdel-Rahman
- Computer Science Division, Department of Mathematics, Faculty of Science, Ain Shams University, Abbassia, Cairo, 11566, Egypt
| | - Fayed F M Ghaleb
- Computer Science Division, Department of Mathematics, Faculty of Science, Ain Shams University, Abbassia, Cairo, 11566, Egypt
| |
Collapse
|
49
|
Phumkuea T, Wongsirichot T, Damkliang K, Navasakulpong A, Andritsch J. MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models. Tomography 2023; 9:2233-2246. [PMID: 38133077 PMCID: PMC10747997 DOI: 10.3390/tomography9060173] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Revised: 12/08/2023] [Accepted: 12/08/2023] [Indexed: 12/23/2023] Open
Abstract
This study introduces a Multi-Stage Automated Classification (MSTAC) system for COVID-19 chest X-ray (CXR) images, utilizing stacked Convolutional Neural Network (CNN) models. Suspected COVID-19 patients often undergo CXR imaging, making it valuable for disease classification. The study collected CXR images from public datasets and aimed to differentiate between COVID-19, non-COVID-19, and healthy cases. MSTAC employs two classification stages: the first distinguishes healthy from unhealthy cases, and the second further classifies COVID-19 and non-COVID-19 cases. Compared to a single CNN-Multiclass model, MSTAC demonstrated superior classification performance, achieving 97.30% accuracy and sensitivity. In contrast, the CNN-Multiclass model showed 94.76% accuracy and sensitivity. MSTAC's effectiveness is highlighted in its promising results over the CNN-Multiclass model, suggesting its potential to assist healthcare professionals in efficiently diagnosing COVID-19 cases. The system outperformed similar techniques, emphasizing its accuracy and efficiency in COVID-19 diagnosis. This research underscores MSTAC as a valuable tool in medical image analysis for enhanced disease classification.
Collapse
Affiliation(s)
- Thanakorn Phumkuea
- College of Digital Science, Prince of Songkla University, Songkhla 90110, Thailand
| | - Thakerng Wongsirichot
- Division of Computational Science, Faculty of Science, Prince of Songkla University, Songkhla 90110, Thailand;
| | - Kasikrit Damkliang
- Division of Computational Science, Faculty of Science, Prince of Songkla University, Songkhla 90110, Thailand;
| | - Asma Navasakulpong
- Division of Respiratory and Respiratory Critical Care Medicine, Prince of Songkla University, Songkhla 90110, Thailand;
| | - Jarutas Andritsch
- Faculty of Business, Law and Digital Technologies, Solent University, Southampton SO14 0YN, UK;
| |
Collapse
|
50
|
Kumar S, Kumar H. Classification of COVID-19 X-ray images using transfer learning with visual geometrical groups and novel sequential convolutional neural networks. MethodsX 2023; 11:102295. [PMID: 37539339 PMCID: PMC10393783 DOI: 10.1016/j.mex.2023.102295] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2022] [Accepted: 07/20/2023] [Indexed: 08/05/2023] Open
Abstract
COVID-19 is a highly transmissible infectious disease that remains a substantial challenge. The utilization of chest radiology, particularly X-ray imaging, has proven to be highly effective, easily accessible, and cost-efficient in detecting COVID-19. A dataset named COVID-Xray-5k, consisting of imbalanced X-ray images of COVID-19-positive and normal subjects, is employed for investigation. The research introduces a novel methodology that utilizes conventional machine learning (ML), such as local binary patterns (LBP) for feature extraction and support vector machines (SVM) for classification. In addition, transfer learning is employed with the Visual Geometry Group 16-layer (VGG16) and 19-layer (VGG19) models. Besides, novel sequential convolutional neural network (CNN) architectures are presented to develop an autonomous system for classifying COVID-19. One of the proposed CNN architectures classifies the test dataset with an F1 score of 91.00% and an accuracy of 99.45% based on an empirical investigation to determine optimal hyper-parameters. The methods presented in the research show promising potential for COVID-19 classification, irrespective of class imbalance.•Employment of ML models to investigate subjective feature engineering and classification.•Transfer learning was employed for VGG16 and VGG19 with eight distinct models.•Illustration of two novel CNN sequential architectures; all the investigation is performed with and without weighted sampling.
Collapse
Affiliation(s)
- Sunil Kumar
- Department of Information Technology, School of Engineering and Technology (UIET), CSJM University, Kanpur, UP, India
- Department of Computer Engineering, J.C. Bose University of Science and Technology, YMCA, Faridabad, India
| | - Harish Kumar
- Department of Computer Engineering, J.C. Bose University of Science and Technology, YMCA, Faridabad, India
| |
Collapse
|