1
|
Ahmed A, Zeng X, Xi R, Hou M, Shah SA. Enhancing multimodal medical image analysis with Slice-Fusion: A novel fusion approach to address modality imbalance. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 261:108615. [PMID: 39904191 DOI: 10.1016/j.cmpb.2025.108615] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/26/2024] [Revised: 01/01/2025] [Accepted: 01/20/2025] [Indexed: 02/06/2025]
Abstract
BACKGROUND AND OBJECTIVE In recent times, medical imaging analysis (MIA) has seen an increasing interest due to its core application in computer-aided diagnosis systems (CADs). A modality in MIA refers to a specific technology used to produce human body images, such as MRI, CT scans, or X-rays. Each modality presents unique challenges and characteristics, often leading to imbalances within datasets. This significant challenge impedes model training and generalization due to the varying convergence rates of different modalities and the suppression of gradients in less dominant modalities. METHODS This paper proposes a novel fusion approach, and we named it Slice-Fusion. The proposed approach aims to mitigate the modality imbalance problem by implementing a "Modality-Specific-Balancing-Factor" fusion strategy. Furthermore, it incorporates an auxiliary (uni-modal) task that generates balanced modality pairs based on the image orientations of different modalities. Subsequently, a novel multimodal classification framework is presented to learn from the generated balanced modalities. The effectiveness of the proposed approach is evaluated through comparative assessments on a publicly available BraTS2021 dataset. The results demonstrate the efficiency of Slice-Fusion in resolving the modality imbalance problem. By enhancing the representation of balanced features and reducing modality bias, this approach holds promise for advancing visual health informatics and facilitating more accurate and reliable medical image analysis. RESULTS In the experiment section, three diverse experiments are conducted such as i) Fusion Loss Metrics Evaluation, ii) Classification, and iii) Visual Health Informatics. Notably, the proposed approach achieved an F1-Score of (100%, 81.25%) on the training and validation sets for the classification generalization task. In addition to the Slice-Fusion's out-performance, the study also created a new modality-aligned dataset (a highly balanced and informative modality-specific image collection) that aids further research and improves MIA's robustness. These advancements not only enhance the capability of medical diagnostic tools but also create opportunities for future innovations in the field. CONCLUSION This study contributes to advancing medical image analysis, such as effective modality fusion, image reconstruction, comparison, and glioma classification, facilitating more accurate and reliable results, and holds promise for further advancements in visual health informatics.
Collapse
Affiliation(s)
- Awais Ahmed
- School of Computer Science, China West Normal University, China; School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China.
| | - Xiaoyang Zeng
- School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China.
| | - Rui Xi
- School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China.
| | - Mengshu Hou
- School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China; School of Big Data and Artificial Intelligence, Chengdu Technological University, Sichuan, 611730, China.
| | - Syed Attique Shah
- School of Computing and Digital Technology, Birmingham City University, STEAMhouse, B4 7RQ, Birmingham, United Kingdom.
| |
Collapse
|
2
|
Baxter JSH, Eagleson R. Exploring the values underlying machine learning research in medical image analysis. Med Image Anal 2025; 102:103494. [PMID: 40020419 DOI: 10.1016/j.media.2025.103494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 01/26/2025] [Accepted: 02/01/2025] [Indexed: 03/03/2025]
Abstract
Machine learning has emerged as a crucial tool for medical image analysis, largely due to recent developments in deep artificial neural networks addressing numerous, diverse clinical problems. As with any conceptual tool, the effective use of machine learning should be predicated on an understanding of its underlying motivations just as much as algorithms or theory - and to do so, we need to explore its philosophical foundations. One of these foundations is the understanding of how values, despite being non-empirical, nevertheless affect scientific research. This article has three goals: to introduce the reader to values in a way that is specific to medical image analysis; to characterise a particular set of technical decisions (what we call the end-to-end vs. separable learning spectrum) that are fundamental to machine learning for medical image analysis; and to create a simple and structured method to show how these values can be rigorously connected to these technical decisions. This better understanding of how the philosophy of science can clarify fundamental elements of how medical image analysis research is performed and can be improved.
Collapse
Affiliation(s)
- John S H Baxter
- Laboratoire Traitement du Signal et de l'Image (LTSI, INSERM UMR 1099), Université de Rennes, Rennes, France.
| | - Roy Eagleson
- Biomedical Engineering Graduate Program, Western University, London, Canada
| |
Collapse
|
3
|
Deng Z, Yang Y, Suzuki K. Federated Active Learning Framework for Efficient Annotation Strategy in Skin-Lesion Classification. J Invest Dermatol 2025; 145:303-311. [PMID: 38909844 DOI: 10.1016/j.jid.2024.05.023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 05/10/2024] [Accepted: 05/27/2024] [Indexed: 06/25/2024]
Abstract
Federated learning (FL) enables multiple institutes to train models collaboratively without sharing private data. Current FL research focuses on communication efficiency, privacy protection, and personalization and assumes that the data of FL have already been ideally collected. However, in medical scenarios, data annotation demands both expertise and intensive labor, which is a critical problem in FL. Active learning (AL) has shown promising performance in reducing the number of data annotations in medical image analysis. We propose a federated AL framework in which AL is executed periodically and interactively under FL. We exploit a local model in each hospital and a global model acquired from FL to construct an ensemble. We use ensemble entropy-based AL as an efficient data-annotation strategy in FL. Therefore, our federated AL framework can decrease the amount of annotated data and preserve patient privacy while maintaining the performance of FL. To our knowledge, this federated AL framework applied to medical images has not been previously reported. We validated our framework on real-world dermoscopic datasets. Using only 50% of samples, our framework was able to achieve state-of-the-art performance on a skin-lesion classification task. Our framework performed better than several state-of-the-art AL methods under FL and achieved comparable performance with full-data FL.
Collapse
Affiliation(s)
- Zhipeng Deng
- Biomedical Artificial Intelligence Research Unit (BMAI), Institute of Innovative Research, Tokyo Institute of Technology, Tokyo, Japan; Department of Information and Communications Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo, Japan
| | - Yuqiao Yang
- Biomedical Artificial Intelligence Research Unit (BMAI), Institute of Innovative Research, Tokyo Institute of Technology, Tokyo, Japan; Department of Information and Communications Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo, Japan
| | - Kenji Suzuki
- Biomedical Artificial Intelligence Research Unit (BMAI), Institute of Innovative Research, Tokyo Institute of Technology, Tokyo, Japan; Department of Information and Communications Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo, Japan.
| |
Collapse
|
4
|
Ramos-Briceño DA, Flammia-D'Aleo A, Fernández-López G, Carrión-Nessi FS, Forero-Peña DA. Deep learning-based malaria parasite detection: convolutional neural networks model for accurate species identification of Plasmodium falciparum and Plasmodium vivax. Sci Rep 2025; 15:3746. [PMID: 39885248 PMCID: PMC11782605 DOI: 10.1038/s41598-025-87979-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Accepted: 01/23/2025] [Indexed: 02/01/2025] Open
Abstract
Accurate malaria diagnosis with precise identification of Plasmodium species is crucial for an effective treatment. While microscopy is still the gold standard in malaria diagnosis, it relies heavily on trained personnel. Artificial intelligence (AI) advances, particularly convolutional neural networks (CNNs), have significantly improved diagnostic capabilities and accuracy by enabling the automated analysis of medical images. Previous models efficiently detected malaria parasites in red blood cells but had difficulty differentiating between species. We propose a CNN-based model for classifying cells infected by P. falciparum, P. vivax, and uninfected white blood cells from thick blood smears. Our best-performing model utilizes a seven-channel input and correctly predicted 12,876 out of 12,954 cases. We also generated a cross-validation confusion matrix that showed the results of five iterations, achieving 63,654 out of 64,126 true predictions. The model's accuracy reached 99.51%, a precision of 99.26%, a recall of 99.26%, a specificity of 99.63%, an F1 score of 99.26%, and a loss of 2.3%. We are now developing a system based on real-world quality images to create a comprehensive detection tool for remote regions where trained microscopists are unavailable.
Collapse
Affiliation(s)
- Diego A Ramos-Briceño
- School of Systems Engineering, Faculty of Engineering, Universidad Metropolitana de Caracas, Caracas, Venezuela
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela
- "Luis Razetti" School of Medicine, Universidad Central de Venezuela, Caracas, Venezuela
| | - Alessandro Flammia-D'Aleo
- School of Systems Engineering, Faculty of Engineering, Universidad Metropolitana de Caracas, Caracas, Venezuela
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela
| | - Gerardo Fernández-López
- Department of Electronics and Circuits, Faculty of Engineering, Universidad Simón Bolívar, Caracas, Venezuela
| | - Fhabián S Carrión-Nessi
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela.
- "Luis Razetti" School of Medicine, Universidad Central de Venezuela, Caracas, Venezuela.
- Immunogenetics Section, Laboratory of Pathophysiology, Centro de Medicina Experimental "Miguel Layrisse", Instituto Venezolano de Investigaciones Científicas, Altos de Pipe, Venezuela.
| | - David A Forero-Peña
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela.
- "Luis Razetti" School of Medicine, Universidad Central de Venezuela, Caracas, Venezuela.
- Department of Infectious Diseases, Hospital Universitario de Caracas, Caracas, Venezuela.
| |
Collapse
|
5
|
Kim JW, Khan AU, Banerjee I. Systematic Review of Hybrid Vision Transformer Architectures for Radiological Image Analysis. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-024-01322-4. [PMID: 39871042 DOI: 10.1007/s10278-024-01322-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/28/2024] [Revised: 10/11/2024] [Accepted: 10/25/2024] [Indexed: 01/29/2025]
Abstract
Vision transformer (ViT)and convolutional neural networks (CNNs) each possess distinct strengths in medical imaging: ViT excels in capturing long-range dependencies through self-attention, while CNNs are adept at extracting local features via spatial convolution filters. While ViT may struggle with capturing detailed local spatial information, critical for tasks like anomaly detection in medical imaging, shallow CNNs often fail to effectively abstract global context. This study aims to explore and evaluate hybrid architectures that integrate ViT and CNN to leverage their complementary strengths for enhanced performance in medical vision tasks, such as segmentation, classification, reconstruction, and prediction. Following PRISMA guideline, a systematic review was conducted on 34 articles published between 2020 and Sept. 2024. These articles proposed novel hybrid ViT-CNN architectures specifically for medical imaging tasks in radiology. The review focused on analyzing architectural variations, merging strategies between ViT and CNN, innovative applications of ViT, and efficiency metrics including parameters, inference time (GFlops), and performance benchmarks. The review identified that integrating ViT and CNN can mitigate the limitations of each architecture offering comprehensive solutions that combine global context understanding with precise local feature extraction. We benchmarked the articles based on architectural variations, merging strategies, innovative uses of ViT, and efficiency metrics (number of parameters, inference time (GFlops), and performance), and derived a ranked list. By synthesizing current literature, this review defines fundamental concepts of hybrid vision transformers and highlights emerging trends in the field. It provides a clear direction for future research aimed at optimizing the integration of ViT and CNN for effective utilization in medical imaging, contributing to advancements in diagnostic accuracy and image analysis. We performed systematic review of hybrid vision transformer architecture using PRISMA guideline and performed thorough comparative analysis to benchmark the architectures.
Collapse
Affiliation(s)
- Ji Woong Kim
- School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ, USA
| | | | - Imon Banerjee
- School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ, USA.
- Department of Radiology, Mayo Clinic, Phoenix, AZ, USA.
- Department of Artificial Intelligence and Informatics (AI&I), Mayo Clinic, Scottsdale, AZ, USA.
| |
Collapse
|
6
|
Dashti M, Londono J, Ghasemi S, Tabatabaei S, Hashemi S, Baghaei K, Palma PJ, Khurshid Z. Evaluation of accuracy of deep learning and conventional neural network algorithms in detection of dental implant type using intraoral radiographic images: A systematic review and meta-analysis. J Prosthet Dent 2025; 133:137-146. [PMID: 38176985 DOI: 10.1016/j.prosdent.2023.11.030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2023] [Revised: 11/23/2023] [Accepted: 11/28/2023] [Indexed: 01/06/2024]
Abstract
STATEMENT OF PROBLEM With the growing importance of implant brand detection in clinical practice, the accuracy of machine learning algorithms in implant brand detection has become a subject of research interest. Recent studies have shown promising results for the use of machine learning in implant brand detection. However, despite these promising findings, a comprehensive evaluation of the accuracy of machine learning in implant brand detection is needed. PURPOSE The purpose of this systematic review and meta-analysis was to assess the accuracy, sensitivity, and specificity of deep learning algorithms in implant brand detection using 2-dimensional images such as from periapical or panoramic radiographs. MATERIAL AND METHODS Electronic searches were conducted in PubMed, Embase, Scopus, Scopus Secondary, and Web of Science databases. Studies that met the inclusion criteria were assessed for quality using the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool. Meta-analyses were performed using the random-effects model to estimate the pooled performance measures and the 95% confidence intervals (CIs) using STATA v.17. RESULTS Thirteen studies were selected for the systematic review, and 3 were used in the meta-analysis. The meta-analysis of the studies found that the overall accuracy of CNN algorithms in detecting dental implants in radiographic images was 95.63%, with a sensitivity of 94.55% and a specificity of 97.91%. The highest reported accuracy was 99.08% for CNN Multitask ResNet152 algorithm, and sensitivity and specificity were 100.00% and 98.70% respectively for the deep CNN (Neuro-T version 2.0.1) algorithm with the Straumann SLActive BLT implant brand. All studies had a low risk of bias. CONCLUSIONS The highest accuracy and sensitivity were reported in studies using CNN Multitask ResNet152 and deep CNN (Neuro-T version 2.0.1) algorithms.
Collapse
Affiliation(s)
- Mahmood Dashti
- Researcher, School of Dentistry, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| | - Jimmy Londono
- Professor and Director of the Prosthodontics Residency Program and the Ronald Goldstein Center for Esthetics and Implant Dentistry, The Dental College of Georgia at Augusta University, Augusta, GA
| | - Shohreh Ghasemi
- Graduate Student, MSc of Trauma and Craniofacial Reconstrution, Faculty of Medicine and Dentistry, Queen Mary College, London, England
| | | | - Sara Hashemi
- Graduate student, Faculty of Dental Medicine and Oral Health Sciences, McGill University, Montreal, Quebec, Canada
| | - Kimia Baghaei
- Researcher, Dental Students' Research Committee, School of Dentistry, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Paulo J Palma
- Researcher, Center for Innovation and Research in Oral Sciences (CIROS), Faculty of Medicine, University of Coimbra, Coimbra, Portugal; and Professor, Institute of Endodontics, Faculty of Medicine, University of Coimbra, Coimbra, Portugal.
| | - Zohaib Khurshid
- Lecturer, Prosthodontics, Department of Prosthodontics and Dental Implantology, King Faisal University, Al-Ahsa, Saudi Arabia; and Professor, Center of Excellence for Regenerative Dentistry, Department of Anatomy, Faculty of Dentistry, Chulalongkorn University, Bangkok, Thailand
| |
Collapse
|
7
|
Obeidat R, Alsmadi I, Baker QB, Al-Njadat A, Srinivasan S. Researching public health datasets in the era of deep learning: a systematic literature review. Health Informatics J 2025; 31:14604582241307839. [PMID: 39794941 DOI: 10.1177/14604582241307839] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/13/2025]
Abstract
Objective: Explore deep learning applications in predictive analytics for public health data, identify challenges and trends, and then understand the current landscape. Materials and Methods: A systematic literature review was conducted in June 2023 to search articles on public health data in the context of deep learning, published from the inception of medical and computer science databases through June 2023. The review focused on diverse datasets, abstracting applications, challenges, and advancements in deep learning. Results: 2004 articles were reviewed, identifying 14 disease categories. Observed trends include explainable-AI, patient embedding learning, and integrating different data sources and employing deep learning models in health informatics. Noted challenges were technical reproducibility and handling sensitive data. Discussion: There has been a notable surge in deep learning applications on public health data publications since 2015. Consistent deep learning applications and models continue to be applied across public health data. Despite the wide applications, a standard approach still does not exist for addressing the outstanding challenges and issues in this field. Conclusion: Guidelines are needed for applying deep learning and models in public health data to improve FAIRness, efficiency, transparency, comparability, and interoperability of research. Interdisciplinary collaboration among data scientists, public health experts, and policymakers is needed to harness the full potential of deep learning.
Collapse
Affiliation(s)
- Rand Obeidat
- Department of Management Information Systems, Bowie State University, Bowie, USA
| | - Izzat Alsmadi
- Department of Computational, Engineering and Mathematical Sciences, Texas A & M San Antonio, San Antonio, USA
| | - Qanita Bani Baker
- Department of Computer Science, Jordan University of Science and Technology, Irbid, Jordan
| | | | - Sriram Srinivasan
- Department of Management Information Systems, Bowie State University, Bowie, USA
| |
Collapse
|
8
|
Viqar M, Sahin E, Stoykova E, Madjarova V. Reconstruction of Optical Coherence Tomography Images from Wavelength Space Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2024; 25:93. [PMID: 39796883 PMCID: PMC11723098 DOI: 10.3390/s25010093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Revised: 12/19/2024] [Accepted: 12/24/2024] [Indexed: 01/13/2025]
Abstract
Conventional Fourier domain Optical Coherence Tomography (FD-OCT) systems depend on resampling into a wavenumber (k) domain to extract the depth profile. This either necessitates additional hardware resources or amplifies the existing computational complexity. Moreover, the OCT images also suffer from speckle noise, due to systemic reliance on low-coherence interferometry. We propose a streamlined and computationally efficient approach based on Deep Learning (DL) which enables reconstructing speckle-reduced OCT images directly from the wavelength (λ) domain. For reconstruction, two encoder-decoder styled networks, namely Spatial Domain Convolution Neural Network (SD-CNN) and Fourier Domain CNN (FD-CNN), are used sequentially. The SD-CNN exploits the highly degraded images obtained by Fourier transforming the (λ) domain fringes to reconstruct the deteriorated morphological structures along with suppression of unwanted noise. The FD-CNN leverages this output to enhance the image quality further by optimization in the Fourier domain (FD). We quantitatively and visually demonstrate the efficacy of the method in obtaining high-quality OCT images. Furthermore, we illustrate the computational complexity reduction by harnessing the power of DL models. We believe that this work lays the framework for further innovations in the realm of OCT image reconstruction.
Collapse
Affiliation(s)
- Maryam Viqar
- Faculty of Information Technology and Communication Sciences, Tampere University, 33720 Tampere, Finland;
- Institute of Optical Materials and Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria; (E.S.); (V.M.)
| | - Erdem Sahin
- Faculty of Information Technology and Communication Sciences, Tampere University, 33720 Tampere, Finland;
| | - Elena Stoykova
- Institute of Optical Materials and Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria; (E.S.); (V.M.)
| | - Violeta Madjarova
- Institute of Optical Materials and Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria; (E.S.); (V.M.)
| |
Collapse
|
9
|
Berghout T. The Neural Frontier of Future Medical Imaging: A Review of Deep Learning for Brain Tumor Detection. J Imaging 2024; 11:2. [PMID: 39852315 PMCID: PMC11766058 DOI: 10.3390/jimaging11010002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2024] [Revised: 12/21/2024] [Accepted: 12/23/2024] [Indexed: 01/26/2025] Open
Abstract
Brain tumor detection is crucial in medical research due to high mortality rates and treatment challenges. Early and accurate diagnosis is vital for improving patient outcomes, however, traditional methods, such as manual Magnetic Resonance Imaging (MRI) analysis, are often time-consuming and error-prone. The rise of deep learning has led to advanced models for automated brain tumor feature extraction, segmentation, and classification. Despite these advancements, comprehensive reviews synthesizing recent findings remain scarce. By analyzing over 100 research papers over past half-decade (2019-2024), this review fills that gap, exploring the latest methods and paradigms, summarizing key concepts, challenges, datasets, and offering insights into future directions for brain tumor detection using deep learning. This review also incorporates an analysis of previous reviews and targets three main aspects: feature extraction, segmentation, and classification. The results revealed that research primarily focuses on Convolutional Neural Networks (CNNs) and their variants, with a strong emphasis on transfer learning using pre-trained models. Other methods, such as Generative Adversarial Networks (GANs) and Autoencoders, are used for feature extraction, while Recurrent Neural Networks (RNNs) are employed for time-sequence modeling. Some models integrate with Internet of Things (IoT) frameworks or federated learning for real-time diagnostics and privacy, often paired with optimization algorithms. However, the adoption of eXplainable AI (XAI) remains limited, despite its importance in building trust in medical diagnostics. Finally, this review outlines future opportunities, focusing on image quality, underexplored deep learning techniques, expanding datasets, and exploring deeper learning representations and model behavior such as recurrent expansion to advance medical imaging diagnostics.
Collapse
Affiliation(s)
- Tarek Berghout
- Laboratory of Automation and Manufacturing Engineering, Department of Industrial Engineering, Batna 2 University, Batna 05000, Algeria
| |
Collapse
|
10
|
Lerch L, Huber LS, Kamath A, Pöllinger A, Pahud de Mortanges A, Obmann VC, Dammann F, Senn W, Reyes M. DreamOn: a data augmentation strategy to narrow the robustness gap between expert radiologists and deep learning classifiers. FRONTIERS IN RADIOLOGY 2024; 4:1420545. [PMID: 39758512 PMCID: PMC11696537 DOI: 10.3389/fradi.2024.1420545] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/20/2024] [Accepted: 11/22/2024] [Indexed: 01/07/2025]
Abstract
Purpose Successful performance of deep learning models for medical image analysis is highly dependent on the quality of the images being analysed. Factors like differences in imaging equipment and calibration, as well as patient-specific factors such as movements or biological variability (e.g., tissue density), lead to a large variability in the quality of obtained medical images. Consequently, robustness against the presence of noise is a crucial factor for the application of deep learning models in clinical contexts. Materials and methods We evaluate the effect of various data augmentation strategies on the robustness of a ResNet-18 trained to classify breast ultrasound images and benchmark the performance against trained human radiologists. Additionally, we introduce DreamOn, a novel, biologically inspired data augmentation strategy for medical image analysis. DreamOn is based on a conditional generative adversarial network (GAN) to generate REM-dream-inspired interpolations of training images. Results We find that while available data augmentation approaches substantially improve robustness compared to models trained without any data augmentation, radiologists outperform models on noisy images. Using DreamOn data augmentation, we obtain a substantial improvement in robustness in the high noise regime. Conclusions We show that REM-dream-inspired conditional GAN-based data augmentation is a promising approach to improving deep learning model robustness against noise perturbations in medical imaging. Additionally, we highlight a gap in robustness between deep learning models and human experts, emphasizing the imperative for ongoing developments in AI to match human diagnostic expertise.
Collapse
Affiliation(s)
- Luc Lerch
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
- Computational Neuroscience Group, Department of Physiology, University of Bern, Bern, Switzerland
| | - Lukas S. Huber
- Cognition, Perception and Research Methods, Department of Psychology, University of Bern, Bern, Switzerland
- Neural Information Processing Group, Department of Computer Science, University of Tübingen, Tübingen, Germany
| | - Amith Kamath
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
| | - Alexander Pöllinger
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Aurélie Pahud de Mortanges
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
| | - Verena C. Obmann
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Florian Dammann
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Walter Senn
- Computational Neuroscience Group, Department of Physiology, University of Bern, Bern, Switzerland
- Center for Artificial Intelligence in Medicine, University of Bern, Bern, Switzerland
| | - Mauricio Reyes
- Center for Artificial Intelligence in Medicine, University of Bern, Bern, Switzerland
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
- Department of Radiation Oncology, University Hospital Bern, University of Bern, Bern, Switzerland
| |
Collapse
|
11
|
Lim H, Gi Y, Ko Y, Jo Y, Hong J, Kim J, Ahn SH, Park HC, Kim H, Chung K, Yoon M. A device-dependent auto-segmentation method based on combined generalized and single-device datasets. Med Phys 2024. [PMID: 39699056 DOI: 10.1002/mp.17570] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2024] [Revised: 11/12/2024] [Accepted: 11/26/2024] [Indexed: 12/20/2024] Open
Abstract
BACKGROUND Although generalized-dataset-based auto-segmentation models that consider various computed tomography (CT) scanners have shown great clinical potential, their application to medical images from unseen scanners remains challenging because of device-dependent image features. PURPOSE This study aims to investigate the performance of a device-dependent auto-segmentation model based on a combined dataset of a generalized dataset and single CT scanner dataset. METHOD We constructed two training datasets for 21 chest and abdominal organs. The generalized dataset comprised 1203 publicly available multi-scanner data. The device-dependent dataset comprised 1253 data, including the 1203 multi-CT scanner data and 50 single CT scanner data. Using these datasets, the generalized-dataset-based model (GDSM) and the device-dependent-dataset-based model (DDSM) were trained. The models were trained using nnU-Net and tested on ten data samples from a single CT scanner. The evaluation metrics included the Dice similarity coefficient (DSC), the Hausdorff distance (HD), and the average symmetric surface distance (ASSD), which were used to assess the overall performance of the models. In addition, DSCdiff, HDratio, and ASSDratio, which are variations of the three existing metrics, were used to compare the performance of the models across different organs. RESULT For the average DSC, the GDSM and DDSM had values of 0.9251 and 0.9323, respectively. For the average HD, the GDSM and DDSM had values of 10.66 and 9.139 mm, respectively; for the average ASSD, the GDSM and DDSM had values of 0.8318 and 0.6656 mm, respectively. Compared with the GDSM, the DDSM showed consistent performance improvements of 0.78%, 14%, and 20% for the DSC, HD, and ASSD metrics, respectively. In addition, compared with the GDSM, the DDSM had better DSCdiff values in 14 of 21 tested organs, better HDratio values in 13 of 21 tested organs, and better ASSDratio values in 14 of 21 tested organs. The three averages of the variant metrics were all better for the DDSM than for the GDSM. CONCLUSION The results suggest that combining the generalized dataset with a single scanner dataset resulted in an overall improvement in model performance for that device image.
Collapse
Affiliation(s)
- Hyeongjin Lim
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yongha Gi
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yousun Ko
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yunhui Jo
- Institute of Global Health Technology (IGHT), Korea University, Seoul, Republic of Korea
| | - Jinyoung Hong
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | | | - Sung Hwan Ahn
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Hee-Chul Park
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Haeyoung Kim
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Kwangzoo Chung
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Myonggeun Yoon
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
- FieldCure Ltd, Seoul, Republic of Korea
| |
Collapse
|
12
|
Lei M, Zhao J, Sahan AZ, Hu J, Zhou J, Lee H, Wu Q, Zhang J, Liu Z. Deep Learning Assisted Plasmonic Dark-Field Microscopy for Super-Resolution Label-Free Imaging. NANO LETTERS 2024; 24:15724-15730. [PMID: 39586837 PMCID: PMC11638943 DOI: 10.1021/acs.nanolett.4c04399] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2024] [Revised: 11/21/2024] [Accepted: 11/22/2024] [Indexed: 11/27/2024]
Abstract
Dark-field microscopy (DFM) is a widely used imaging tool, due to its high-contrast capability in imaging label-free specimens. Traditional DFM requires optical alignment to block the oblique illumination, and the resolution is diffraction-limited to the wavelength scale. In this work, we present deep-learning assisted plasmonic dark-field microscopy (DAPD), which is a single-frame super-resolution method using plasmonic dark-field (PDF) microscopy and deep-learning assisted image reconstruction. Specifically, we fabricated a designed PDF substrate with surface plasmon polaritons (SPPs) illuminating specimens on the substrate. Dark field images formed by scattered light from the specimen are further processed by a pretrained convolutional neural network (CNN) using a simulation dataset based on the designed substrate and parameters of the detection optics. We demonstrated a resolution enhancement of 2.8 times on various label-free objects with a large potential for future improvement. We highlight our technique as a compact alternative to traditional DFM with a significantly enhanced spatial resolution.
Collapse
Affiliation(s)
- Ming Lei
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Junxiang Zhao
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Ayse Z. Sahan
- Department
of Pharmacology, University of California−San
Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Biomedical
Sciences Graduate Program, University of
California, San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Jie Hu
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Junxiao Zhou
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Hongki Lee
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Qianyi Wu
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Jin Zhang
- Department
of Pharmacology, University of California−San
Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Department
of Chemistry and Biochemistry, University
of California, San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Shu
Chien − Gene Lay Department of Bioengineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Zhaowei Liu
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Materials
Science and Engineering, University of California−San
Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| |
Collapse
|
13
|
Singh D, Marathe A, Roy S, Walambe R, Kotecha K. Explainable rotation-invariant self-supervised representation learning. MethodsX 2024; 13:102959. [PMID: 39329154 PMCID: PMC11426157 DOI: 10.1016/j.mex.2024.102959] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2024] [Accepted: 09/11/2024] [Indexed: 09/28/2024] Open
Abstract
This paper describes a method that can perform robust detection and classification in out-of-distribution rotated images in the medical domain. In real-world medical imaging tools, noise due to the rotation of the body part is frequently observed. This noise reduces the accuracy of AI-based classification and prediction models. Hence, it is important to develop models which are rotation invariant. To that end, the proposed method - RISC (rotation invariant self-supervised vision framework) addresses this issue of rotational corruption. We present state-of-the-art rotation-invariant classification results and provide explainability for the performance in the domain. The evaluation of the proposed method is carried out on real-world adversarial examples in Medical Imagery-OrganAMNIST, RetinaMNIST and PneumoniaMNIST. It is observed that RISC outperforms the rotation-affected benchmark methods by obtaining 22\%, 17\% and 2\% accuracy boost on OrganAMNIST, PneumoniaMNIST and RetinaMNIST rotated baselines respectively. Further, explainability results are demonstrated. This methods paper describes:•a representation learning approach that can perform robust detection and classification in out-of-distribution rotated images in the medical domain.•It presents a method that incorporates self-supervised rotation invariance for correcting rotational corruptions.•GradCAM-based explainability for the rotational SSL pretext task and the downstream classification outcomes for the three benchmark datasets are presented.
Collapse
Affiliation(s)
- Devansh Singh
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, India
| | - Aboli Marathe
- Machine Learning Department, Carnegie Mellon University, Pittsburgh, PA, USA
| | - Sidharth Roy
- Department of Computer and Information Science, University of Pennsylvania, USA
| | - Rahee Walambe
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, India
| | - Ketan Kotecha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, India
| |
Collapse
|
14
|
Carvalho BKG, Nolden EL, Wenning AS, Kiss-Dala S, Agócs G, Róth I, Kerémi B, Géczi Z, Hegyi P, Kivovics M. Diagnostic accuracy of artificial intelligence for approximal caries on bitewing radiographs: A systematic review and meta-analysis. J Dent 2024; 151:105388. [PMID: 39396775 DOI: 10.1016/j.jdent.2024.105388] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Revised: 09/13/2024] [Accepted: 10/01/2024] [Indexed: 10/15/2024] Open
Abstract
OBJECTIVES This systematic review and meta-analysis aimed to investigate the diagnostic accuracy of Artificial Intelligence (AI) for approximal carious lesions on bitewing radiographs. METHODS This study included randomized controlled trials (RCTs) and non-randomized controlled trials (non-RCTs) reporting on the diagnostic accuracy of AI for approximal carious lesions on bitewing radiographs. The risk of bias was assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool. A systematic search was conducted on November 4, 2023, in PubMed, Cochrane, and Embase databases and an updated search was performed on August 28, 2024. The primary outcomes assessed were sensitivity, specificity, and overall accuracy. Sensitivity and specificity were pooled using a bivariate model. RESULTS Of the 2,442 studies identified, 21 met the inclusion criteria. The pooled sensitivity and specificity of AI were 0.94 (confidence interval (CI): ± 0.78-0.99) and 0.91 (CI: ± 0.84-0.95), respectively. The positive predictive value (PPV) ranged from 0.15 to 0.87, indicating a moderate capacity for identifying true positives among decayed teeth. The negative predictive value (NPV) ranged from 0.79 to 1.00, demonstrating a high ability to exclude healthy teeth. The diagnostic odds ratio was high, indicating strong overall diagnostic performance. CONCLUSIONS AI models demonstrate clinically acceptable diagnostic accuracy for approximal caries on bitewing radiographs. Although AI can be valuable for preliminary screening, positive findings should be verified by dental experts to prevent unnecessary treatments and ensure timely diagnosis. AI models are highly reliable in excluding healthy approximal surfaces. CLINICAL SIGNIFICANCE AI can assist dentists in detecting approximal caries on bitewing radiographs. However, expert supervision is required to prevent iatrogenic damage and ensure timely diagnosis.
Collapse
Affiliation(s)
| | - Elias-Leon Nolden
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary
| | - Alexander Schulze Wenning
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary
| | - Szilvia Kiss-Dala
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary
| | - Gergely Agócs
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Biophysics and Radiation Biology, Semmelweis University, Tűzoltó utca 37-47, 1072, Budapest, Hungary
| | - Ivett Róth
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Prosthodontics, Semmelweis University, Szentkirályi utca 47 1088, Budapest, Hungary
| | - Beáta Kerémi
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Restorative Dentistry and Endodontics, Semmelweis University, Szentkirályi utca 47, 1088, Budapest, Hungary
| | - Zoltán Géczi
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Prosthodontics, Semmelweis University, Szentkirályi utca 47 1088, Budapest, Hungary
| | - Péter Hegyi
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Institute of Pancreatic Diseases, Semmelweis University, Tömő utca 25-29 1083, Budapest, Hungary; Institute for Translational Medicine, Medical School, University of Pécs, Szigeti utca 12 7624, Pécs, Hungary
| | - Márton Kivovics
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Community Dentistry, Semmelweis University, Szentkirályi utca 40 1088, Budapest, Hungary.
| |
Collapse
|
15
|
Song S, Wang J, Wang Z, Wang H, Su J, Ding X, Dang K. DualStreamFoveaNet: A Dual Stream Fusion Architecture With Anatomical Awareness for Robust Fovea Localization. IEEE J Biomed Health Inform 2024; 28:7217-7229. [PMID: 39150813 DOI: 10.1109/jbhi.2024.3445112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/18/2024]
Abstract
Accurate fovea localization is essential for analyzing retinal diseases to prevent irreversible vision loss. While current deep learning-based methods outperform traditional ones, they still face challenges such as the lack of local anatomical landmarks around the fovea, the inability to robustly handle diseased retinal images, and the variations in image conditions. In this paper, we propose a novel transformer-based architecture called DualStreamFoveaNet (DSFN) for multi-cue fusion. This architecture explicitly incorporates long-range connections and global features using retina and vessel distributions for robust fovea localization. We introduce a spatial attention mechanism in the dual-stream encoder to extract and fuse self-learned anatomical information, focusing more on features distributed along blood vessels and significantly reducing computational costs by decreasing token numbers. Our extensive experiments show that the proposed architecture achieves state-of-the-art performance on two public datasets and one large-scale private dataset. Furthermore, we demonstrate that the DSFN is more robust on both normal and diseased retina images and has better generalization capacity in cross-dataset experiments.
Collapse
|
16
|
Raina D, Chandrashekhara SH, Voyles R, Wachs J, Saha SK. Deep-Learning Model for Quality Assessment of Urinary Bladder Ultrasound Images Using Multiscale and Higher-Order Processing. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2024; 71:1451-1463. [PMID: 38598406 DOI: 10.1109/tuffc.2024.3386919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/12/2024]
Abstract
Autonomous ultrasound image quality assessment (US-IQA) is a promising tool to aid the interpretation by practicing sonographers and to enable the future robotization of ultrasound procedures. However, autonomous US-IQA has several challenges. Ultrasound images contain many spurious artifacts, such as noise due to handheld probe positioning, errors in the selection of probe parameters, and patient respiration during the procedure. Furthermore, these images are highly variable in appearance with respect to the individual patient's physiology. We propose to use a deep convolutional neural network (CNN), USQNet, which utilizes a multiscale and local-to-global second-order pooling (MS-L2GSoP) classifier to conduct the sonographer-like assessment of image quality. This classifier first extracts features at multiple scales to encode the interpatient anatomical variations, similar to a sonographer's understanding of anatomy. Then, it uses SoP in the intermediate layers (local) and at the end of the network (global) to exploit the second-order statistical dependency of MS structural and multiregion textural features. The L2GSoP will capture the higher-order relationships between different spatial locations and provide the seed for correlating local patches, much like a sonographer prioritizes regions across the image. We experimentally validated the USQNet for a new dataset of the human urinary bladder (UB) ultrasound images. The validation involved first with the subjective assessment by experienced radiologists' annotation, and then with state-of-the-art (SOTA) CNN networks for US-IQA and its ablated counterparts. The results demonstrate that USQNet achieves a remarkable accuracy of 92.4% and outperforms the SOTA models by 3%-14% while requiring comparable computation time.
Collapse
|
17
|
Fantini M, Ciravegna G, Koudounas A, Cerquitelli T, Baralis E, Succo G, Crosetti E. The Rapidly Evolving Scenario of Acoustic Voice Analysis in Otolaryngology. Cureus 2024; 16:e73491. [PMID: 39669823 PMCID: PMC11635181 DOI: 10.7759/cureus.73491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/11/2024] [Indexed: 12/14/2024] Open
Abstract
The field of voice analysis has experienced significant transformations, evolving from basic perceptual assessments to the incorporation of advanced digital signal processing and computational tools. This progression has facilitated a deeper understanding of the complex dynamics of vocal function, particularly through the use of acoustic voice analysis within a multidimensional evaluation framework. Traditionally, voice analysis relied on parameters such as fundamental frequency, jitter, shimmer, and noise-to-harmonic ratio, which, despite their utility, have faced criticism for variability and lack of robustness. Recent developments have led to a shift toward more reliable metrics such as cepstral measures, which offer improved accuracy in voice quality assessments. Furthermore, the integration of multiparametric constructs underscores a comprehensive approach to evaluating vocal quality, blending sustained vowels, and continuous speech analyses. Current trends in clinical practice increasingly favor these advanced measures over traditional parameters due to their greater reliability and clinical utility. Additionally, the emergence of artificial intelligence (AI), particularly deep learning, holds promise for revolutionizing voice analysis by enhancing diagnostic precision and enabling efficient, non-invasive screening methods. This shift toward AI-driven approaches signifies a potential paradigm change in voice health, suggesting a future where AI not only aids in diagnosis but also the early detection and treatment of voice-related pathologies.
Collapse
Affiliation(s)
- Marco Fantini
- Ear, Nose, and Throat Unit, Koelliker Hospital, Turin, ITA
- Ear, Nose, and Throat Unit, San Feliciano Hospital, Rome, Italy
| | - Gabriele Ciravegna
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Alkis Koudounas
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Tania Cerquitelli
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Elena Baralis
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Giovanni Succo
- Ear, Nose, and Throat Clinic, Head and Neck Cancer Unit, San Giovanni Bosco Hospital, Turin, ITA
- Department of Oncology, University of Turin, Turin, ITA
| | - Erika Crosetti
- Ear, Nose, and Throat Clinic, Head and Neck Cancer Unit, San Giovanni Bosco Hospital, Turin, ITA
| |
Collapse
|
18
|
Báskay J, Kivovics M, Pénzes D, Kontsek E, Pesti A, Kiss A, Szócska M, Németh O, Pollner P. Reconstructing 3D histological structures using machine learning (artificial intelligence) algorithms. PATHOLOGIE (HEIDELBERG, GERMANY) 2024; 45:67-73. [PMID: 39570395 DOI: 10.1007/s00292-024-01387-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 10/11/2024] [Indexed: 11/22/2024]
Abstract
BACKGROUND Histomorphometry is currently the gold standard for bone microarchitectural examinations. This relies on two-dimensional (2D) sections to deduce the spatial properties of structures. Micromorphometric parameters are calculated from these sections based on the assumption of a plate-like 3D microarchitecture, resulting in the loss of 3D structure due to the destructive nature of classical histological processing. MATERIALS AND METHODS To overcome the limitation of histomorphometry and reconstruct the 3D architecture of bone core biopsy samples from 2D histological sections, bone core biopsy samples were decalcified and embedded in paraffin. Subsequently, 5 µm thick serial sections were stained with hematoxylin and eosin and scanned using a 3DHISTECH PANNORAMIC 1000 Digital Slide Scanner (3DHISTECH, Budapest, Hungary). A modified U‑Net architecture was trained to categorize tissues on the sections. LoFTR feature matching combined with affine transformations was employed to create the histologic reconstruction. Micromorphometric parameters were calculated using Bruker's CTAn software (v. 1.18.8.0, Bruker, Kontich, Belgium) for both histological and microCT datasets. RESULTS Our method achieved an overall accuracy of 95.26% (95% confidence interval (CI): [94.15%, 96.37%]) with an F‑score of 0.9320 (95% CI: [0.9211, 0.9429]) averaged across all classes. Correlation coefficients between micromorphometric parameters measured on microCT imaging and histological reconstruction showed a strong linear relationship, with Spearman's ρ‑values of 0.777, 0.717, 0.705, 0.666, and 0.687 for bone volume/tissue volume (BV/TV), bone surface/TV, trabecular pattern factor, trabecular thickness, and trabecular separation, respectively. Bland-Altman and mountain plots indicated good agreement between the methods for BV/TV measurements. CONCLUSION This method enables examination of tissue microarchitecture in 3D with an even higher resolution than microcomputed tomography (microCT), without losing information on cellularity. However, the limitation of this procedure is its destructive nature, which precludes subsequent mechanical testing of the sample or any further secondary measurements. Furthermore, the number of histological sections that can be created from a single sample is limited.
Collapse
Affiliation(s)
- J Báskay
- Department of Biological Physics, Eötvös Loránd University, Pázmány Péter Sétány 1/a, 1117, Budapest, Hungary
| | - M Kivovics
- Department of Community Dentistry, Semmelweis University, Szentkirályi Utca 40, 1088, Budapest, Hungary
| | - D Pénzes
- Department of Community Dentistry, Semmelweis University, Szentkirályi Utca 40, 1088, Budapest, Hungary
| | - E Kontsek
- Department of Pathology, Forensic and Insurance Medicine, Semmelweis University, Üllői út 93, 1091, Budapest, Hungary.
| | - A Pesti
- Department of Pathology, Forensic and Insurance Medicine, Semmelweis University, Üllői út 93, 1091, Budapest, Hungary
| | - A Kiss
- Department of Pathology, Forensic and Insurance Medicine, Semmelweis University, Üllői út 93, 1091, Budapest, Hungary
| | - M Szócska
- Data-Driven Health Division of National Laboratory for Health Security, Health Services Management Training Centre, Semmelweis University, Kútvölgyi út 2, 1125, Budapest, Hungary
| | - O Németh
- Department of Community Dentistry, Semmelweis University, Szentkirályi Utca 40, 1088, Budapest, Hungary
| | - P Pollner
- Department of Biological Physics, Eötvös Loránd University, Pázmány Péter Sétány 1/a, 1117, Budapest, Hungary
| |
Collapse
|
19
|
Soliman-Aboumarie H, Geers J, Lowcock D, Suji T, Kok K, Cameli M, Galiatsou E. Artificial intelligence-assisted focused cardiac ultrasound training: A survey among undergraduate medical students. ULTRASOUND (LEEDS, ENGLAND) 2024:1742271X241287923. [PMID: 39555149 PMCID: PMC11563526 DOI: 10.1177/1742271x241287923] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/06/2024] [Accepted: 09/05/2024] [Indexed: 11/19/2024]
Abstract
Objectives Focused cardiac ultrasound (FoCUS) is increasingly applied in many specialities, and adequate education and training of physicians is therefore mandatory. This study aimed to assess the impact of artificial intelligence (AI)-assisted interactive focused cardiac ultrasound (FoCUS) teaching session on undergraduate medical students' confidence level and knowledge in cardiac ultrasound. Methods The AI-assisted interactive FoCUS teaching session was held during the 9th National Undergraduate Cardiovascular Conference in London in March 2023 and all undergraduate medical students were invited to attend, and 79 students enrolled and attended the training. Two workshops were conducted each over 3-hour period. Each workshop consisted of a theoretical lecture followed by a supervised hands-on session by experts, first workshop trained 39 students and the second workshop trained 40 students. The students' pre- and post-session knowledge and confidence levels were assessed by Likert-type-scale questionnaires filled in by the students before and immediately after the workshop. Results A total of 61 pre-session and 52 post-session questionnaires were completed. Confidence level in ultrasound skills increased significantly for all six domains after the workshop, with the greatest improvement seen in obtaining basic cardiac views (p < 0.001 for all six domains). Students strongly agreed about the effectiveness of the teaching session and supported the integration of ultrasound training into their medical curriculum. Conclusions AI-assisted interactive FoCUS training can be an effective and powerful tool to increase ultrasound skills and confidence levels of undergraduate medical students. Integration of such ultrasound courses into the medical curriculum should therefore be considered.
Collapse
Affiliation(s)
- Hatem Soliman-Aboumarie
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
- School of Cardiovascular and Metabolic Medicine & Sciences, King’s College London, London, UK
| | - Jolien Geers
- Department of Cardiology, The Brussels University Hospital, Brussels, Belgium
| | - Dominic Lowcock
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
| | - Trisha Suji
- School of Cardiovascular and Metabolic Medicine & Sciences, King’s College London, London, UK
| | - Kimberley Kok
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
| | - Matteo Cameli
- School of Cardiovascular Medicine, University of Siena, Siena, Italy
| | - Eftychia Galiatsou
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
| |
Collapse
|
20
|
Chiu IM, Chen TY, Zheng YC, Lin XH, Cheng FJ, Ouyang D, Cheng CY. Prospective clinical evaluation of deep learning for ultrasonographic screening of abdominal aortic aneurysms. NPJ Digit Med 2024; 7:282. [PMID: 39406888 PMCID: PMC11480325 DOI: 10.1038/s41746-024-01269-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Accepted: 09/23/2024] [Indexed: 10/19/2024] Open
Abstract
Abdominal aortic aneurysm (AAA) often remains undetected until rupture due to limited access to diagnostic ultrasound. This trial evaluated a deep learning (DL) algorithm to guide AAA screening by novice nurses with no prior ultrasonography experience. Ten nurses performed 15 scans each on patients over 65, assisted by a DL object detection algorithm, and compared against physician-performed scans. Ultrasound scan quality, assessed by three blinded expert physicians, was the primary outcome. Among 184 patients, DL-guided novices achieved adequate scan quality in 87.5% of cases, comparable to the 91.3% by physicians (p = 0.310). The DL model predicted AAA with an AUC of 0.975, 100% sensitivity, and 97.8% specificity, with a mean absolute error of 2.8 mm in predicting aortic width compared to physicians. This study demonstrates that DL-guided POCUS has the potential to democratize AAA screening, offering performance comparable to experienced physicians and improving early detection.
Collapse
Affiliation(s)
- I-Min Chiu
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
- Department of Cardiology, Smidt Heart Institute, Cedars-Sinai Medical Center, Los Angeles, CA, USA
- Department of Computer Science and Engineering, National Sun Yat-sen University, Kaohsiung, Taiwan
| | - Tien-Yu Chen
- Division of Cardiology, Department of Internal Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - You-Cheng Zheng
- Division of Cardiology, Department of Internal Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - Xin-Hong Lin
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - Fu-Jen Cheng
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - David Ouyang
- Department of Cardiology, Smidt Heart Institute, Cedars-Sinai Medical Center, Los Angeles, CA, USA.
| | - Chi-Yung Cheng
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan.
- Department of Computer Science and Engineering, National Sun Yat-sen University, Kaohsiung, Taiwan.
| |
Collapse
|
21
|
Bargagna F, Zigrino D, De Santi LA, Genovesi D, Scipioni M, Favilli B, Vergaro G, Emdin M, Giorgetti A, Positano V, Santarelli MF. Automated Neural Architecture Search for Cardiac Amyloidosis Classification from [18F]-Florbetaben PET Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01275-8. [PMID: 39356368 DOI: 10.1007/s10278-024-01275-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/18/2024] [Revised: 08/30/2024] [Accepted: 09/08/2024] [Indexed: 10/03/2024]
Abstract
Medical image classification using convolutional neural networks (CNNs) is promising but often requires extensive manual tuning for optimal model definition. Neural architecture search (NAS) automates this process, reducing human intervention significantly. This study applies NAS to [18F]-Florbetaben PET cardiac images for classifying cardiac amyloidosis (CA) sub-types (amyloid light chain (AL) and transthyretin amyloid (ATTR)) and controls. Following data preprocessing and augmentation, an evolutionary cell-based NAS approach with a fixed network macro-structure is employed, automatically deriving cells' micro-structure. The algorithm is executed five times, evaluating 100 mutating architectures per run on an augmented dataset of 4048 images (originally 597), totaling 5000 architectures evaluated. The best network (NAS-Net) achieves 76.95% overall accuracy. K-fold analysis yields mean ± SD percentages of sensitivity, specificity, and accuracy on the test dataset: AL subjects (98.7 ± 2.9, 99.3 ± 1.1, 99.7 ± 0.7), ATTR-CA subjects (93.3 ± 7.8, 78.0 ± 2.9, 70.9 ± 3.7), and controls (35.8 ± 14.6, 77.1 ± 2.0, 96.7 ± 4.4). NAS-derived network performance rivals manually determined networks in the literature while using fewer parameters, validating its automatic approach's efficacy.
Collapse
Affiliation(s)
- Filippo Bargagna
- Department of Information Engineering, University of Pisa, Via G. Caruso 16, 56122, Pisa, Italy.
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy.
| | - Donato Zigrino
- Department of Information Engineering, University of Pisa, Via G. Caruso 16, 56122, Pisa, Italy
| | - Lisa Anita De Santi
- Department of Information Engineering, University of Pisa, Via G. Caruso 16, 56122, Pisa, Italy
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Dario Genovesi
- Nuclear Medicine Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Michele Scipioni
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital and Harvard Medical School, Boston, MA, USA
| | - Brunella Favilli
- Nuclear Medicine Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Giuseppe Vergaro
- Division of Cardiology and Cardiovascular Medicine, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Michele Emdin
- Division of Cardiology and Cardiovascular Medicine, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
- Health Science Interdisciplinary Center, Scuola Universitaria Superiore 'S. Anna", Piazza Martiri della Libertà 33, 56127, Pisa, Italy
| | - Assuero Giorgetti
- Nuclear Medicine Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Vincenzo Positano
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Maria Filomena Santarelli
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
- CNR Institute of Clinical Physiology, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| |
Collapse
|
22
|
Wu J, Li R, Gan J, Zheng Q, Wang G, Tao W, Yang M, Li W, Ji G, Li W. Application of artificial intelligence in lung cancer screening: A real-world study in a Chinese physical examination population. Thorac Cancer 2024; 15:2061-2072. [PMID: 39206529 PMCID: PMC11444925 DOI: 10.1111/1759-7714.15428] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2024] [Revised: 07/29/2024] [Accepted: 07/31/2024] [Indexed: 09/04/2024] Open
Abstract
BACKGROUND With the rapid increase of chest computed tomography (CT) images, the workload faced by radiologists has increased dramatically. It is undeniable that the use of artificial intelligence (AI) image-assisted diagnosis system in clinical treatment is a major trend in medical development. Therefore, in order to explore the value and diagnostic accuracy of the current AI system in clinical application, we aim to compare the detection and differentiation of benign and malignant pulmonary nodules between AI system and physicians, so as to provide a theoretical basis for clinical application. METHODS Our study encompassed a cohort of 23 336 patients who underwent chest low-dose spiral CT screening for lung cancer at the Health Management Center of West China Hospital. We conducted a comparative analysis between AI-assisted reading and manual interpretation, focusing on the detection and differentiation of benign and malignant pulmonary nodules. RESULTS The AI-assisted reading exhibited a significantly higher screening positive rate and probability of diagnosing malignant pulmonary nodules compared with manual interpretation (p < 0.001). Moreover, AI scanning demonstrated a markedly superior detection rate of malignant pulmonary nodules compared with manual scanning (97.2% vs. 86.4%, p < 0.001). Additionally, the lung cancer detection rate was substantially higher in the AI reading group compared with the manual reading group (98.9% vs. 90.3%, p < 0.001). CONCLUSIONS Our findings underscore the superior screening positive rate and lung cancer detection rate achieved through AI-assisted reading compared with manual interpretation. Thus, AI exhibits considerable potential as an adjunctive tool in lung cancer screening within clinical practice settings.
Collapse
Affiliation(s)
- Jiaxuan Wu
- Department of Pulmonary and Critical Care Medicine, West China HospitalSichuan UniversityChengduSichuanChina
- State Key Laboratory of Respiratory Health and MultimorbidityWest China HospitalChengduSichuanChina
- Institute of Respiratory Health and Multimorbidity, West China HospitalSichuan UniversityChengduSichuanChina
| | - Ruicen Li
- Health Management Center, General Practice Medical Center, West China HospitalSichuan UniversityChengduChina
| | - Jiadi Gan
- Department of Pulmonary and Critical Care Medicine, West China HospitalSichuan UniversityChengduSichuanChina
- State Key Laboratory of Respiratory Health and MultimorbidityWest China HospitalChengduSichuanChina
- Institute of Respiratory Health and Multimorbidity, West China HospitalSichuan UniversityChengduSichuanChina
| | - Qian Zheng
- West China Clinical Medical CollegeSichuan UniversityChengduChina
| | - Guoqing Wang
- State Key Laboratory of Biotherapy and Cancer Center, West China HospitalSichuan UniversityChengduSichuanChina
| | - Wenjuan Tao
- Institute of Hospital Management, West China HospitalSichuan UniversityChengduChina
| | - Ming Yang
- National Clinical Research Center for Geriatrics (WCH), West China HospitalSichuan UniversityChengduChina
- Center of Gerontology and Geriatrics, West China HospitalSichuan UniversityChengduChina
| | - Wenyu Li
- Health Management Center, General Practice Medical Center, West China HospitalSichuan UniversityChengduChina
| | - Guiyi Ji
- Health Management Center, General Practice Medical Center, West China HospitalSichuan UniversityChengduChina
| | - Weimin Li
- Department of Pulmonary and Critical Care Medicine, West China HospitalSichuan UniversityChengduSichuanChina
- State Key Laboratory of Respiratory Health and MultimorbidityWest China HospitalChengduSichuanChina
- Institute of Respiratory Health and Multimorbidity, West China HospitalSichuan UniversityChengduSichuanChina
- Institute of Respiratory Health, Frontiers Science Center for Disease‐related Molecular Network, West China HospitalSichuan UniversityChengduSichuanChina
- Precision Medicine Center, Precision Medicine Key Laboratory of Sichuan Province, West China HospitalSichuan UniversityChengduSichuanChina
- The Research Units of West China, Chinese Academy of Medical SciencesWest China HospitalChengduSichuanChina
| |
Collapse
|
23
|
Yıldız Potter İ, Rodriguez EK, Wu J, Nazarian A, Vaziri A. An Automated Vertebrae Localization, Segmentation, and Osteoporotic Compression Fracture Detection Pipeline for Computed Tomographic Imaging. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2428-2443. [PMID: 38717516 PMCID: PMC11522205 DOI: 10.1007/s10278-024-01135-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Revised: 04/30/2024] [Accepted: 05/01/2024] [Indexed: 06/29/2024]
Abstract
Osteoporosis is the most common chronic metabolic bone disease worldwide. Vertebral compression fracture (VCF) is the most common type of osteoporotic fracture. Approximately 700,000 osteoporotic VCFs are diagnosed annually in the USA alone, resulting in an annual economic burden of ~$13.8B. With an aging population, the rate of osteoporotic VCFs and their associated burdens are expected to rise. Those burdens include pain, functional impairment, and increased medical expenditure. Therefore, it is of utmost importance to develop an analytical tool to aid in the identification of VCFs. Computed Tomography (CT) imaging is commonly used to detect occult injuries. Unlike the existing VCF detection approaches based on CT, the standard clinical criteria for determining VCF relies on the shape of vertebrae, such as loss of vertebral body height. We developed a novel automated vertebrae localization, segmentation, and osteoporotic VCF detection pipeline for CT scans using state-of-the-art deep learning models to bridge this gap. To do so, we employed a publicly available dataset of spine CT scans with 325 scans annotated for segmentation, 126 of which also graded for VCF (81 with VCFs and 45 without VCFs). Our approach attained 96% sensitivity and 81% specificity in detecting VCF at the vertebral-level, and 100% accuracy at the subject-level, outperforming deep learning counterparts tested for VCF detection without segmentation. Crucially, we showed that adding predicted vertebrae segments as inputs significantly improved VCF detection at both vertebral and subject levels by up to 14% Sensitivity and 20% Specificity (p-value = 0.028).
Collapse
Affiliation(s)
| | - Edward K Rodriguez
- Carl J. Shapiro Department of Orthopedic Surgery, Beth Israel Deaconess Medical Center (BIDMC), Harvard Medical School, 330 Brookline Avenue, Stoneman 10, Boston, MA, 02215, USA
- Musculoskeletal Translational Innovation Initiative, Beth Israel Deaconess Medical Center, Harvard Medical School, 330 Brookline Avenue, RN123, Boston, MA, 02215, USA
| | - Jim Wu
- Department of Radiology, Beth Israel Deaconess Medical Center (BIDMC), Harvard Medical School, 330 Brookline Avenue, Shapiro 4, Boston, MA, 02215, USA
| | - Ara Nazarian
- Carl J. Shapiro Department of Orthopedic Surgery, Beth Israel Deaconess Medical Center (BIDMC), Harvard Medical School, 330 Brookline Avenue, Stoneman 10, Boston, MA, 02215, USA
- Musculoskeletal Translational Innovation Initiative, Beth Israel Deaconess Medical Center, Harvard Medical School, 330 Brookline Avenue, RN123, Boston, MA, 02215, USA
- Department of Orthopaedics Surgery, Yerevan State University, Yerevan, Armenia
| | - Ashkan Vaziri
- BioSensics, LLC, 57 Chapel Street, Newton, MA, 02458, USA
| |
Collapse
|
24
|
Li Y, Cai P, Huang Y, Yu W, Liu Z, Liu P. Deep learning based detection and classification of fetal lip in ultrasound images. J Perinat Med 2024; 52:769-777. [PMID: 39028804 DOI: 10.1515/jpm-2024-0122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/19/2024] [Accepted: 07/07/2024] [Indexed: 07/21/2024]
Abstract
OBJECTIVES Fetal cleft lip is a common congenital defect. Considering the delicacy and difficulty of observing fetal lips, we have utilized deep learning technology to develop a new model aimed at quickly and accurately assessing the development of fetal lips during prenatal examinations. This model can detect ultrasound images of the fetal lips and classify them, aiming to provide a more objective prediction for the development of fetal lips. METHODS This study included 632 pregnant women in their mid-pregnancy stage, who underwent ultrasound examinations of the fetal lips, collecting both normal and abnormal fetal lip ultrasound images. To improve the accuracy of the detection and classification of fetal lips, we proposed and validated the Yolov5-ECA model. RESULTS The experimental results show that, compared with the currently popular 10 models, our model achieved the best results in the detection and classification of fetal lips. In terms of the detection of fetal lips, the mean average precision (mAP) at 0.5 and mAP at 0.5:0.95 were 0.920 and 0.630, respectively. In the classification of fetal lip ultrasound images, the accuracy reached 0.925. CONCLUSIONS The deep learning algorithm has accuracy consistent with manual evaluation in the detection and classification process of fetal lips. This automated recognition technology can provide a powerful tool for inexperienced young doctors, helping them to accurately conduct examinations and diagnoses of fetal lips.
Collapse
Affiliation(s)
- Yapeng Li
- School of Medicine, Huaqiao University, Quanzhou, China
| | - Peiya Cai
- Department of Gynecology and Obstetrics, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, China
| | - Yubing Huang
- Department of Ultrasound, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, China
| | - Weifeng Yu
- Department of Ultrasound, Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou, China
| | - Zhonghua Liu
- Department of Ultrasound, Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou, China
| | - Peizhong Liu
- School of Medicine, Huaqiao University, Quanzhou, China
- College of Engineering, Huaqiao University, Quanzhou, China
| |
Collapse
|
25
|
Shah STH, Shah SAH, Khan II, Imran A, Shah SBH, Mehmood A, Qureshi SA, Raza M, Di Terlizzi A, Cavaglià M, Deriu MA. Data-driven classification and explainable-AI in the field of lung imaging. Front Big Data 2024; 7:1393758. [PMID: 39364222 PMCID: PMC11446784 DOI: 10.3389/fdata.2024.1393758] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Accepted: 09/03/2024] [Indexed: 10/05/2024] Open
Abstract
Detecting lung diseases in medical images can be quite challenging for radiologists. In some cases, even experienced experts may struggle with accurately diagnosing chest diseases, leading to potential inaccuracies due to complex or unseen biomarkers. This review paper delves into various datasets and machine learning techniques employed in recent research for lung disease classification, focusing on pneumonia analysis using chest X-ray images. We explore conventional machine learning methods, pretrained deep learning models, customized convolutional neural networks (CNNs), and ensemble methods. A comprehensive comparison of different classification approaches is presented, encompassing data acquisition, preprocessing, feature extraction, and classification using machine vision, machine and deep learning, and explainable-AI (XAI). Our analysis highlights the superior performance of transfer learning-based methods using CNNs and ensemble models/features for lung disease classification. In addition, our comprehensive review offers insights for researchers in other medical domains too who utilize radiological images. By providing a thorough overview of various techniques, our work enables the establishment of effective strategies and identification of suitable methods for a wide range of challenges. Currently, beyond traditional evaluation metrics, researchers emphasize the importance of XAI techniques in machine and deep learning models and their applications in classification tasks. This incorporation helps in gaining a deeper understanding of their decision-making processes, leading to improved trust, transparency, and overall clinical decision-making. Our comprehensive review serves as a valuable resource for researchers and practitioners seeking not only to advance the field of lung disease detection using machine learning and XAI but also from other diverse domains.
Collapse
Affiliation(s)
- Syed Taimoor Hussain Shah
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
| | - Syed Adil Hussain Shah
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
- Department of Research and Development (R&D), GPI SpA, Trento, Italy
| | - Iqra Iqbal Khan
- Department of Computer Science, Bahauddin Zakariya University, Multan, Pakistan
| | - Atif Imran
- College of Electrical and Mechanical Engineering, National University of Sciences and Technology, Rawalpindi, Pakistan
| | - Syed Baqir Hussain Shah
- Department of Computer Science, Commission on Science and Technology for Sustainable Development in the South (COMSATS) University Islamabad (CUI), Wah Campus, Wah, Pakistan
| | - Atif Mehmood
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua, China
- Zhejiang Institute of Photoelectronics & Zhejiang Institute for Advanced Light Source, Zhejiang Normal University, Jinhua, Zhejiang, China
| | - Shahzad Ahmad Qureshi
- Department of Computer and Information Sciences, Pakistan Institute of Engineering and Applied Sciences (PIEAS), Islamabad, Pakistan
| | - Mudassar Raza
- Department of Computer Science, Namal University Mianwali, Mianwali, Pakistan
- Department of Computer Science, Heavy Industries Taxila Education City (HITEC), University of Taxila, Taxila, Pakistan
| | | | - Marco Cavaglià
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
| | - Marco Agostino Deriu
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
| |
Collapse
|
26
|
Li F, Xu Y, Lemus OD, Wang TJC, Sisti MB, Wuu CS. Synthetic CT for gamma knife radiosurgery dose calculation: A feasibility study. Phys Med 2024; 125:104504. [PMID: 39197262 DOI: 10.1016/j.ejmp.2024.104504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Revised: 06/24/2024] [Accepted: 08/22/2024] [Indexed: 09/01/2024] Open
Abstract
PURPOSE To determine if MRI-based synthetic CTs (sCT), generated with no predefined pulse sequence, can be used for inhomogeneity correction in routine gamma knife radiosurgery (GKRS) treatment planning dose calculation. METHODS Two sets of sCTs were generated from T1post and T2 images using cycleGAN. Twenty-eight patients (18 training, 10 validation) were retrospectively selected. The image quality of the generated sCTs was compared with the original CT (oCT) regarding the HU value preservation using histogram comparison, RMSE and MAE, and structural integrity. Dosimetric comparisons were also made among GKRS plans from 3 calculation approaches: TMR10 (oCT), and convolution (oCT and sCT), at four locations: original disease site, bone/tissue interface, air/tissue interface, and mid-brain. RESULTS The study showed that sCTs and oCTs' HU were similar, with T2-sCT performing better. TMR10 significantly underdosed the target by a mean of 5.4% compared to the convolution algorithm. There was no significant difference in convolution algorithm shot time between the oCT and sCT generated with T2. The highest and lowest dosimetric differences between the two CTs were observed in the bone and air interface, respectively. Dosimetric differences of 3.3% were observed in sCT predicted from MRI with stereotactic frames, which was not included in the training sets. CONCLUSIONS MRI-based sCT can be utilized for GKRS convolution dose calculation without the unnecessary radiation dose, and sCT without metal artifacts could be generated in framed cases. Larger datasets inclusive of all pulse sequences can improve the training set. Further investigation and validation studies are needed before clinical implementation.
Collapse
Affiliation(s)
- Fiona Li
- Department of Radiation Oncology, Columbia University, New York, NY, USA.
| | - Yuanguang Xu
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| | - Olga D Lemus
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| | - Tony J C Wang
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| | - Michael B Sisti
- Department of Neurological Surgery, Columbia University, New York, NY, USA
| | - Cheng-Shie Wuu
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| |
Collapse
|
27
|
Cai M, Zhao L, Qiang Y, Wang L, Zhao J. CHNet: A multi-task global-local Collaborative Hybrid Network for KRAS mutation status prediction in colorectal cancer. Artif Intell Med 2024; 155:102931. [PMID: 39094228 DOI: 10.1016/j.artmed.2024.102931] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 06/29/2024] [Accepted: 07/03/2024] [Indexed: 08/04/2024]
Abstract
Accurate prediction of Kirsten rat sarcoma (KRAS) mutation status is crucial for personalized treatment of advanced colorectal cancer patients. However, despite the excellent performance of deep learning models in certain aspects, they often overlook the synergistic promotion among multiple tasks and the consideration of both global and local information, which can significantly reduce prediction accuracy. To address these issues, this paper proposes an innovative method called the Multi-task Global-Local Collaborative Hybrid Network (CHNet) aimed at more accurately predicting patients' KRAS mutation status. CHNet consists of two branches that can extract global and local features from segmentation and classification tasks, respectively, and exchange complementary information to collaborate in executing these tasks. Within the two branches, we have designed a Channel-wise Hybrid Transformer (CHT) and a Spatial-wise Hybrid Transformer (SHT). These transformers integrate the advantages of both Transformer and CNN, employing cascaded hybrid attention and convolution to capture global and local information from the two tasks. Additionally, we have created an Adaptive Collaborative Attention (ACA) module to facilitate the collaborative fusion of segmentation and classification features through guidance. Furthermore, we introduce a novel Class Activation Map (CAM) loss to encourage CHNet to learn complementary information between the two tasks. We evaluate CHNet on the T2-weighted MRI dataset, and achieve an accuracy of 88.93% in KRAS mutation status prediction, which outperforms the performance of representative KRAS mutation status prediction methods. The results suggest that our CHNet can more accurately predict KRAS mutation status in patients via a multi-task collaborative facilitation and considering global-local information way, which can assist doctors in formulating more personalized treatment strategies for patients.
Collapse
Affiliation(s)
- Meiling Cai
- College of computer science and technology (College of data science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China.
| | - Lin Zhao
- Southeast University, Nanjing, 210037, Jiangsu, China
| | - Yan Qiang
- College of computer science and technology (College of data science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China
| | - Long Wang
- Jinzhong College of Information, Jinzhong, 030800, Shanxi, China
| | - Juanjuan Zhao
- College of computer science and technology (College of data science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China.
| |
Collapse
|
28
|
Li K, Han X, Meng Y, Li J, Hong Y, Chen X, You JY, Yao L, Hu W, Xia Z, Ke G, Zhang L, Zhang J, Zhao X. Single-Image-Based Deep Learning for Precise Atomic Defect Identification. NANO LETTERS 2024; 24:10275-10283. [PMID: 39106329 DOI: 10.1021/acs.nanolett.4c02654] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/09/2024]
Abstract
Defect engineering is widely used to impart the desired functionalities on materials. Despite the widespread application of atomic-resolution scanning transmission electron microscopy (STEM), traditional methods for defect analysis are highly sensitive to random noise and human bias. While deep learning (DL) presents a viable alternative, it requires extensive amounts of training data with labeled ground truth. Herein, employing cycle generative adversarial networks (CycleGAN) and U-Nets, we propose a method based on a single experimental STEM image to tackle high annotation costs and image noise for defect detection. Not only atomic defects but also oxygen dopants in monolayer MoS2 are visualized. The method can be readily extended to other two-dimensional systems, as the training is based on unit-cell-level images. Therefore, our results outline novel ways to train the model with minimal data sets, offering great opportunities to fully exploit the power of DL in the materials science community.
Collapse
Affiliation(s)
- Kangshu Li
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Xiaocang Han
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Yuan Meng
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Junxian Li
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | | | - Xiang Chen
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Jing-Yang You
- Department of Physics, National University of Singapore, Singapore 117551
| | - Lin Yao
- DP Technology, Beijing 100080, China
| | - Wenchao Hu
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Zhiyi Xia
- DP Technology, Beijing 100080, China
| | - Guolin Ke
- DP Technology, Beijing 100080, China
| | - Linfeng Zhang
- DP Technology, Beijing 100080, China
- AI for Science Institute, Beijing 100084, China
| | - Jin Zhang
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Xiaoxu Zhao
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
- AI for Science Institute, Beijing 100084, China
| |
Collapse
|
29
|
Stripelis D, Gupta U, Saleem H, Dhinagar N, Ghai T, Anastasiou C, Sánchez R, Steeg GV, Ravi S, Naveed M, Thompson PM, Ambite JL. A federated learning architecture for secure and private neuroimaging analysis. PATTERNS (NEW YORK, N.Y.) 2024; 5:101031. [PMID: 39233693 PMCID: PMC11368680 DOI: 10.1016/j.patter.2024.101031] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/10/2023] [Revised: 04/04/2024] [Accepted: 06/06/2024] [Indexed: 09/06/2024]
Abstract
The amount of biomedical data continues to grow rapidly. However, collecting data from multiple sites for joint analysis remains challenging due to security, privacy, and regulatory concerns. To overcome this challenge, we use federated learning, which enables distributed training of neural network models over multiple data sources without sharing data. Each site trains the neural network over its private data for some time and then shares the neural network parameters (i.e., weights and/or gradients) with a federation controller, which in turn aggregates the local models and sends the resulting community model back to each site, and the process repeats. Our federated learning architecture, MetisFL, provides strong security and privacy. First, sample data never leave a site. Second, neural network parameters are encrypted before transmission and the global neural model is computed under fully homomorphic encryption. Finally, we use information-theoretic methods to limit information leakage from the neural model to prevent a "curious" site from performing model inversion or membership attacks. We present a thorough evaluation of the performance of secure, private federated learning in neuroimaging tasks, including for predicting Alzheimer's disease and for brain age gap estimation (BrainAGE) from magnetic resonance imaging (MRI) studies in challenging, heterogeneous federated environments where sites have different amounts of data and statistical distributions.
Collapse
Affiliation(s)
- Dimitris Stripelis
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Umang Gupta
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Hamza Saleem
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Nikhil Dhinagar
- University of Southern California, Imaging Genetics Center, Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, Marina del Rey, CA 90292, USA
| | - Tanmay Ghai
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | | | - Rafael Sánchez
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Greg Ver Steeg
- University of California, Riverside, Riverside, CA 92521, USA
| | - Srivatsan Ravi
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Muhammad Naveed
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Paul M. Thompson
- University of Southern California, Imaging Genetics Center, Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, Marina del Rey, CA 90292, USA
| | - José Luis Ambite
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| |
Collapse
|
30
|
Zhao Y, Dohi O, Ishida T, Yoshida N, Ochiai T, Mukai H, Seya M, Yamauchi K, Miyazaki H, Fukui H, Yasuda T, Iwai N, Inoue K, Itoh Y, Liu X, Zhang R, Zhu X. Linked Color Imaging with Artificial Intelligence Improves the Detection of Early Gastric Cancer. Dig Dis 2024:1-9. [PMID: 39102801 DOI: 10.1159/000540728] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/22/2024] [Accepted: 07/31/2024] [Indexed: 08/07/2024]
Abstract
INTRODUCTION Esophagogastroduodenoscopy is the most important tool to detect gastric cancer (GC). In this study, we developed a computer-aided detection (CADe) system to detect GC with white light imaging (WLI) and linked color imaging (LCI) modes and aimed to compare the performance of CADe with that of endoscopists. METHODS The system was developed based on the deep learning framework from 9,021 images in 385 patients between 2017 and 2020. A total of 116 LCI and WLI videos from 110 patients between 2017 and 2023 were used to evaluate per-case sensitivity and per-frame specificity. RESULTS The per-case sensitivity and per-frame specificity of CADe with a confidence level of 0.5 in detecting GC were 78.6% and 93.4% for WLI and 94.0% and 93.3% for LCI, respectively (p < 0.001). The per-case sensitivities of nonexpert endoscopists for WLI and LCI were 45.8% and 80.4%, whereas those of expert endoscopists were 66.7% and 90.6%, respectively. Regarding detectability between CADe and endoscopists, the per-case sensitivities for WLI and LCI were 78.6% and 94.0% in CADe, respectively, which were significantly higher than those for LCI in experts (90.6%, p = 0.004) and those for WLI and LCI in nonexperts (45.8% and 80.4%, respectively, p < 0.001); however, no significant difference for WLI was observed between CADe and experts (p = 0.134). CONCLUSIONS Our CADe system showed significantly better sensitivity in detecting GC when used in LCI compared with WLI mode. Moreover, the sensitivity of CADe using LCI is significantly higher than those of expert endoscopists using LCI to detect GC.
Collapse
Affiliation(s)
- Youshen Zhao
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| | - Osamu Dohi
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Tsugitaka Ishida
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Naohisa Yoshida
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Tomoko Ochiai
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Hiroki Mukai
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Mayuko Seya
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Katsuma Yamauchi
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Hajime Miyazaki
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Hayato Fukui
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Takeshi Yasuda
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Naoto Iwai
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Ken Inoue
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Yoshito Itoh
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Xinkai Liu
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| | - Ruiyao Zhang
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| | - Xin Zhu
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| |
Collapse
|
31
|
Dunenova G, Kalmataeva Z, Kaidarova D, Dauletbaev N, Semenova Y, Mansurova M, Grjibovski A, Kassymbekova F, Sarsembayev A, Semenov D, Glushkova N. The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review. Cancers (Basel) 2024; 16:2761. [PMID: 39123488 PMCID: PMC11311684 DOI: 10.3390/cancers16152761] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Revised: 07/28/2024] [Accepted: 07/30/2024] [Indexed: 08/12/2024] Open
Abstract
This systematic review aims to address the research gap in the performance of computational algorithms for the digital image analysis of HER2 images in clinical settings. While numerous studies have explored various aspects of these algorithms, there is a lack of comprehensive evaluation regarding their effectiveness in real-world clinical applications. We conducted a search of the Web of Science and PubMed databases for studies published from 31 December 2013 to 30 June 2024, focusing on performance effectiveness and components such as dataset size, diversity and source, ground truth, annotation, and validation methods. The study was registered with PROSPERO (CRD42024525404). Key questions guiding this review include the following: How effective are current computational algorithms at detecting HER2 status in digital images? What are the common validation methods and dataset characteristics used in these studies? Is there standardization of algorithm evaluations of clinical applications that can improve the clinical utility and reliability of computational tools for HER2 detection in digital image analysis? We identified 6833 publications, with 25 meeting the inclusion criteria. The accuracy rate with clinical datasets varied from 84.19% to 97.9%. The highest accuracy was achieved on the publicly available Warwick dataset at 98.8% in synthesized datasets. Only 12% of studies used separate datasets for external validation; 64% of studies used a combination of accuracy, precision, recall, and F1 as a set of performance measures. Despite the high accuracy rates reported in these studies, there is a notable absence of direct evidence supporting their clinical application. To facilitate the integration of these technologies into clinical practice, there is an urgent need to address real-world challenges and overreliance on internal validation. Standardizing study designs on real clinical datasets can enhance the reliability and clinical applicability of computational algorithms in improving the detection of HER2 cancer.
Collapse
Affiliation(s)
- Gauhar Dunenova
- Department of Epidemiology, Biostatistics and Evidence-Based Medicine, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
| | - Zhanna Kalmataeva
- Rector Office, Asfendiyarov Kazakh National Medical University, Almaty 050000, Kazakhstan;
| | - Dilyara Kaidarova
- Kazakh Research Institute of Oncology and Radiology, Almaty 050022, Kazakhstan;
| | - Nurlan Dauletbaev
- Department of Internal, Respiratory and Critical Care Medicine, Philipps University of Marburg, 35037 Marburg, Germany;
- Department of Pediatrics, Faculty of Medicine and Health Sciences, McGill University, Montreal, QC H4A 3J1, Canada
- Faculty of Medicine and Health Care, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
| | - Yuliya Semenova
- School of Medicine, Nazarbayev University, Astana 010000, Kazakhstan;
| | - Madina Mansurova
- Department of Artificial Intelligence and Big Data, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan;
| | - Andrej Grjibovski
- Central Scientific Research Laboratory, Northern State Medical University, Arkhangelsk 163000, Russia;
- Department of Epidemiology and Modern Vaccination Technologies, I.M. Sechenov First Moscow State Medical University, Moscow 105064, Russia
- Department of Biology, Ecology and Biotechnology, Northern (Arctic) Federal University, Arkhangelsk 163000, Russia
- Department of Health Policy and Management, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
| | - Fatima Kassymbekova
- Department of Public Health and Social Sciences, Kazakhstan Medical University “KSPH”, Almaty 050060, Kazakhstan;
| | - Aidos Sarsembayev
- School of Digital Technologies, Almaty Management University, Almaty 050060, Kazakhstan;
- Health Research Institute, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan;
| | - Daniil Semenov
- Computer Science and Engineering Program, Astana IT University, Astana 020000, Kazakhstan;
| | - Natalya Glushkova
- Department of Epidemiology, Biostatistics and Evidence-Based Medicine, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
- Health Research Institute, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan;
| |
Collapse
|
32
|
Badkul A, Vamsi I, Sudha R. Comparative study of DCNN and image processing based classification of chest X-rays for identification of COVID-19 patients using fine-tuning. J Med Eng Technol 2024; 48:213-222. [PMID: 39648993 DOI: 10.1080/03091902.2024.2438158] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 11/20/2024] [Accepted: 11/30/2024] [Indexed: 12/10/2024]
Abstract
The conventional detection of COVID-19 by evaluating the CT scan images is tiresome, often experiences high inter-observer variability and uncertainty issues. This work proposes the automatic detection and classification of COVID-19 by analysing the chest X-ray images (CXR) with the deep convolutional neural network (DCNN) models through a fine-tuning and pre-training approach. CXR images pertaining to four health scenarios, namely, healthy, COVID-19, bacterial pneumonia and viral pneumonia, are considered and subjected to data augmentation. Two types of input datasets are prepared; in which dataset I contains the original image dataset categorised under four classes, whereas the original CXR images are subjected to image pre-processing via Contrast Limited Adaptive Histogram Equalisation (CLAHE) algorithm and Blackhat Morphological Operation (BMO) for devising the input dataset II. Both datasets are supplied as input to various DCNN models such as DenseNet, MobileNet, ResNet, VGG16, and Xception for achieving multi-class classification. It is observed that the classification accuracies are improved, and the classification errors are reduced with the image pre-processing. Overall, the VGG16 model resulted in better classification accuracies and reduced classification errors while accomplishing multi-class classification. Thus, the proposed work would assist the clinical diagnosis, and reduce the workload of the front-line healthcare workforce and medical professionals.
Collapse
Affiliation(s)
- Amitesh Badkul
- Department of Electrical and Electronics, Birla Institute of Technology and Science-Pilani, Hyderabad, India
| | - Inturi Vamsi
- Mechanical Engineering Department, Chaitanya Bharathi Institute of Technology (A), Hyderabad, India
| | - Radhika Sudha
- Department of Electrical and Electronics, Birla Institute of Technology and Science-Pilani, Hyderabad, India
| |
Collapse
|
33
|
Rozhyna A, Somfai GM, Atzori M, DeBuc DC, Saad A, Zoellin J, Müller H. Exploring Publicly Accessible Optical Coherence Tomography Datasets: A Comprehensive Overview. Diagnostics (Basel) 2024; 14:1668. [PMID: 39125544 PMCID: PMC11312046 DOI: 10.3390/diagnostics14151668] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Revised: 07/15/2024] [Accepted: 07/25/2024] [Indexed: 08/12/2024] Open
Abstract
Artificial intelligence has transformed medical diagnostic capabilities, particularly through medical image analysis. AI algorithms perform well in detecting abnormalities with a strong performance, enabling computer-aided diagnosis by analyzing the extensive amounts of patient data. The data serve as a foundation upon which algorithms learn and make predictions. Thus, the importance of data cannot be underestimated, and clinically corresponding datasets are required. Many researchers face a lack of medical data due to limited access, privacy concerns, or the absence of available annotations. One of the most widely used diagnostic tools in ophthalmology is Optical Coherence Tomography (OCT). Addressing the data availability issue is crucial for enhancing AI applications in the field of OCT diagnostics. This review aims to provide a comprehensive analysis of all publicly accessible retinal OCT datasets. Our main objective is to compile a list of OCT datasets and their properties, which can serve as an accessible reference, facilitating data curation for medical image analysis tasks. For this review, we searched through the Zenodo repository, Mendeley Data repository, MEDLINE database, and Google Dataset search engine. We systematically evaluated all the identified datasets and found 23 open-access datasets containing OCT images, which significantly vary in terms of size, scope, and ground-truth labels. Our findings indicate the need for improvement in data-sharing practices and standardized documentation. Enhancing the availability and quality of OCT datasets will support the development of AI algorithms and ultimately improve diagnostic capabilities in ophthalmology. By providing a comprehensive list of accessible OCT datasets, this review aims to facilitate better utilization and development of AI in medical image analysis.
Collapse
Affiliation(s)
- Anastasiia Rozhyna
- Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO), 3960 Sierre, Switzerland
- Medical Informatics, University of Geneva, 1205 Geneva, Switzerland
| | - Gábor Márk Somfai
- Department of Ophthalmology, Stadtspital Zürich, 8063 Zurich, Switzerland
- Spross Research Institute, 8063 Zurich, Switzerland
| | - Manfredo Atzori
- Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO), 3960 Sierre, Switzerland
- Department of Neuroscience, University of Padua, 35121 Padova, Italy
| | - Delia Cabrera DeBuc
- Bascom Palmer Eye Institute, Miller School of Medicine, University of Miami, Miami, FL 33136, USA
| | - Amr Saad
- Department of Ophthalmology, Stadtspital Zürich, 8063 Zurich, Switzerland
- Spross Research Institute, 8063 Zurich, Switzerland
| | - Jay Zoellin
- Department of Ophthalmology, Stadtspital Zürich, 8063 Zurich, Switzerland
- Spross Research Institute, 8063 Zurich, Switzerland
| | - Henning Müller
- Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO), 3960 Sierre, Switzerland
- Medical Informatics, University of Geneva, 1205 Geneva, Switzerland
- The Sense Research and Innovation Center, 1007 Lausanne, Switzerland
| |
Collapse
|
34
|
Sureshkumar V, Prasad RSN, Balasubramaniam S, Jagannathan D, Daniel J, Dhanasekaran S. Breast Cancer Detection and Analytics Using Hybrid CNN and Extreme Learning Machine. J Pers Med 2024; 14:792. [PMID: 39201984 PMCID: PMC11355507 DOI: 10.3390/jpm14080792] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2024] [Revised: 07/08/2024] [Accepted: 07/15/2024] [Indexed: 09/03/2024] Open
Abstract
Early detection of breast cancer is essential for increasing survival rates, as it is one of the primary causes of death for women globally. Mammograms are extensively used by physicians for diagnosis, but selecting appropriate algorithms for image enhancement, segmentation, feature extraction, and classification remains a significant research challenge. This paper presents a computer-aided diagnosis (CAD)-based hybrid model combining convolutional neural networks (CNN) with a pruned ensembled extreme learning machine (HCPELM) to enhance breast cancer detection, segmentation, feature extraction, and classification. The model employs the rectified linear unit (ReLU) activation function to enhance data analytics after removing artifacts and pectoral muscles, and the HCPELM hybridized with the CNN model improves feature extraction. The hybrid elements are convolutional and fully connected layers. Convolutional layers extract spatial features like edges, textures, and more complex features in deeper layers. The fully connected layers take these features and combine them in a non-linear manner to perform the final classification. ELM performs classification and recognition tasks, aiming for state-of-the-art performance. This hybrid classifier is used for transfer learning by freezing certain layers and modifying the architecture to reduce parameters, easing cancer detection. The HCPELM classifier was trained using the MIAS database and evaluated against benchmark methods. It achieved a breast image recognition accuracy of 86%, outperforming benchmark deep learning models. HCPELM is demonstrating superior performance in early detection and diagnosis, thus aiding healthcare practitioners in breast cancer diagnosis.
Collapse
Affiliation(s)
- Vidhushavarshini Sureshkumar
- Department of Computer Science and Engineering, SRM Institute of Science and Technology, Vadapalani, Chennai 600026, India
| | | | | | - Dhayanithi Jagannathan
- Department of Computer Science and Engineering, Sona College of Technology, Salem 636005, India; (S.B.); (D.J.)
| | - Jayanthi Daniel
- Department of Electronics and Communication Engineering, Rajalakshmi Engineering College, Chennai 602105, India;
| | | |
Collapse
|
35
|
Ahmad J, Akram S, Jaffar A, Ali Z, Bhatti SM, Ahmad A, Rehman SU. Deep learning empowered breast cancer diagnosis: Advancements in detection and classification. PLoS One 2024; 19:e0304757. [PMID: 38990817 PMCID: PMC11239011 DOI: 10.1371/journal.pone.0304757] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Accepted: 05/18/2024] [Indexed: 07/13/2024] Open
Abstract
Recent advancements in AI, driven by big data technologies, have reshaped various industries, with a strong focus on data-driven approaches. This has resulted in remarkable progress in fields like computer vision, e-commerce, cybersecurity, and healthcare, primarily fueled by the integration of machine learning and deep learning models. Notably, the intersection of oncology and computer science has given rise to Computer-Aided Diagnosis (CAD) systems, offering vital tools to aid medical professionals in tumor detection, classification, recurrence tracking, and prognosis prediction. Breast cancer, a significant global health concern, is particularly prevalent in Asia due to diverse factors like lifestyle, genetics, environmental exposures, and healthcare accessibility. Early detection through mammography screening is critical, but the accuracy of mammograms can vary due to factors like breast composition and tumor characteristics, leading to potential misdiagnoses. To address this, an innovative CAD system leveraging deep learning and computer vision techniques was introduced. This system enhances breast cancer diagnosis by independently identifying and categorizing breast lesions, segmenting mass lesions, and classifying them based on pathology. Thorough validation using the Curated Breast Imaging Subset of Digital Database for Screening Mammography (CBIS-DDSM) demonstrated the CAD system's exceptional performance, with a 99% success rate in detecting and classifying breast masses. While the accuracy of detection is 98.5%, when segmenting breast masses into separate groups for examination, the method's performance was approximately 95.39%. Upon completing all the analysis, the system's classification phase yielded an overall accuracy of 99.16% for classification. The potential for this integrated framework to outperform current deep learning techniques is proposed, despite potential challenges related to the high number of trainable parameters. Ultimately, this recommended framework offers valuable support to researchers and physicians in breast cancer diagnosis by harnessing cutting-edge AI and image processing technologies, extending recent advances in deep learning to the medical domain.
Collapse
Affiliation(s)
- Jawad Ahmad
- Faculty of Computer Science & Information Technology, The Superior University, Lahore, Pakistan
- Intelligent Data Visual Computing Research (IDVCR), Lahore, Pakistan
| | - Sheeraz Akram
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - Arfan Jaffar
- Faculty of Computer Science & Information Technology, The Superior University, Lahore, Pakistan
- Intelligent Data Visual Computing Research (IDVCR), Lahore, Pakistan
| | - Zulfiqar Ali
- School of Computer Science and Electronic Engineering (CSEE), University of Essex, Wivenhoe Park, Colchester, United Kingdom
| | - Sohail Masood Bhatti
- Faculty of Computer Science & Information Technology, The Superior University, Lahore, Pakistan
- Intelligent Data Visual Computing Research (IDVCR), Lahore, Pakistan
| | - Awais Ahmad
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - Shafiq Ur Rehman
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| |
Collapse
|
36
|
Li C, Zhang F, Du Y, Li H. Classification of brain tumor types through MRIs using parallel CNNs and firefly optimization. Sci Rep 2024; 14:15057. [PMID: 38956224 PMCID: PMC11219740 DOI: 10.1038/s41598-024-65714-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2024] [Accepted: 06/24/2024] [Indexed: 07/04/2024] Open
Abstract
Image segmentation is a critical and challenging endeavor in the field of medicine. A magnetic resonance imaging (MRI) scan is a helpful method for locating any abnormal brain tissue these days. It is a difficult undertaking for radiologists to diagnose and classify the tumor from several pictures. This work develops an intelligent method for accurately identifying brain tumors. This research investigates the identification of brain tumor types from MRI data using convolutional neural networks and optimization strategies. Two novel approaches are presented: the first is a novel segmentation technique based on firefly optimization (FFO) that assesses segmentation quality based on many parameters, and the other is a combination of two types of convolutional neural networks to categorize tumor traits and identify the kind of tumor. These upgrades are intended to raise the general efficacy of the MRI scan technique and increase identification accuracy. Using MRI scans from BBRATS2018, the testing is carried out, and the suggested approach has shown improved performance with an average accuracy of 98.6%.
Collapse
Affiliation(s)
- Chen Li
- Department of Neurosurgery, Shandong Provincial Third Hospital, Shandong University, No.12 Wuyingshan Middle Road, Jinan, 250031, Shandong, China
| | - Faxue Zhang
- Department of Neurosurgery, Shandong Provincial Third Hospital, Shandong University, No.12 Wuyingshan Middle Road, Jinan, 250031, Shandong, China
| | - Yongjian Du
- Department of Neurosurgery, The Fifth People's Hospital of Jinan, No.24297, Jingshi Road, Jinan, 250022, Shandong, China
| | - Huachao Li
- Department of Neurosurgery, Shandong Provincial Third Hospital, Shandong University, No.12 Wuyingshan Middle Road, Jinan, 250031, Shandong, China.
| |
Collapse
|
37
|
Jiang C, Ji T, Qiao Q. Application and progress of artificial intelligence in radiation therapy dose prediction. Clin Transl Radiat Oncol 2024; 47:100792. [PMID: 38779524 PMCID: PMC11109740 DOI: 10.1016/j.ctro.2024.100792] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2024] [Accepted: 05/07/2024] [Indexed: 05/25/2024] Open
Abstract
Radiation therapy (RT) nowadays is a main treatment modality of cancer. To ensure the therapeutic efficacy of patients, accurate dose distribution is often required, which is a time-consuming and labor-intensive process. In addition, due to the differences in knowledge and experience among participants and diverse institutions, the predicted dose are often inconsistent. In last several decades, artificial intelligence (AI) has been applied in various aspects of RT, several products have been implemented in clinical practice and confirmed superiority. In this paper, we will review the research of AI in dose prediction, focusing on the progress in deep learning (DL).
Collapse
Affiliation(s)
- Chen Jiang
- Department of Radiation Oncology, The First Hospital of China Medical University, Shenyang, China
| | - Tianlong Ji
- Department of Radiation Oncology, The First Hospital of China Medical University, Shenyang, China
| | - Qiao Qiao
- Department of Radiation Oncology, The First Hospital of China Medical University, Shenyang, China
| |
Collapse
|
38
|
Zhu K, Shen Z, Wang M, Jiang L, Zhang Y, Yang T, Zhang H, Zhang M. Visual Knowledge Domain of Artificial Intelligence in Computed Tomography: A Review Based on Bibliometric Analysis. J Comput Assist Tomogr 2024; 48:652-662. [PMID: 38271538 DOI: 10.1097/rct.0000000000001585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2024]
Abstract
ABSTRACT Artificial intelligence (AI)-assisted medical imaging technology is a new research area of great interest that has developed rapidly over the last decade. However, there has been no bibliometric analysis of published studies in this field. The present review focuses on AI-related studies on computed tomography imaging in the Web of Science database and uses CiteSpace and VOSviewer to generate a knowledge map and conduct the basic information analysis, co-word analysis, and co-citation analysis. A total of 7265 documents were included and the number of documents published had an overall upward trend. Scholars from the United States and China have made outstanding achievements, and there is a general lack of extensive cooperation in this field. In recent years, the research areas of great interest and difficulty have been the optimization and upgrading of algorithms, and the application of theoretical models to practical clinical applications. This review will help researchers understand the developments, research areas of great interest, and research frontiers in this field and provide reference and guidance for future studies.
Collapse
|
39
|
Yilmaz S, Tasyurek M, Amuk M, Celik M, Canger EM. Developing deep learning methods for classification of teeth in dental panoramic radiography. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 138:118-127. [PMID: 37316425 DOI: 10.1016/j.oooo.2023.02.021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2022] [Revised: 09/13/2022] [Accepted: 02/10/2023] [Indexed: 06/16/2023]
Abstract
OBJECTIVES We aimed to develop an artificial intelligence-based clinical dental decision-support system using deep-learning methods to reduce diagnostic interpretation error and time and increase the effectiveness of dental treatment and classification. STUDY DESIGN We compared the performance of 2 deep-learning methods, You Only Look Once V4 (YOLO-V4) and Faster Regions with the Convolutional Neural Networks (R-CNN), for tooth classification in dental panoramic radiography for tooth classification in dental panoramic radiography to determine which is more successful in terms of accuracy, time, and detection ability. Using a method based on deep-learning models trained on a semantic segmentation task, we analyzed 1200 panoramic radiographs selected retrospectively. In the classification process, our model identified 36 classes, including 32 teeth and 4 impacted teeth. RESULTS The YOLO-V4 method achieved a mean 99.90% precision, 99.18% recall, and 99.54% F1 score. The Faster R-CNN method achieved a mean 93.67% precision, 90.79% recall, and 92.21% F1 score. Experimental evaluations showed that the YOLO-V4 method outperformed the Faster R-CNN method in terms of accuracy of predicted teeth in the tooth classification process, speed of tooth classification, and ability to detect impacted and erupted third molars. CONCLUSIONS The YOLO-V4 method outperforms the Faster R-CNN method in terms of accuracy of tooth prediction, speed of detection, and ability to detect impacted third molars and erupted third molars. The proposed deep learning based methods can assist dentists in clinical decision making, save time, and reduce the negative effects of stress and fatigue in daily practice.
Collapse
Affiliation(s)
- Serkan Yilmaz
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Erciyes University, Kayseri, Turkey
| | - Murat Tasyurek
- Department of Computer Engineering, Kayseri University, Kayseri, Turkey
| | - Mehmet Amuk
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Erciyes University, Kayseri, Turkey
| | - Mete Celik
- Department of Computer Engineering, Erciyes University, Kayseri, Turkey
| | - Emin Murat Canger
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Erciyes University, Kayseri, Turkey.
| |
Collapse
|
40
|
Altındağ A, Bahrilli S, Çelik Ö, Bayrakdar İŞ, Orhan K. Tooth numbering and classification on bitewing radiographs: an artificial intelligence pilot study. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 137:679-689. [PMID: 38632035 DOI: 10.1016/j.oooo.2024.02.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Revised: 01/13/2024] [Accepted: 02/08/2024] [Indexed: 04/19/2024]
Abstract
OBJECTIVE The aim of this study is to assess the efficacy of employing a deep learning methodology for the automated identification and enumeration of permanent teeth in bitewing radiographs. The experimental procedures and techniques employed in this study are described in the following section. STUDY DESIGN A total of 1248 bitewing radiography images were annotated using the CranioCatch labeling program, developed in Eskişehir, Turkey. The dataset has been partitioned into 3 subsets: training (n = 1000, 80% of the total), validation (n = 124, 10% of the total), and test (n = 124, 10% of the total) sets. The images were subjected to a 3 × 3 clash operation in order to enhance the clarity of the labeled regions. RESULTS The F1, sensitivity and precision results of the artificial intelligence model obtained using the Yolov5 architecture in the test dataset were found to be 0.9913, 0.9954, and 0.9873, respectively. CONCLUSION The utilization of numerical identification for teeth within deep learning-based artificial intelligence algorithms applied to bitewing radiographs has demonstrated notable efficacy. The utilization of clinical decision support system software, which is augmented by artificial intelligence, has the potential to enhance the efficiency and effectiveness of dental practitioners.
Collapse
Affiliation(s)
- Ali Altındağ
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Necmettin Erbakan University, Konya, Turkey.
| | - Serkan Bahrilli
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Necmettin Erbakan University, Konya, Turkey
| | - Özer Çelik
- Department of Mathematics-Computer, Eskisehir Osmangazi University Faculty of Science, Eskisehir, Turkey
| | - İbrahim Şevki Bayrakdar
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Eskişehir Osmangazi University, Eskişehir, Turkey
| | - Kaan Orhan
- Department of Oral and Maxillofacial Radiology, Faculty of Dentistry, Ankara University, Ankara, Turkey
| |
Collapse
|
41
|
Botnari A, Kadar M, Patrascu JM. A Comprehensive Evaluation of Deep Learning Models on Knee MRIs for the Diagnosis and Classification of Meniscal Tears: A Systematic Review and Meta-Analysis. Diagnostics (Basel) 2024; 14:1090. [PMID: 38893617 PMCID: PMC11172202 DOI: 10.3390/diagnostics14111090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2024] [Revised: 05/19/2024] [Accepted: 05/21/2024] [Indexed: 06/21/2024] Open
Abstract
OBJECTIVES This study delves into the cutting-edge field of deep learning techniques, particularly deep convolutional neural networks (DCNNs), which have demonstrated unprecedented potential in assisting radiologists and orthopedic surgeons in precisely identifying meniscal tears. This research aims to evaluate the effectiveness of deep learning models in recognizing, localizing, describing, and categorizing meniscal tears in magnetic resonance images (MRIs). MATERIALS AND METHODS This systematic review was rigorously conducted, strictly following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. Extensive searches were conducted on MEDLINE (PubMed), Web of Science, Cochrane Library, and Google Scholar. All identified articles underwent a comprehensive risk of bias analysis. Predictive performance values were either extracted or calculated for quantitative analysis, including sensitivity and specificity. The meta-analysis was performed for all prediction models that identified the presence and location of meniscus tears. RESULTS This study's findings underscore that a range of deep learning models exhibit robust performance in detecting and classifying meniscal tears, in one case surpassing the expertise of musculoskeletal radiologists. Most studies in this review concentrated on identifying tears in the medial or lateral meniscus and even precisely locating tears-whether in the anterior or posterior horn-with exceptional accuracy, as demonstrated by AUC values ranging from 0.83 to 0.94. CONCLUSIONS Based on these findings, deep learning models have showcased significant potential in analyzing knee MR images by learning intricate details within images. They offer precise outcomes across diverse tasks, including segmenting specific anatomical structures and identifying pathological regions. Contributions: This study focused exclusively on DL models for identifying and localizing meniscus tears. It presents a meta-analysis that includes eight studies for detecting the presence of a torn meniscus and a meta-analysis of three studies with low heterogeneity that localize and classify the menisci. Another novelty is the analysis of arthroscopic surgery as ground truth. The quality of the studies was assessed against the CLAIM checklist, and the risk of bias was determined using the QUADAS-2 tool.
Collapse
Affiliation(s)
- Alexei Botnari
- Department of Orthopedics, Faculty of Medicine, “Victor Babes” University of Medicine and Pharmacy, 300041 Timisoara, Romania
| | - Manuella Kadar
- Department of Computer Science, Faculty of Informatics and Engineering, “1 Decembrie 1918” University of Alba Iulia, 510009 Alba Iulia, Romania
| | - Jenel Marian Patrascu
- Department of Orthopedics-Traumatology, Faculty of Medicine, “Victor Babes” University of Medicine and Pharmacy, 300041 Timisoara, Romania;
| |
Collapse
|
42
|
Jiao Z, Liang Z, Liao Q, Chen S, Yang H, Hong G, Gui H. Deep learning for automatic detection of cephalometric landmarks on lateral cephalometric radiographs using the Mask Region-based Convolutional Neural Network: a pilot study. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 137:554-562. [PMID: 38480069 DOI: 10.1016/j.oooo.2024.02.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2023] [Revised: 01/24/2024] [Accepted: 02/04/2024] [Indexed: 06/20/2024]
Abstract
OBJECTIVE We examined the effectiveness and feasibility of the Mask Region-based Convolutional Neural Network (Mask R-CNN) for automatic detection of cephalometric landmarks on lateral cephalometric radiographs (LCRs). STUDY DESIGN In total, 400 LCRs, each with 19 manually identified landmarks, were collected. Of this total, 320 images were randomly selected as the training dataset for Mask R-CNN, and the remaining 80 images were used for testing the automatic detection of the 19 cephalometric landmarks, for a total of 1520 landmarks. Detection rate, average error, and detection accuracy rate were calculated to assess Mask R-CNN performance. RESULTS Of the 1520 landmarks, 1494 were detected, for a detection rate of 98.29%. The average error, or linear deviation distance between the detected points and the originally marked points of each detected landmark, ranged from 0.56 to 9.51 mm, with an average of 2.19 mm. For detection accuracy rate, 649 landmarks (43.44%) had a linear deviation distance less than 1 mm, 1020 (68.27%) less than 2 mm, and 1281 (85.74%) less than 4 mm in deviation from the manually marked point. The average detection time was 1.48 seconds per image. CONCLUSIONS Deep learning Mask R-CNN shows promise in enhancing cephalometric analysis by automating landmark detection on LCRs, addressing the limitations of manual analysis, and demonstrating effectiveness and feasibility.
Collapse
Affiliation(s)
- Zhentao Jiao
- Department of Oral and Maxillofacial Surgery, Dalian Stomatological Hospital, Dalian, China; Division for Globalization Initiative, Liaison Center for Innovative Dentistry, Graduate School of Dentistry, Tohoku University, Sendai, Japan
| | - Zhuangzhuang Liang
- School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Qian Liao
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China
| | - Sheng Chen
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China
| | - Hui Yang
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China
| | - Guang Hong
- Division for Globalization Initiative, Liaison Center for Innovative Dentistry, Graduate School of Dentistry, Tohoku University, Sendai, Japan.
| | - Haijun Gui
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China.
| |
Collapse
|
43
|
Yang X, Huang X, Wei C, Yu J, Yu X, Dong C, Chen J, Chen R, Wu X, Yu Z, Sun B, Wang J, Liu H, Han W, Sun B, Jiang Z, Ding J, Liu Z, Peng J, Ni D, Deng X, Liu L, Gou Z. An intelligent quantification system for fetal heart rhythm assessment: A multicenter prospective study. Heart Rhythm 2024; 21:600-609. [PMID: 38266752 DOI: 10.1016/j.hrthm.2024.01.024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Revised: 01/04/2024] [Accepted: 01/15/2024] [Indexed: 01/26/2024]
Abstract
BACKGROUND The motion relationship and time intervals of the pulsed-wave Doppler (PWD) spectrum are essential for diagnosing fetal arrhythmia. However, few technologies currently are available to automatically calculate fetal cardiac time intervals (CTIs). OBJECTIVE The purpose of this study was to develop a fetal heart rhythm intelligent quantification system (HR-IQS) for the automatic extraction of CTIs and establish the normal reference range for fetal CTIs. METHODS A total of 6498 PWD spectrums of 2630 fetuses over the junction between the left ventricular inflow and outflow tracts were recorded across 14 centers. E, A, and V waves were manually labeled by 3 experienced fetal cardiologists, with 17 CTIs extracted. Five-fold cross-validation was performed for training and testing of the deep learning model. Agreement between the manual and HR-IQS-based values was evaluated using the intraclass correlation coefficient and Spearman's rank correlation coefficient. The Jarque-Bera test was applied to evaluate the normality of CTIs' distributions, and the normal reference range of 17 CTIs was established with quantile regression. Arrhythmia subset was compared with the non-arrhythmia subset using the Mann-Whitney U test. RESULTS Significant positive correlation (P <.001) and moderate-to-excellent consistency (P <.001) between the manual and HR-IQS automated measurements of CTIs was found. The distribution of CTIs was non-normal (P <.001). The normal range (2.5th to 97.5th percentiles) was successfully established for the 17 CTIs. CONCLUSIONS Using our HR-IQS is feasible for the automated calculation of CTIs in practice and thus could provide a promising tool for the assessment of fetal rhythm and function.
Collapse
Affiliation(s)
- Xin Yang
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China
| | - Xiaoqiong Huang
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China
| | - Chenchen Wei
- Center for Cardiovascular Disease, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou, Jiangsu, China
| | - Junxuan Yu
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China; Shenzhen RayShape Medical Technology Co., Ltd, Shenzhen, Guangdong, China
| | - Xuejuan Yu
- Department of Ultrasonography, Suzhou Xiangcheng People's Hospital, Suzhou, Jiangsu, China
| | - Caixia Dong
- Department of Ultrasonography, Wulin Hospital, Hangzhou, Zhejiang, China
| | - Ju Chen
- Department of Ultrasonography, Taicang First People's Hospital, Suzhou, Jiangsu, China
| | - Ruifeng Chen
- Department of Ultrasound Diagnosis, The Affiliated Changsha Central Hospital, Hengyang Medical School, University of South China, Changsha, Hunan, China
| | - Xiafang Wu
- Department of Ultrasonography, Taizhou Hospital of Zhejiang Province Affiliated to Wenzhou Medical University, Linhai, Zhejiang, China
| | - Zhuan Yu
- Department of Ultrasonography, The Affiliated Hospital of Jiangsu University, Zhenjiang, Jiangsu, China
| | - Baojuan Sun
- Department of Ultrasonography, Huai'an Maternal and Child Health Hospital, Huai'an, Jiangsu, China
| | - Junli Wang
- Department of Ultrasonography, Wuhu No.2 People's Hospital, Wuhu, Anhui, China
| | - Hongmei Liu
- Department of Ultrasonography, Panzhou Emerging Hospital, Panzhou, Guizhou, China
| | - Wen Han
- Department of Ultrasonography, Suzhou Gaoxin District People's Hospital, Suzhou, Jiangsu, China
| | - Biyun Sun
- Department of Ultrasonography, The Affiliated Yijishan Hospital of Wannan Medical University, Wuhu, Anhui, China
| | - Zhiyong Jiang
- Department of Ultrasonography, The Huaren Hospital, Wuhu, Zhejiang, China
| | - Jie Ding
- Department of Ultrasonography, The Affiliated Suzhou Hospital of Nanjing University, Suzhou, Jiangsu, China
| | - Zhe Liu
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China; Shenzhen RayShape Medical Technology Co., Ltd, Shenzhen, Guangdong, China
| | - Jin Peng
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China; Shenzhen RayShape Medical Technology Co., Ltd, Shenzhen, Guangdong, China
| | - Dong Ni
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China
| | - Xuedong Deng
- Center for Medical Ultrasound, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou, Jiangsu, China
| | - Lian Liu
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, China; Shenzhen RayShape Medical Technology Co., Ltd, Shenzhen, Guangdong, China.
| | - Zhongshan Gou
- Center for Cardiovascular Disease, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou, Jiangsu, China.
| |
Collapse
|
44
|
Thakur GK, Thakur A, Kulkarni S, Khan N, Khan S. Deep Learning Approaches for Medical Image Analysis and Diagnosis. Cureus 2024; 16:e59507. [PMID: 38826977 PMCID: PMC11144045 DOI: 10.7759/cureus.59507] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Accepted: 05/01/2024] [Indexed: 06/04/2024] Open
Abstract
In addition to enhancing diagnostic accuracy, deep learning techniques offer the potential to streamline workflows, reduce interpretation time, and ultimately improve patient outcomes. The scalability and adaptability of deep learning algorithms enable their deployment across diverse clinical settings, ranging from radiology departments to point-of-care facilities. Furthermore, ongoing research efforts focus on addressing the challenges of data heterogeneity, model interpretability, and regulatory compliance, paving the way for seamless integration of deep learning solutions into routine clinical practice. As the field continues to evolve, collaborations between clinicians, data scientists, and industry stakeholders will be paramount in harnessing the full potential of deep learning for advancing medical image analysis and diagnosis. Furthermore, the integration of deep learning algorithms with other technologies, including natural language processing and computer vision, may foster multimodal medical data analysis and clinical decision support systems to improve patient care. The future of deep learning in medical image analysis and diagnosis is promising. With each success and advancement, this technology is getting closer to being leveraged for medical purposes. Beyond medical image analysis, patient care pathways like multimodal imaging, imaging genomics, and intelligent operating rooms or intensive care units can benefit from deep learning models.
Collapse
Affiliation(s)
- Gopal Kumar Thakur
- Department of Data Sciences, Harrisburg University of Science and Technology, Harrisburg, USA
| | - Abhishek Thakur
- Department of Data Sciences, Harrisburg University of Science and Technology, Harrisburg, USA
| | - Shridhar Kulkarni
- Department of Data Sciences, Harrisburg University of Science and Technology, Harrisburg, USA
| | - Naseebia Khan
- Department of Data Sciences, Harrisburg University of Science and Technology, Harrisburg, USA
| | - Shahnawaz Khan
- Department of Computer Application, Bundelkhand University, Jhansi, IND
| |
Collapse
|
45
|
Chakraborty S, Pradhan B. Editorial for the Special Issue "Machine Learning in Computer Vision and Image Sensing: Theory and Applications". SENSORS (BASEL, SWITZERLAND) 2024; 24:2874. [PMID: 38732978 PMCID: PMC11086158 DOI: 10.3390/s24092874] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Accepted: 03/25/2024] [Indexed: 05/13/2024]
Abstract
Machine learning (ML) models have experienced remarkable growth in their application for multimodal data analysis over the past decade [...].
Collapse
Affiliation(s)
- Subrata Chakraborty
- School of Science and Technology, University of New England, Armidale, NSW 2351, Australia
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), School of Civil and Environmental Engineering, Faculty of Engineering & IT, University of Technology Sydney, Sydney, NSW 2007, Australia;
- Griffith Business School, Griffith University, Nathan, QLD 4111, Australia
| | - Biswajeet Pradhan
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), School of Civil and Environmental Engineering, Faculty of Engineering & IT, University of Technology Sydney, Sydney, NSW 2007, Australia;
- Earth Observation Centre, Institute of Climate Change, Universiti Kebangsaan Malaysia (UKM), Bangi 43600, Selangor, Malaysia
| |
Collapse
|
46
|
Theodoropoulos D, Karabetsos DA, Vakis A, Papadaki E, Karantanas A, Marias K. The current status of noninvasive intracranial pressure monitoring: A literature review. Clin Neurol Neurosurg 2024; 239:108209. [PMID: 38430649 DOI: 10.1016/j.clineuro.2024.108209] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 02/25/2024] [Accepted: 02/26/2024] [Indexed: 03/05/2024]
Abstract
Elevated intracranial pressure (ICP) is a life-threatening condition that must be promptly diagnosed. However, the gold standard methods for ICP monitoring are invasive, time-consuming, and they involve certain risks. To address these risks, many noninvasive approaches have been proposed. This study undertakes a literature review of the existing noninvasive methods, which have reported promising results. The experimental base on which they are established, however, prevents their application in emergency conditions and thus none of them are capable of replacing the traditional invasive methods to date. On the other hand, contemporary methods leverage Machine Learning (ML) which has already shown unprecedented results in several medical research areas. That said, only a few publications exist on ML-based approaches for ICP estimation, which are not appropriate for emergency conditions due to their restricted capability of employing the medical imaging data available in intensive care units. The lack of such image-based ML models to estimate ICP is attributed to the scarcity of annotated datasets requiring directly measured ICP data. This ascertainment highlights an active and unexplored scientific frontier, calling for further research and development in the field of ICP estimation, particularly leveraging the untapped potential of ML techniques.
Collapse
Affiliation(s)
| | - Dimitrios A Karabetsos
- Department of Neurosurgery, Heraklion University Hospital, Voutes, Heraklion, Crete 715 00, Greece.
| | - Antonios Vakis
- University of Crete, Medical School, Andrea Kalokerinou 13, Heraklion, Crete 715 00, Greece; Department of Neurosurgery, Heraklion University Hospital, Voutes, Heraklion, Crete 715 00, Greece
| | - Efrosini Papadaki
- University of Crete, Medical School, Andrea Kalokerinou 13, Heraklion, Crete 715 00, Greece; Department Of Radiology, Heraklion University Hospital, Voutes, Heraklion, Crete 715 00, Greece; FORTH-ICS, Computational Biomedicine Laboratory, Vassilika Vouton, Heraklion
| | - Apostolos Karantanas
- University of Crete, Medical School, Andrea Kalokerinou 13, Heraklion, Crete 715 00, Greece; Department Of Radiology, Heraklion University Hospital, Voutes, Heraklion, Crete 715 00, Greece; FORTH-ICS, Computational Biomedicine Laboratory, Vassilika Vouton, Heraklion
| | - Kostas Marias
- FORTH-ICS, Computational Biomedicine Laboratory, Vassilika Vouton, Heraklion; Department of Electrical and Computer Engineering, Hellenic Mediterranean University, Estavromenos, Heraklion, Crete 71410, Greece
| |
Collapse
|
47
|
Russo C, Bria A, Marrocco C. GravityNet for end-to-end small lesion detection. Artif Intell Med 2024; 150:102842. [PMID: 38553147 DOI: 10.1016/j.artmed.2024.102842] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Revised: 03/01/2024] [Accepted: 03/11/2024] [Indexed: 04/02/2024]
Abstract
This paper introduces a novel one-stage end-to-end detector specifically designed to detect small lesions in medical images. Precise localization of small lesions presents challenges due to their appearance and the diverse contextual backgrounds in which they are found. To address this, our approach introduces a new type of pixel-based anchor that dynamically moves towards the targeted lesion for detection. We refer to this new architecture as GravityNet, and the novel anchors as gravity points since they appear to be "attracted" by the lesions. We conducted experiments on two well-established medical problems involving small lesions to evaluate the performance of the proposed approach: microcalcifications detection in digital mammograms and microaneurysms detection in digital fundus images. Our method demonstrates promising results in effectively detecting small lesions in these medical imaging tasks.
Collapse
Affiliation(s)
- Ciro Russo
- Department of Electrical and Information Engineering, University of Cassino and L.M., Via G. Di Biasio 43, 03043 Cassino (FR), Italy.
| | - Alessandro Bria
- Department of Electrical and Information Engineering, University of Cassino and L.M., Via G. Di Biasio 43, 03043 Cassino (FR), Italy.
| | - Claudio Marrocco
- Department of Electrical and Information Engineering, University of Cassino and L.M., Via G. Di Biasio 43, 03043 Cassino (FR), Italy.
| |
Collapse
|
48
|
Zheng J, Xiong Y, Zheng Y, Zhang H, Wu R. Evaluating the Stroke Risk of Patients using Machine Learning: A New Perspective from Sichuan and Chongqing. EVALUATION REVIEW 2024; 48:346-369. [PMID: 37533403 DOI: 10.1177/0193841x231193468] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/04/2023]
Abstract
Stroke is the leading cause of death and disability among people in China, and it leads to heavy burdens for patients, their families and society. An accurate prediction of the risk of stroke has important implications for early intervention and treatment. In light of recent advances in machine learning, the application of this technique in stroke prediction has achieved plentiful promising results. To detect the relationship between potential factors and the risk of stroke and examine which machine learning method significantly can enhance the prediction accuracy of stroke. We employed six machine learning methods including logistic regression, naive Bayes, decision tree, random forest, K-nearest neighbor and support vector machine, to model and predict the risk of stroke. Participants were 233 patients from Sichuan and Chongqing. Four indicators (accuracy, precision, recall and F1 metric) were examined to evaluate the predictive performance of the different models. The empirical results indicate that random forest yields the best accuracy, recall and F1 in predicting the risk of stroke, with an accuracy of .7548, precision of .7805, recall of .7619 and F1 of .7711. Additionally, the findings show that age, cerebral infarction, PM 8 (an anti-atrial fibrillation drug), and drinking are independent risk factors for stroke. Further studies should adopt a broader assortment of machine learning methods to analyze the risk of stroke, by which better accuracy can be expected. In particular, RF can successfully enhance the forecasting accuracy for stroke.
Collapse
Affiliation(s)
- Jin Zheng
- Institute of Traditional Chinese Medicine, Sichuan Academy of Chinese Medicine Sciences, Chengdu, China
| | - Yao Xiong
- Department of Neurology, The Third People's Hospital of Chengdu & The Affilliate Hosipital of Southwest Jiaotong University, Chengdu, China
| | - Yimei Zheng
- School of Mathematics, Southwest Jiao Tong University, Chengdu, China
| | - Haitao Zhang
- Department of Neurology, The Third People's Hospital of Chengdu & The Affilliate Hosipital of Southwest Jiaotong University, Chengdu, China
| | - Rui Wu
- School of Mathematics, Southwest Jiao Tong University, Chengdu, China
| |
Collapse
|
49
|
Özbay Y, Kazangirler BY, Özcan C, Pekince A. Detection of the separated endodontic instrument on periapical radiographs using a deep learning-based convolutional neural network algorithm. AUST ENDOD J 2024; 50:131-139. [PMID: 38062627 DOI: 10.1111/aej.12822] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2023] [Revised: 10/27/2023] [Accepted: 11/24/2023] [Indexed: 04/07/2024]
Abstract
The study evaluated the diagnostic performance of an artificial intelligence system to detect separated endodontic instruments on periapical radiograph radiographs. Three hundred seven periapical radiographs were collected and divided into 222 for training and 85 for testing to be fed to the Mask R-CNN model. Periapical radiographs were assigned to the training and test set and labelled on the DentiAssist labeling platform. Labelled polygonal objects had their bounding boxes automatically generated by the DentiAssist system. Fractured instruments were classified and segmented. As a result of the proposed method, the mean average precision (mAP) metric was 98.809%, the precision value was 95.238, while the recall reached 98.765 and the f1 score 96.969%. The threshold value of 80% was chosen for the bounding boxes working with the Intersection over Union (IoU) technique. The Mask R-CNN distinguished separated endodontic instruments on periapical radiographs.
Collapse
Affiliation(s)
- Yağız Özbay
- Department of Endodontics, Faculty of Dentistry, Karabuk University, Karabuk, Turkey
| | | | - Caner Özcan
- Department of Software Engineering, Karabuk University, Karabuk, Turkey
| | - Adem Pekince
- Department of Oral and Maxillofacial Radiology, Faculty of Dentistry, Karabuk University, Karabuk, Turkey
| |
Collapse
|
50
|
Kowadlo G, Mittelberg Y, Ghomlaghi M, Stiglitz DK, Kishore K, Guha R, Nazareth J, Weinberg L. Development and validation of 'Patient Optimizer' (POP) algorithms for predicting surgical risk with machine learning. BMC Med Inform Decis Mak 2024; 24:70. [PMID: 38468330 DOI: 10.1186/s12911-024-02463-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2023] [Accepted: 02/20/2024] [Indexed: 03/13/2024] Open
Abstract
BACKGROUND Pre-operative risk assessment can help clinicians prepare patients for surgery, reducing the risk of perioperative complications, length of hospital stay, readmission and mortality. Further, it can facilitate collaborative decision-making and operational planning. OBJECTIVE To develop effective pre-operative risk assessment algorithms (referred to as Patient Optimizer or POP) using Machine Learning (ML) that predict the development of post-operative complications and provide pilot data to inform the design of a larger prospective study. METHODS After institutional ethics approval, we developed a base model that encapsulates the standard manual approach of combining patient-risk and procedure-risk. In an automated process, additional variables were included and tested with 10-fold cross-validation, and the best performing features were selected. The models were evaluated and confidence intervals calculated using bootstrapping. Clinical expertise was used to restrict the cardinality of categorical variables (e.g. pathology results) by including the most clinically relevant values. The models were created with logistic regression (LR) and extreme gradient-boosted trees using XGBoost (Chen and Guestrin, 2016). We evaluated performance using the area under the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AUPRC). Data was obtained from a metropolitan university teaching hospital from January 2015 to July 2020. Data collection was restricted to adult patients undergoing elective surgery. RESULTS A total of 11,475 adult admissions were included. The performance of XGBoost and LR was very similar across endpoints and metrics. For predicting the risk of any post-operative complication, kidney failure and length-of-stay (LOS), POP with XGBoost achieved an AUROC (95%CI) of 0.755 (0.744, 0.767), 0.869 (0.846, 0.891) and 0.841 (0.833, 0.847) respectively and AUPRC of 0.651 (0.632, 0.669), 0.336 (0.282, 0.390) and 0.741 (0.729, 0.753) respectively. For 30-day readmission and in-patient mortality, POP with XGBoost achieved an AUROC (95%CI) of 0.610 (0.587, 0.635) and 0.866 (0.777, 0.943) respectively and AUPRC of 0.116 (0.104, 0.132) and 0.031 (0.015, 0.072) respectively. CONCLUSION The POP algorithms effectively predicted any post-operative complication, kidney failure and LOS in the sample population. A larger study is justified to improve the algorithm to better predict complications and length of hospital stay. A larger dataset may also improve the prediction of additional specific complications, readmission and mortality.
Collapse
Affiliation(s)
| | | | | | - Daniel K Stiglitz
- Atidia Health, Melbourne, Australia
- Department of Anaesthesiology and Perioperative Medicine, Alfred Health, Melbourne, Australia
| | - Kartik Kishore
- Data Analytics Research and Evaluation Centre, Austin Health, Melbourne, Australia
| | - Ranjan Guha
- Department of Anaesthesia, Austin Health, Heidelberg, Australia
| | - Justin Nazareth
- Department of Anaesthesia, Austin Health, Heidelberg, Australia
| | - Laurence Weinberg
- Department of Anaesthesia, Austin Health, Heidelberg, Australia
- Department of Critical Care, The University of Melbourne, Austin Health, Heidelberg, Australia
| |
Collapse
|