1
|
Baxter JSH, Eagleson R. Exploring the values underlying machine learning research in medical image analysis. Med Image Anal 2025; 102:103494. [PMID: 40020419 DOI: 10.1016/j.media.2025.103494] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 01/26/2025] [Accepted: 02/01/2025] [Indexed: 03/03/2025]
Abstract
Machine learning has emerged as a crucial tool for medical image analysis, largely due to recent developments in deep artificial neural networks addressing numerous, diverse clinical problems. As with any conceptual tool, the effective use of machine learning should be predicated on an understanding of its underlying motivations just as much as algorithms or theory - and to do so, we need to explore its philosophical foundations. One of these foundations is the understanding of how values, despite being non-empirical, nevertheless affect scientific research. This article has three goals: to introduce the reader to values in a way that is specific to medical image analysis; to characterise a particular set of technical decisions (what we call the end-to-end vs. separable learning spectrum) that are fundamental to machine learning for medical image analysis; and to create a simple and structured method to show how these values can be rigorously connected to these technical decisions. This better understanding of how the philosophy of science can clarify fundamental elements of how medical image analysis research is performed and can be improved.
Collapse
Affiliation(s)
- John S H Baxter
- Laboratoire Traitement du Signal et de l'Image (LTSI, INSERM UMR 1099), Université de Rennes, Rennes, France.
| | - Roy Eagleson
- Biomedical Engineering Graduate Program, Western University, London, Canada
| |
Collapse
|
2
|
Sappey-Marinier E, Dutra Vieira T, Schmidt A, Aït Si Selmi T, Bonnin M. How New Technologies Will Transform Total Knee Arthroplasty from a Singular Surgical Procedure to a Holistic Standardized Process. J Clin Med 2025; 14:3102. [PMID: 40364133 DOI: 10.3390/jcm14093102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2025] [Revised: 04/25/2025] [Accepted: 04/28/2025] [Indexed: 05/15/2025] Open
Abstract
Many new technologies focused mainly on improving surgical accuracy were first developed in total knee arthroplasty and have not yet shown significant value. These non-significant clinical improvements could potentially be explained by an inadequate target. In this current concept paper, the authors will describe how artificial intelligence (AI), robotic-assisted surgery, and custom implants allow the definition of new targets and the standardization of the TKA process. As paradoxical as it may be, new technologies in TKA will allow for better standardization in the overall way in which this procedure is carried out. Achieving this goal can be accomplished by incorporating AI-driven tools into the medical field. These tools are intended to enhance decision making, refine surgical planning, and increase the precision and consistency of surgical procedures. Moreover, custom implants with personalized alignment, beyond restoring native anatomy, will define new targets and standardize the whole TKA process.
Collapse
Affiliation(s)
| | - Thais Dutra Vieira
- Centre Orthopédique Santy, Hôpital Privé Jean Mermoz, Ramsay Santé, 69008 Lyon, France
| | - Axel Schmidt
- Centre Orthopédique Santy, Hôpital Privé Jean Mermoz, Ramsay Santé, 69008 Lyon, France
| | - Tarik Aït Si Selmi
- Centre Orthopédique Santy, Hôpital Privé Jean Mermoz, Ramsay Santé, 69008 Lyon, France
| | - Michel Bonnin
- Centre Orthopédique Santy, Hôpital Privé Jean Mermoz, Ramsay Santé, 69008 Lyon, France
| |
Collapse
|
3
|
Hadilou M, Mahdavi N, Keykha E, Ghofrani A, Tahmasebi E, Arabfard M. Artificial intelligence based vision transformer application for grading histopathological images of oral epithelial dysplasia: a step towards AI-driven diagnosis. BMC Cancer 2025; 25:780. [PMID: 40281456 PMCID: PMC12032748 DOI: 10.1186/s12885-025-14193-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2024] [Accepted: 04/21/2025] [Indexed: 04/29/2025] Open
Abstract
BACKGROUND This study aimed to classify dysplastic and healthy oral epithelial histopathological images, according to WHO and binary grading systems, using the Vision Transformer (ViT) deep learning algorithm-a state-of-the-art Artificial Intelligence (AI) approach and compare it with established Convolutional Neural Network models (VGG16 and ConvNet). METHODS A total of 218 histopathological slide images were collected from the Department of Oral and Maxillofacial Pathology at Tehran University of Medical Sciences archive and combined with two online databases. Two oral pathologists independently labeled the images based on the 2022 World Health Organization (WHO) grading system (mild, moderate and severe), the binary grading system (low risk and high risk), including an additional normal tissue class. After preprocessing, the images were fed to the ViT, VGG16 and ConvNet models. RESULTS Image preprocessing yielded 2,545 low-risk, 2,054 high-risk, 726 mild, 831 moderate, 449 severe, and 937 normal tissue patches. The proposed ViT model outperformed both CNNs with the accuracy of 94% (VGG16:86% and ConvNet: 88%) in 3-class scenario and 97% (VGG16:79% and ConvNet: 88%) in 4-class scenario. CONCLUSIONS The ViT model successfully classified oral epithelial dysplastic tissues with a high accuracy, paving the way for AI to serve as an adjunct or independent tool alongside oral and maxillofacial pathologists for detecting and grading oral epithelial dysplasia.
Collapse
Affiliation(s)
- Mahdi Hadilou
- Research Center for Prevention of Oral and Dental Diseases, Baqiyatallah University of Medical Sciences, Tehran, Iran
| | - Nazanin Mahdavi
- Department of Oral and Maxillofacial Pathology, School of Dentistry, Tehran University of Medical Sciences, Tehran, Iran
| | - Elham Keykha
- Research Center for Prevention of Oral and Dental Diseases, School of Dentistry, Baqiyatallah University of Medical Sciences, Tehran, Iran.
| | - Ali Ghofrani
- Student Research Committee, Baqiyatallah University of Medical Sciences, Tehran, Iran
| | - Elahe Tahmasebi
- Research Center for Prevention of Oral and Dental Diseases, School of Dentistry, Baqiyatallah University of Medical Sciences, Tehran, Iran
| | - Masoud Arabfard
- Chemical Injuries Research Center, Systems Biology and Poisonings Institute, Baqiyatallah University of Medical Sciences, Tehran, Iran
| |
Collapse
|
4
|
Raza A, Guzzo A, Ianni M, Lappano R, Zanolini A, Maggiolini M, Fortino G. Federated Learning in radiomics: A comprehensive meta-survey on medical image analysis. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 267:108768. [PMID: 40279838 DOI: 10.1016/j.cmpb.2025.108768] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/06/2024] [Revised: 03/03/2025] [Accepted: 04/09/2025] [Indexed: 04/29/2025]
Abstract
Federated Learning (FL) has emerged as a promising approach for collaborative medical image analysis while preserving data privacy, making it particularly suitable for radiomics tasks. This paper presents a systematic meta-analysis of recent surveys on Federated Learning in Medical Imaging (FL-MI), published in reputable venues over the past five years. We adopt the PRISMA methodology, categorizing and analyzing the existing body of research in FL-MI. Our analysis identifies common trends, challenges, and emerging strategies for implementing FL in medical imaging, including handling data heterogeneity, privacy concerns, and model performance in non-IID settings. The paper also highlights the most widely used datasets and a comparison of adopted machine learning models. Moreover, we examine FL frameworks in FL-MI applications, such as tumor detection, organ segmentation, and disease classification. We identify several research gaps, including the need for more robust privacy protection. Our findings provide a comprehensive overview of the current state of FL-MI and offer valuable directions for future research and development in this rapidly evolving field.
Collapse
Affiliation(s)
- Asaf Raza
- Department of Informatics, Modeling, Electronics, and Systems, University of Calabria, Rende, Italy
| | - Antonella Guzzo
- Department of Informatics, Modeling, Electronics, and Systems, University of Calabria, Rende, Italy
| | - Michele Ianni
- Department of Informatics, Modeling, Electronics, and Systems, University of Calabria, Rende, Italy.
| | - Rosamaria Lappano
- Department of Pharmacy, Health and Nutritional Sciences, University of Calabria, Rende, Italy
| | - Alfredo Zanolini
- Radiology Unit, "Annunziata" Hospital, Azienda Ospedaliera di Cosenza, Cosenza, Italy
| | - Marcello Maggiolini
- Department of Pharmacy, Health and Nutritional Sciences, University of Calabria, Rende, Italy
| | - Giancarlo Fortino
- Department of Informatics, Modeling, Electronics, and Systems, University of Calabria, Rende, Italy
| |
Collapse
|
5
|
Wang J, Cai J, Tang W, Dudurych I, van Tuinen M, Vliegenthart R, van Ooijen P. A comparison of an integrated and image-only deep learning model for predicting the disappearance of indeterminate pulmonary nodules. Comput Med Imaging Graph 2025; 123:102553. [PMID: 40239430 DOI: 10.1016/j.compmedimag.2025.102553] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2024] [Revised: 03/18/2025] [Accepted: 04/03/2025] [Indexed: 04/18/2025]
Abstract
BACKGROUND Indeterminate pulmonary nodules (IPNs) require follow-up CT to assess potential growth; however, benign nodules may disappear. Accurately predicting whether IPNs will resolve is a challenge for radiologists. Therefore, we aim to utilize deep-learning (DL) methods to predict the disappearance of IPNs. MATERIAL AND METHODS This retrospective study utilized data from the Dutch-Belgian Randomized Lung Cancer Screening Trial (NELSON) and Imaging in Lifelines (ImaLife) cohort. Participants underwent follow-up CT to determine the evolution of baseline IPNs. The NELSON data was used for model training. External validation was performed in ImaLife. We developed integrated DL-based models that incorporated CT images and demographic data (age, sex, smoking status, and pack years). We compared the performance of integrated methods with those limited to CT images only and calculated sensitivity, specificity, and area under the receiver operating characteristic curve (AUC). From a clinical perspective, ensuring high specificity is critical, as it minimizes false predictions of non-resolving nodules that should be monitored for evolution on follow-up CTs. Feature importance was calculated using SHapley Additive exPlanations (SHAP) values. RESULTS The training dataset included 840 IPNs (134 resolving) in 672 participants. The external validation dataset included 111 IPNs (46 resolving) in 65 participants. On the external validation set, the performance of the integrated model (sensitivity, 0.50; 95 % CI, 0.35-0.65; specificity, 0.91; 95 % CI, 0.80-0.96; AUC, 0.82; 95 % CI, 0.74-0.90) was comparable to that solely trained on CT image (sensitivity, 0.41; 95 % CI, 0.27-0.57; specificity, 0.89; 95 % CI, 0.78-0.95; AUC, 0.78; 95 % CI, 0.69-0.86; P = 0.39). The top 10 most important features were all image related. CONCLUSION Deep learning-based models can predict the disappearance of IPNs with high specificity. Integrated models using CT scans and clinical data had comparable performance to those using only CT images.
Collapse
Affiliation(s)
- Jingxuan Wang
- Department of Radiology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands; Data Science in Health (DASH), University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands
| | - Jiali Cai
- Department of Epidemiology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands
| | - Wei Tang
- Department of Neurology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands; Data Science in Health (DASH), University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands
| | - Ivan Dudurych
- Department of Radiology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands
| | - Marcel van Tuinen
- Department of Radiology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands
| | - Rozemarijn Vliegenthart
- Department of Radiology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands; Data Science in Health (DASH), University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands
| | - Peter van Ooijen
- Department of Radiation Oncology, University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands; Data Science in Health (DASH), University of Groningen, University Medical Center of Groningen, Groningen, the Netherlands.
| |
Collapse
|
6
|
Ahmed A, Zeng X, Xi R, Hou M, Shah SA. Enhancing multimodal medical image analysis with Slice-Fusion: A novel fusion approach to address modality imbalance. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 261:108615. [PMID: 39904191 DOI: 10.1016/j.cmpb.2025.108615] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/26/2024] [Revised: 01/01/2025] [Accepted: 01/20/2025] [Indexed: 02/06/2025]
Abstract
BACKGROUND AND OBJECTIVE In recent times, medical imaging analysis (MIA) has seen an increasing interest due to its core application in computer-aided diagnosis systems (CADs). A modality in MIA refers to a specific technology used to produce human body images, such as MRI, CT scans, or X-rays. Each modality presents unique challenges and characteristics, often leading to imbalances within datasets. This significant challenge impedes model training and generalization due to the varying convergence rates of different modalities and the suppression of gradients in less dominant modalities. METHODS This paper proposes a novel fusion approach, and we named it Slice-Fusion. The proposed approach aims to mitigate the modality imbalance problem by implementing a "Modality-Specific-Balancing-Factor" fusion strategy. Furthermore, it incorporates an auxiliary (uni-modal) task that generates balanced modality pairs based on the image orientations of different modalities. Subsequently, a novel multimodal classification framework is presented to learn from the generated balanced modalities. The effectiveness of the proposed approach is evaluated through comparative assessments on a publicly available BraTS2021 dataset. The results demonstrate the efficiency of Slice-Fusion in resolving the modality imbalance problem. By enhancing the representation of balanced features and reducing modality bias, this approach holds promise for advancing visual health informatics and facilitating more accurate and reliable medical image analysis. RESULTS In the experiment section, three diverse experiments are conducted such as i) Fusion Loss Metrics Evaluation, ii) Classification, and iii) Visual Health Informatics. Notably, the proposed approach achieved an F1-Score of (100%, 81.25%) on the training and validation sets for the classification generalization task. In addition to the Slice-Fusion's out-performance, the study also created a new modality-aligned dataset (a highly balanced and informative modality-specific image collection) that aids further research and improves MIA's robustness. These advancements not only enhance the capability of medical diagnostic tools but also create opportunities for future innovations in the field. CONCLUSION This study contributes to advancing medical image analysis, such as effective modality fusion, image reconstruction, comparison, and glioma classification, facilitating more accurate and reliable results, and holds promise for further advancements in visual health informatics.
Collapse
Affiliation(s)
- Awais Ahmed
- School of Computer Science, China West Normal University, China; School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China.
| | - Xiaoyang Zeng
- School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China.
| | - Rui Xi
- School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China.
| | - Mengshu Hou
- School of Computer Science and Engineering, University of Electronic Science and Technology of China - UESTC, Sichuan, 611731, China; School of Big Data and Artificial Intelligence, Chengdu Technological University, Sichuan, 611730, China.
| | - Syed Attique Shah
- School of Computing and Digital Technology, Birmingham City University, STEAMhouse, B4 7RQ, Birmingham, United Kingdom.
| |
Collapse
|
7
|
Lim H, Gi Y, Ko Y, Jo Y, Hong J, Kim J, Ahn SH, Park HC, Kim H, Chung K, Yoon M. A device-dependent auto-segmentation method based on combined generalized and single-device datasets. Med Phys 2025; 52:2375-2383. [PMID: 39699056 DOI: 10.1002/mp.17570] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2024] [Revised: 11/12/2024] [Accepted: 11/26/2024] [Indexed: 12/20/2024] Open
Abstract
BACKGROUND Although generalized-dataset-based auto-segmentation models that consider various computed tomography (CT) scanners have shown great clinical potential, their application to medical images from unseen scanners remains challenging because of device-dependent image features. PURPOSE This study aims to investigate the performance of a device-dependent auto-segmentation model based on a combined dataset of a generalized dataset and single CT scanner dataset. METHOD We constructed two training datasets for 21 chest and abdominal organs. The generalized dataset comprised 1203 publicly available multi-scanner data. The device-dependent dataset comprised 1253 data, including the 1203 multi-CT scanner data and 50 single CT scanner data. Using these datasets, the generalized-dataset-based model (GDSM) and the device-dependent-dataset-based model (DDSM) were trained. The models were trained using nnU-Net and tested on ten data samples from a single CT scanner. The evaluation metrics included the Dice similarity coefficient (DSC), the Hausdorff distance (HD), and the average symmetric surface distance (ASSD), which were used to assess the overall performance of the models. In addition, DSCdiff, HDratio, and ASSDratio, which are variations of the three existing metrics, were used to compare the performance of the models across different organs. RESULT For the average DSC, the GDSM and DDSM had values of 0.9251 and 0.9323, respectively. For the average HD, the GDSM and DDSM had values of 10.66 and 9.139 mm, respectively; for the average ASSD, the GDSM and DDSM had values of 0.8318 and 0.6656 mm, respectively. Compared with the GDSM, the DDSM showed consistent performance improvements of 0.78%, 14%, and 20% for the DSC, HD, and ASSD metrics, respectively. In addition, compared with the GDSM, the DDSM had better DSCdiff values in 14 of 21 tested organs, better HDratio values in 13 of 21 tested organs, and better ASSDratio values in 14 of 21 tested organs. The three averages of the variant metrics were all better for the DDSM than for the GDSM. CONCLUSION The results suggest that combining the generalized dataset with a single scanner dataset resulted in an overall improvement in model performance for that device image.
Collapse
Affiliation(s)
- Hyeongjin Lim
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yongha Gi
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yousun Ko
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yunhui Jo
- Institute of Global Health Technology (IGHT), Korea University, Seoul, Republic of Korea
| | - Jinyoung Hong
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | | | - Sung Hwan Ahn
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Hee-Chul Park
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Haeyoung Kim
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Kwangzoo Chung
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Myonggeun Yoon
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
- FieldCure Ltd, Seoul, Republic of Korea
| |
Collapse
|
8
|
Sun Y, Chen Y, Dong L, Hu D, Zhang X, Jin C, Zhou R, Zhang J, Dou X, Wang J, Xue L, Xiao M, Zhong Y, Tian M, Zhang H. Diagnostic performance of deep learning-assisted [ 18F]FDG PET imaging for Alzheimer's disease: a systematic review and meta-analysis. Eur J Nucl Med Mol Imaging 2025:10.1007/s00259-025-07228-9. [PMID: 40159544 DOI: 10.1007/s00259-025-07228-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2025] [Accepted: 03/17/2025] [Indexed: 04/02/2025]
Abstract
PURPOSE This study aims to calculate the diagnostic performance of deep learning (DL)-assisted 18F-fluorodeoxyglucose ([18F]FDG) PET imaging in Alzheimer's disease (AD). METHODS The Ovid MEDLINE, Ovid Embase, Web of Science Core Collection, Cochrane, and IEEE Xplore databases were searched for related studies from inception to May 24, 2024. We included original studies that developed a DL algorithm for [18F]FDG PET imaging to assess diagnostic performance in classifying AD, mild cognitive impairment (MCI), and normal control (NC). A bivariate random-effects model was employed to assess the area under the curve (AUC). RESULTS We identified 36 studies that met the inclusion criteria. Of these, 35 studies distinguished AD from NC, with a pooled AUC of 0.98 (95% CI: 0.96-0.99). Thirteen studies distinguished AD from MCI, with a pooled AUC of 0.95 (95% CI: 0.92-0.96). Nineteen studies distinguished MCI from NC, with a pooled AUC of 0.94 (95% CI: 0.91-0.95). Additionally, we found large amounts of heterogeneity across studies which could be partially attributed to variations in DL methods and imaging modalities. CONCLUSION This systematic review and meta-analysis shows that DL-assisted [18F]FDG PET imaging has high diagnostic performance in identifying AD. The significant heterogeneity among studies underscores the necessity for future research to incorporate external validation, utilize large sample size, and adhere to rigorous guideline to provide robust support for clinical decision-making.
Collapse
Affiliation(s)
- Yuan Sun
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Yuhan Chen
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - La Dong
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Daoyan Hu
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
- College of Biomedical Engineering & Instrument Science, Zhejiang University, Hangzhou, 310014, Zhejiang, China
| | - Xiaohui Zhang
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Chentao Jin
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
- Key Laboratory for Biomedical Engineering of Ministry of Education, Zhejiang University, Hangzhou, 310014, Zhejiang, China
| | - Rui Zhou
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Jucheng Zhang
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
- Department of Clinical Engineering, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
| | - Xiaofeng Dou
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Jing Wang
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Le Xue
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
- Huashan Hospital and Human Phenome Institute, Fudan University, Shanghai, 200040, China
| | - Meiling Xiao
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China
| | - Yan Zhong
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China.
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China.
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China.
- Key Laboratory for Biomedical Engineering of Ministry of Education, Zhejiang University, Hangzhou, 310014, Zhejiang, China.
| | - Mei Tian
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China.
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China.
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China.
- Huashan Hospital and Human Phenome Institute, Fudan University, Shanghai, 200040, China.
| | - Hong Zhang
- Department of Nuclear Medicine and PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, Zhejiang, China.
- Institute of Nuclear Medicine and Molecular Imaging, Zhejiang University, Hangzhou, 310009, Zhejiang, China.
- Key Laboratory of Medical Molecular Imaging of Zhejiang Province, Hangzhou, 310009, Zhejiang, China.
- College of Biomedical Engineering & Instrument Science, Zhejiang University, Hangzhou, 310014, Zhejiang, China.
- Key Laboratory for Biomedical Engineering of Ministry of Education, Zhejiang University, Hangzhou, 310014, Zhejiang, China.
| |
Collapse
|
9
|
Squires S, Weedon MN, Oram RA. Exploring the application of deep learning methods for polygenic risk score estimation. Biomed Phys Eng Express 2025; 11:025056. [PMID: 40020248 DOI: 10.1088/2057-1976/adbb71] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2024] [Accepted: 02/28/2025] [Indexed: 03/05/2025]
Abstract
Background. Polygenic risk scores (PRS) summarise genetic information into a single number with clinical and research uses. Deep learning (DL) has revolutionised multiple fields, however, the impact of DL on PRSs has been less significant. We explore how DL can improve the generation of PRSs.Methods. We train DL models on known PRSs using UK Biobank data. We explore whether the models can recreate human programmed PRSs, including using a single model to generate multiple PRSs, and DL difficulties in PRS generation. We investigate how DL can compensate for missing data and constraints on performance.Results. We demonstrate almost perfect generation of multiple PRSs with little loss of performance with reduced quantity of training data. For an example set of missing SNPs the DL model produces predictions that enable separation of cases from population samples with an area under the receiver operating characteristic curve of 0.847 (95% CI: 0.828-0.864) compared to 0.798 (95% CI: 0.779-0.818) for the PRS.Conclusions. DL can accurately generate PRSs, including with one model for multiple PRSs. The models are transferable and have high longevity. With certain missing SNPs the DL models can improve on PRS generation; further improvements would likely require additional input data.
Collapse
Affiliation(s)
- Steven Squires
- Clinical and Biomedical Sciences, Faculty of Health and Life Sciences, University of Exeter, Exeter, United Kingdom
| | - Michael N Weedon
- Clinical and Biomedical Sciences, Faculty of Health and Life Sciences, University of Exeter, Exeter, United Kingdom
| | - Richard A Oram
- Clinical and Biomedical Sciences, Faculty of Health and Life Sciences, University of Exeter, Exeter, United Kingdom
- The Academic Renal Unit, Royal Devon University Healthcare Foundation Trust, United Kingdom
| |
Collapse
|
10
|
Deng Z, Yang Y, Suzuki K. Federated Active Learning Framework for Efficient Annotation Strategy in Skin-Lesion Classification. J Invest Dermatol 2025; 145:303-311. [PMID: 38909844 DOI: 10.1016/j.jid.2024.05.023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 05/10/2024] [Accepted: 05/27/2024] [Indexed: 06/25/2024]
Abstract
Federated learning (FL) enables multiple institutes to train models collaboratively without sharing private data. Current FL research focuses on communication efficiency, privacy protection, and personalization and assumes that the data of FL have already been ideally collected. However, in medical scenarios, data annotation demands both expertise and intensive labor, which is a critical problem in FL. Active learning (AL) has shown promising performance in reducing the number of data annotations in medical image analysis. We propose a federated AL framework in which AL is executed periodically and interactively under FL. We exploit a local model in each hospital and a global model acquired from FL to construct an ensemble. We use ensemble entropy-based AL as an efficient data-annotation strategy in FL. Therefore, our federated AL framework can decrease the amount of annotated data and preserve patient privacy while maintaining the performance of FL. To our knowledge, this federated AL framework applied to medical images has not been previously reported. We validated our framework on real-world dermoscopic datasets. Using only 50% of samples, our framework was able to achieve state-of-the-art performance on a skin-lesion classification task. Our framework performed better than several state-of-the-art AL methods under FL and achieved comparable performance with full-data FL.
Collapse
Affiliation(s)
- Zhipeng Deng
- Biomedical Artificial Intelligence Research Unit (BMAI), Institute of Innovative Research, Tokyo Institute of Technology, Tokyo, Japan; Department of Information and Communications Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo, Japan
| | - Yuqiao Yang
- Biomedical Artificial Intelligence Research Unit (BMAI), Institute of Innovative Research, Tokyo Institute of Technology, Tokyo, Japan; Department of Information and Communications Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo, Japan
| | - Kenji Suzuki
- Biomedical Artificial Intelligence Research Unit (BMAI), Institute of Innovative Research, Tokyo Institute of Technology, Tokyo, Japan; Department of Information and Communications Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo, Japan.
| |
Collapse
|
11
|
Ramos-Briceño DA, Flammia-D'Aleo A, Fernández-López G, Carrión-Nessi FS, Forero-Peña DA. Deep learning-based malaria parasite detection: convolutional neural networks model for accurate species identification of Plasmodium falciparum and Plasmodium vivax. Sci Rep 2025; 15:3746. [PMID: 39885248 PMCID: PMC11782605 DOI: 10.1038/s41598-025-87979-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Accepted: 01/23/2025] [Indexed: 02/01/2025] Open
Abstract
Accurate malaria diagnosis with precise identification of Plasmodium species is crucial for an effective treatment. While microscopy is still the gold standard in malaria diagnosis, it relies heavily on trained personnel. Artificial intelligence (AI) advances, particularly convolutional neural networks (CNNs), have significantly improved diagnostic capabilities and accuracy by enabling the automated analysis of medical images. Previous models efficiently detected malaria parasites in red blood cells but had difficulty differentiating between species. We propose a CNN-based model for classifying cells infected by P. falciparum, P. vivax, and uninfected white blood cells from thick blood smears. Our best-performing model utilizes a seven-channel input and correctly predicted 12,876 out of 12,954 cases. We also generated a cross-validation confusion matrix that showed the results of five iterations, achieving 63,654 out of 64,126 true predictions. The model's accuracy reached 99.51%, a precision of 99.26%, a recall of 99.26%, a specificity of 99.63%, an F1 score of 99.26%, and a loss of 2.3%. We are now developing a system based on real-world quality images to create a comprehensive detection tool for remote regions where trained microscopists are unavailable.
Collapse
Affiliation(s)
- Diego A Ramos-Briceño
- School of Systems Engineering, Faculty of Engineering, Universidad Metropolitana de Caracas, Caracas, Venezuela
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela
- "Luis Razetti" School of Medicine, Universidad Central de Venezuela, Caracas, Venezuela
| | - Alessandro Flammia-D'Aleo
- School of Systems Engineering, Faculty of Engineering, Universidad Metropolitana de Caracas, Caracas, Venezuela
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela
| | - Gerardo Fernández-López
- Department of Electronics and Circuits, Faculty of Engineering, Universidad Simón Bolívar, Caracas, Venezuela
| | - Fhabián S Carrión-Nessi
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela.
- "Luis Razetti" School of Medicine, Universidad Central de Venezuela, Caracas, Venezuela.
- Immunogenetics Section, Laboratory of Pathophysiology, Centro de Medicina Experimental "Miguel Layrisse", Instituto Venezolano de Investigaciones Científicas, Altos de Pipe, Venezuela.
| | - David A Forero-Peña
- Biomedical Research and Therapeutic Vaccines Institute, Ciudad Bolívar, Venezuela.
- "Luis Razetti" School of Medicine, Universidad Central de Venezuela, Caracas, Venezuela.
- Department of Infectious Diseases, Hospital Universitario de Caracas, Caracas, Venezuela.
| |
Collapse
|
12
|
Kim JW, Khan AU, Banerjee I. Systematic Review of Hybrid Vision Transformer Architectures for Radiological Image Analysis. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-024-01322-4. [PMID: 39871042 DOI: 10.1007/s10278-024-01322-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/28/2024] [Revised: 10/11/2024] [Accepted: 10/25/2024] [Indexed: 01/29/2025]
Abstract
Vision transformer (ViT)and convolutional neural networks (CNNs) each possess distinct strengths in medical imaging: ViT excels in capturing long-range dependencies through self-attention, while CNNs are adept at extracting local features via spatial convolution filters. While ViT may struggle with capturing detailed local spatial information, critical for tasks like anomaly detection in medical imaging, shallow CNNs often fail to effectively abstract global context. This study aims to explore and evaluate hybrid architectures that integrate ViT and CNN to leverage their complementary strengths for enhanced performance in medical vision tasks, such as segmentation, classification, reconstruction, and prediction. Following PRISMA guideline, a systematic review was conducted on 34 articles published between 2020 and Sept. 2024. These articles proposed novel hybrid ViT-CNN architectures specifically for medical imaging tasks in radiology. The review focused on analyzing architectural variations, merging strategies between ViT and CNN, innovative applications of ViT, and efficiency metrics including parameters, inference time (GFlops), and performance benchmarks. The review identified that integrating ViT and CNN can mitigate the limitations of each architecture offering comprehensive solutions that combine global context understanding with precise local feature extraction. We benchmarked the articles based on architectural variations, merging strategies, innovative uses of ViT, and efficiency metrics (number of parameters, inference time (GFlops), and performance), and derived a ranked list. By synthesizing current literature, this review defines fundamental concepts of hybrid vision transformers and highlights emerging trends in the field. It provides a clear direction for future research aimed at optimizing the integration of ViT and CNN for effective utilization in medical imaging, contributing to advancements in diagnostic accuracy and image analysis. We performed systematic review of hybrid vision transformer architecture using PRISMA guideline and performed thorough comparative analysis to benchmark the architectures.
Collapse
Affiliation(s)
- Ji Woong Kim
- School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ, USA
| | | | - Imon Banerjee
- School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ, USA.
- Department of Radiology, Mayo Clinic, Phoenix, AZ, USA.
- Department of Artificial Intelligence and Informatics (AI&I), Mayo Clinic, Scottsdale, AZ, USA.
| |
Collapse
|
13
|
Dashti M, Londono J, Ghasemi S, Tabatabaei S, Hashemi S, Baghaei K, Palma PJ, Khurshid Z. Evaluation of accuracy of deep learning and conventional neural network algorithms in detection of dental implant type using intraoral radiographic images: A systematic review and meta-analysis. J Prosthet Dent 2025; 133:137-146. [PMID: 38176985 DOI: 10.1016/j.prosdent.2023.11.030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2023] [Revised: 11/23/2023] [Accepted: 11/28/2023] [Indexed: 01/06/2024]
Abstract
STATEMENT OF PROBLEM With the growing importance of implant brand detection in clinical practice, the accuracy of machine learning algorithms in implant brand detection has become a subject of research interest. Recent studies have shown promising results for the use of machine learning in implant brand detection. However, despite these promising findings, a comprehensive evaluation of the accuracy of machine learning in implant brand detection is needed. PURPOSE The purpose of this systematic review and meta-analysis was to assess the accuracy, sensitivity, and specificity of deep learning algorithms in implant brand detection using 2-dimensional images such as from periapical or panoramic radiographs. MATERIAL AND METHODS Electronic searches were conducted in PubMed, Embase, Scopus, Scopus Secondary, and Web of Science databases. Studies that met the inclusion criteria were assessed for quality using the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool. Meta-analyses were performed using the random-effects model to estimate the pooled performance measures and the 95% confidence intervals (CIs) using STATA v.17. RESULTS Thirteen studies were selected for the systematic review, and 3 were used in the meta-analysis. The meta-analysis of the studies found that the overall accuracy of CNN algorithms in detecting dental implants in radiographic images was 95.63%, with a sensitivity of 94.55% and a specificity of 97.91%. The highest reported accuracy was 99.08% for CNN Multitask ResNet152 algorithm, and sensitivity and specificity were 100.00% and 98.70% respectively for the deep CNN (Neuro-T version 2.0.1) algorithm with the Straumann SLActive BLT implant brand. All studies had a low risk of bias. CONCLUSIONS The highest accuracy and sensitivity were reported in studies using CNN Multitask ResNet152 and deep CNN (Neuro-T version 2.0.1) algorithms.
Collapse
Affiliation(s)
- Mahmood Dashti
- Researcher, School of Dentistry, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| | - Jimmy Londono
- Professor and Director of the Prosthodontics Residency Program and the Ronald Goldstein Center for Esthetics and Implant Dentistry, The Dental College of Georgia at Augusta University, Augusta, GA
| | - Shohreh Ghasemi
- Graduate Student, MSc of Trauma and Craniofacial Reconstrution, Faculty of Medicine and Dentistry, Queen Mary College, London, England
| | | | - Sara Hashemi
- Graduate student, Faculty of Dental Medicine and Oral Health Sciences, McGill University, Montreal, Quebec, Canada
| | - Kimia Baghaei
- Researcher, Dental Students' Research Committee, School of Dentistry, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Paulo J Palma
- Researcher, Center for Innovation and Research in Oral Sciences (CIROS), Faculty of Medicine, University of Coimbra, Coimbra, Portugal; and Professor, Institute of Endodontics, Faculty of Medicine, University of Coimbra, Coimbra, Portugal.
| | - Zohaib Khurshid
- Lecturer, Prosthodontics, Department of Prosthodontics and Dental Implantology, King Faisal University, Al-Ahsa, Saudi Arabia; and Professor, Center of Excellence for Regenerative Dentistry, Department of Anatomy, Faculty of Dentistry, Chulalongkorn University, Bangkok, Thailand
| |
Collapse
|
14
|
Obeidat R, Alsmadi I, Baker QB, Al-Njadat A, Srinivasan S. Researching public health datasets in the era of deep learning: a systematic literature review. Health Informatics J 2025; 31:14604582241307839. [PMID: 39794941 DOI: 10.1177/14604582241307839] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/13/2025]
Abstract
Objective: Explore deep learning applications in predictive analytics for public health data, identify challenges and trends, and then understand the current landscape. Materials and Methods: A systematic literature review was conducted in June 2023 to search articles on public health data in the context of deep learning, published from the inception of medical and computer science databases through June 2023. The review focused on diverse datasets, abstracting applications, challenges, and advancements in deep learning. Results: 2004 articles were reviewed, identifying 14 disease categories. Observed trends include explainable-AI, patient embedding learning, and integrating different data sources and employing deep learning models in health informatics. Noted challenges were technical reproducibility and handling sensitive data. Discussion: There has been a notable surge in deep learning applications on public health data publications since 2015. Consistent deep learning applications and models continue to be applied across public health data. Despite the wide applications, a standard approach still does not exist for addressing the outstanding challenges and issues in this field. Conclusion: Guidelines are needed for applying deep learning and models in public health data to improve FAIRness, efficiency, transparency, comparability, and interoperability of research. Interdisciplinary collaboration among data scientists, public health experts, and policymakers is needed to harness the full potential of deep learning.
Collapse
Affiliation(s)
- Rand Obeidat
- Department of Management Information Systems, Bowie State University, Bowie, USA
| | - Izzat Alsmadi
- Department of Computational, Engineering and Mathematical Sciences, Texas A & M San Antonio, San Antonio, USA
| | - Qanita Bani Baker
- Department of Computer Science, Jordan University of Science and Technology, Irbid, Jordan
| | | | - Sriram Srinivasan
- Department of Management Information Systems, Bowie State University, Bowie, USA
| |
Collapse
|
15
|
Dong B, Xue H, Li Y, Li P, Chen J, Zhang T, Chen L, Pan D, Liu P, Sun P. Classification and diagnosis of cervical lesions based on colposcopy images using deep fully convolutional networks: A man-machine comparison cohort study. FUNDAMENTAL RESEARCH 2025; 5:419-428. [PMID: 40166111 PMCID: PMC11955021 DOI: 10.1016/j.fmre.2022.09.032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2022] [Revised: 07/26/2022] [Accepted: 09/19/2022] [Indexed: 11/11/2022] Open
Abstract
Colposcopy is an important technique in the diagnosis of cervical cancer. The development of computer-aided diagnosis methods can mitigate the shortage of colposcopists and improve the accuracy and efficiency of colposcopy examinations in China. This study proposes the Dense-U-Net model for colposcopy image recognition. This was a man-machine comparison cohort study. It presents a novel artificial intelligence (AI) model for the diagnosis of cervical lesions through colposcopy images using a Dense-U-Net image semantic segmentation algorithm. The Dense-U-Net model was created by applying the methods of "deepening the network structure," "applying dropout" and "max pooling." Moreover, image-based and population-based diagnostic performances of the AI algorithm and physicians with different levels of specialist experience were compared. In total, 2,475 participants were recruited, and 13,084 colposcopy images were included in this study. The diagnostic accuracy of the Dense-U-Net model increased significantly with increasing colposcopy images per patient. As the number of images in the training set increased, the diagnostic accuracy of the Dense-U-Net model for cervical intraepithelial neoplasm 3 or worse (CIN3+) diagnosis increased (P = 0.035). The rate of diagnostic accuracy (0.89 vs 0.85, P < 0.001) of CIN3+ lesions using the Dense-U-Net model was higher than that of expert colposcopists, and the missed diagnosis (0.06 vs 0.07, P = 0.002) and false positive (0.05 vs 0.08, P < 0.001) were lower. Moreover, Dense-U-Net is more accurate in diagnosing the type III cervical transformation zone, which is difficult to diagnose by experts (P < 0.001). The Dense-U-Net model also showed higher diagnostic accuracy for CIN3+ in an independent test set (P < 0.001). To diagnose the same 870 test images, the Dense-U-Net system took 1.76 ± 0.09 min, while the expert, senior, and junior colposcopists took 716.3 ± 49.76, 892.1 ± 92.30, and 3034.7 ± 259.51 min, respectively. The study successfully built a reliable, quick, and effective Dense-U-Net model to assist with colposcopy examinations.
Collapse
Affiliation(s)
- Binhua Dong
- Laboratory of Gynecologic Oncology, Department of Gynecology, Fujian Maternity and Child Health Hospital, College of Clinical Medicine for Obstetrics & Gynecology and Pediatrics, Fujian Medical University, Fuzhou 350001, China
- Fujian Key Laboratory of Women and Children's Critical Diseases Research, Fuzhou 350001, China
| | - Huifeng Xue
- Medical Center of Cervical Disease and Colposcopy, Fujian Maternity and Child Health Hospital, Fuzhou 350001, China
| | - Ye Li
- Laboratory of Gynecologic Oncology, Department of Gynecology, Fujian Maternity and Child Health Hospital, College of Clinical Medicine for Obstetrics & Gynecology and Pediatrics, Fujian Medical University, Fuzhou 350001, China
| | - Ping Li
- Department of Gynecology and Obstetrics, Quanzhou First Hospital, Fujian Medical University, Quanzhou 362000, China
| | - Jiancui Chen
- Medical Center of Cervical Disease and Colposcopy, Fujian Maternity and Child Health Hospital, Fuzhou 350001, China
| | - Tao Zhang
- College of Computer Science and Technology, Huaqiao University, Xiamen 361021, China
| | - Lihua Chen
- Laboratory of Gynecologic Oncology, Department of Gynecology, Fujian Maternity and Child Health Hospital, College of Clinical Medicine for Obstetrics & Gynecology and Pediatrics, Fujian Medical University, Fuzhou 350001, China
| | - Diling Pan
- Department of Pathology, Fujian Maternity and Child Health Hospital, Fuzhou 350001, China
| | - Peizhong Liu
- College of Engineering, Huaqiao University, Quanzhou 362021, China
- School of Medicine, Huaqiao University, Quanzhou 362021, China
| | - Pengming Sun
- Laboratory of Gynecologic Oncology, Department of Gynecology, Fujian Maternity and Child Health Hospital, College of Clinical Medicine for Obstetrics & Gynecology and Pediatrics, Fujian Medical University, Fuzhou 350001, China
- Fujian Key Laboratory of Women and Children's Critical Diseases Research, Fuzhou 350001, China
| |
Collapse
|
16
|
Viqar M, Sahin E, Stoykova E, Madjarova V. Reconstruction of Optical Coherence Tomography Images from Wavelength Space Using Deep Learning. SENSORS (BASEL, SWITZERLAND) 2024; 25:93. [PMID: 39796883 PMCID: PMC11723098 DOI: 10.3390/s25010093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Revised: 12/19/2024] [Accepted: 12/24/2024] [Indexed: 01/13/2025]
Abstract
Conventional Fourier domain Optical Coherence Tomography (FD-OCT) systems depend on resampling into a wavenumber (k) domain to extract the depth profile. This either necessitates additional hardware resources or amplifies the existing computational complexity. Moreover, the OCT images also suffer from speckle noise, due to systemic reliance on low-coherence interferometry. We propose a streamlined and computationally efficient approach based on Deep Learning (DL) which enables reconstructing speckle-reduced OCT images directly from the wavelength (λ) domain. For reconstruction, two encoder-decoder styled networks, namely Spatial Domain Convolution Neural Network (SD-CNN) and Fourier Domain CNN (FD-CNN), are used sequentially. The SD-CNN exploits the highly degraded images obtained by Fourier transforming the (λ) domain fringes to reconstruct the deteriorated morphological structures along with suppression of unwanted noise. The FD-CNN leverages this output to enhance the image quality further by optimization in the Fourier domain (FD). We quantitatively and visually demonstrate the efficacy of the method in obtaining high-quality OCT images. Furthermore, we illustrate the computational complexity reduction by harnessing the power of DL models. We believe that this work lays the framework for further innovations in the realm of OCT image reconstruction.
Collapse
Affiliation(s)
- Maryam Viqar
- Faculty of Information Technology and Communication Sciences, Tampere University, 33720 Tampere, Finland;
- Institute of Optical Materials and Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria; (E.S.); (V.M.)
| | - Erdem Sahin
- Faculty of Information Technology and Communication Sciences, Tampere University, 33720 Tampere, Finland;
| | - Elena Stoykova
- Institute of Optical Materials and Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria; (E.S.); (V.M.)
| | - Violeta Madjarova
- Institute of Optical Materials and Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria; (E.S.); (V.M.)
| |
Collapse
|
17
|
Berghout T. The Neural Frontier of Future Medical Imaging: A Review of Deep Learning for Brain Tumor Detection. J Imaging 2024; 11:2. [PMID: 39852315 PMCID: PMC11766058 DOI: 10.3390/jimaging11010002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2024] [Revised: 12/21/2024] [Accepted: 12/23/2024] [Indexed: 01/26/2025] Open
Abstract
Brain tumor detection is crucial in medical research due to high mortality rates and treatment challenges. Early and accurate diagnosis is vital for improving patient outcomes, however, traditional methods, such as manual Magnetic Resonance Imaging (MRI) analysis, are often time-consuming and error-prone. The rise of deep learning has led to advanced models for automated brain tumor feature extraction, segmentation, and classification. Despite these advancements, comprehensive reviews synthesizing recent findings remain scarce. By analyzing over 100 research papers over past half-decade (2019-2024), this review fills that gap, exploring the latest methods and paradigms, summarizing key concepts, challenges, datasets, and offering insights into future directions for brain tumor detection using deep learning. This review also incorporates an analysis of previous reviews and targets three main aspects: feature extraction, segmentation, and classification. The results revealed that research primarily focuses on Convolutional Neural Networks (CNNs) and their variants, with a strong emphasis on transfer learning using pre-trained models. Other methods, such as Generative Adversarial Networks (GANs) and Autoencoders, are used for feature extraction, while Recurrent Neural Networks (RNNs) are employed for time-sequence modeling. Some models integrate with Internet of Things (IoT) frameworks or federated learning for real-time diagnostics and privacy, often paired with optimization algorithms. However, the adoption of eXplainable AI (XAI) remains limited, despite its importance in building trust in medical diagnostics. Finally, this review outlines future opportunities, focusing on image quality, underexplored deep learning techniques, expanding datasets, and exploring deeper learning representations and model behavior such as recurrent expansion to advance medical imaging diagnostics.
Collapse
Affiliation(s)
- Tarek Berghout
- Laboratory of Automation and Manufacturing Engineering, Department of Industrial Engineering, Batna 2 University, Batna 05000, Algeria
| |
Collapse
|
18
|
Lerch L, Huber LS, Kamath A, Pöllinger A, Pahud de Mortanges A, Obmann VC, Dammann F, Senn W, Reyes M. DreamOn: a data augmentation strategy to narrow the robustness gap between expert radiologists and deep learning classifiers. FRONTIERS IN RADIOLOGY 2024; 4:1420545. [PMID: 39758512 PMCID: PMC11696537 DOI: 10.3389/fradi.2024.1420545] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/20/2024] [Accepted: 11/22/2024] [Indexed: 01/07/2025]
Abstract
Purpose Successful performance of deep learning models for medical image analysis is highly dependent on the quality of the images being analysed. Factors like differences in imaging equipment and calibration, as well as patient-specific factors such as movements or biological variability (e.g., tissue density), lead to a large variability in the quality of obtained medical images. Consequently, robustness against the presence of noise is a crucial factor for the application of deep learning models in clinical contexts. Materials and methods We evaluate the effect of various data augmentation strategies on the robustness of a ResNet-18 trained to classify breast ultrasound images and benchmark the performance against trained human radiologists. Additionally, we introduce DreamOn, a novel, biologically inspired data augmentation strategy for medical image analysis. DreamOn is based on a conditional generative adversarial network (GAN) to generate REM-dream-inspired interpolations of training images. Results We find that while available data augmentation approaches substantially improve robustness compared to models trained without any data augmentation, radiologists outperform models on noisy images. Using DreamOn data augmentation, we obtain a substantial improvement in robustness in the high noise regime. Conclusions We show that REM-dream-inspired conditional GAN-based data augmentation is a promising approach to improving deep learning model robustness against noise perturbations in medical imaging. Additionally, we highlight a gap in robustness between deep learning models and human experts, emphasizing the imperative for ongoing developments in AI to match human diagnostic expertise.
Collapse
Affiliation(s)
- Luc Lerch
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
- Computational Neuroscience Group, Department of Physiology, University of Bern, Bern, Switzerland
| | - Lukas S. Huber
- Cognition, Perception and Research Methods, Department of Psychology, University of Bern, Bern, Switzerland
- Neural Information Processing Group, Department of Computer Science, University of Tübingen, Tübingen, Germany
| | - Amith Kamath
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
| | - Alexander Pöllinger
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Aurélie Pahud de Mortanges
- Medical Image Analysis Group, ARTORG Centre for Biomedical Research, University of Bern, Bern, Switzerland
| | - Verena C. Obmann
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Florian Dammann
- Department of Diagnostic, Interventional, and Pediatric Radiology, Inselspital Bern, University of Bern, Bern, Switzerland
| | - Walter Senn
- Computational Neuroscience Group, Department of Physiology, University of Bern, Bern, Switzerland
- Center for Artificial Intelligence in Medicine, University of Bern, Bern, Switzerland
| | - Mauricio Reyes
- Center for Artificial Intelligence in Medicine, University of Bern, Bern, Switzerland
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
- Department of Radiation Oncology, University Hospital Bern, University of Bern, Bern, Switzerland
| |
Collapse
|
19
|
Lei M, Zhao J, Sahan AZ, Hu J, Zhou J, Lee H, Wu Q, Zhang J, Liu Z. Deep Learning Assisted Plasmonic Dark-Field Microscopy for Super-Resolution Label-Free Imaging. NANO LETTERS 2024; 24:15724-15730. [PMID: 39586837 PMCID: PMC11638943 DOI: 10.1021/acs.nanolett.4c04399] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2024] [Revised: 11/21/2024] [Accepted: 11/22/2024] [Indexed: 11/27/2024]
Abstract
Dark-field microscopy (DFM) is a widely used imaging tool, due to its high-contrast capability in imaging label-free specimens. Traditional DFM requires optical alignment to block the oblique illumination, and the resolution is diffraction-limited to the wavelength scale. In this work, we present deep-learning assisted plasmonic dark-field microscopy (DAPD), which is a single-frame super-resolution method using plasmonic dark-field (PDF) microscopy and deep-learning assisted image reconstruction. Specifically, we fabricated a designed PDF substrate with surface plasmon polaritons (SPPs) illuminating specimens on the substrate. Dark field images formed by scattered light from the specimen are further processed by a pretrained convolutional neural network (CNN) using a simulation dataset based on the designed substrate and parameters of the detection optics. We demonstrated a resolution enhancement of 2.8 times on various label-free objects with a large potential for future improvement. We highlight our technique as a compact alternative to traditional DFM with a significantly enhanced spatial resolution.
Collapse
Affiliation(s)
- Ming Lei
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Junxiang Zhao
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Ayse Z. Sahan
- Department
of Pharmacology, University of California−San
Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Biomedical
Sciences Graduate Program, University of
California, San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Jie Hu
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Junxiao Zhou
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Hongki Lee
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Qianyi Wu
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Jin Zhang
- Department
of Pharmacology, University of California−San
Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Department
of Chemistry and Biochemistry, University
of California, San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Shu
Chien − Gene Lay Department of Bioengineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| | - Zhaowei Liu
- Department
of Electrical and Computer Engineering, University of California−San Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
- Materials
Science and Engineering, University of California−San
Diego, 9500 Gilman Drive, La Jolla, California 92093, United States
| |
Collapse
|
20
|
Singh D, Marathe A, Roy S, Walambe R, Kotecha K. Explainable rotation-invariant self-supervised representation learning. MethodsX 2024; 13:102959. [PMID: 39329154 PMCID: PMC11426157 DOI: 10.1016/j.mex.2024.102959] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2024] [Accepted: 09/11/2024] [Indexed: 09/28/2024] Open
Abstract
This paper describes a method that can perform robust detection and classification in out-of-distribution rotated images in the medical domain. In real-world medical imaging tools, noise due to the rotation of the body part is frequently observed. This noise reduces the accuracy of AI-based classification and prediction models. Hence, it is important to develop models which are rotation invariant. To that end, the proposed method - RISC (rotation invariant self-supervised vision framework) addresses this issue of rotational corruption. We present state-of-the-art rotation-invariant classification results and provide explainability for the performance in the domain. The evaluation of the proposed method is carried out on real-world adversarial examples in Medical Imagery-OrganAMNIST, RetinaMNIST and PneumoniaMNIST. It is observed that RISC outperforms the rotation-affected benchmark methods by obtaining 22\%, 17\% and 2\% accuracy boost on OrganAMNIST, PneumoniaMNIST and RetinaMNIST rotated baselines respectively. Further, explainability results are demonstrated. This methods paper describes:•a representation learning approach that can perform robust detection and classification in out-of-distribution rotated images in the medical domain.•It presents a method that incorporates self-supervised rotation invariance for correcting rotational corruptions.•GradCAM-based explainability for the rotational SSL pretext task and the downstream classification outcomes for the three benchmark datasets are presented.
Collapse
Affiliation(s)
- Devansh Singh
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, India
| | - Aboli Marathe
- Machine Learning Department, Carnegie Mellon University, Pittsburgh, PA, USA
| | - Sidharth Roy
- Department of Computer and Information Science, University of Pennsylvania, USA
| | - Rahee Walambe
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, India
| | - Ketan Kotecha
- Symbiosis Centre for Applied Artificial Intelligence, Symbiosis Institute of Technology, India
| |
Collapse
|
21
|
Carvalho BKG, Nolden EL, Wenning AS, Kiss-Dala S, Agócs G, Róth I, Kerémi B, Géczi Z, Hegyi P, Kivovics M. Diagnostic accuracy of artificial intelligence for approximal caries on bitewing radiographs: A systematic review and meta-analysis. J Dent 2024; 151:105388. [PMID: 39396775 DOI: 10.1016/j.jdent.2024.105388] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Revised: 09/13/2024] [Accepted: 10/01/2024] [Indexed: 10/15/2024] Open
Abstract
OBJECTIVES This systematic review and meta-analysis aimed to investigate the diagnostic accuracy of Artificial Intelligence (AI) for approximal carious lesions on bitewing radiographs. METHODS This study included randomized controlled trials (RCTs) and non-randomized controlled trials (non-RCTs) reporting on the diagnostic accuracy of AI for approximal carious lesions on bitewing radiographs. The risk of bias was assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool. A systematic search was conducted on November 4, 2023, in PubMed, Cochrane, and Embase databases and an updated search was performed on August 28, 2024. The primary outcomes assessed were sensitivity, specificity, and overall accuracy. Sensitivity and specificity were pooled using a bivariate model. RESULTS Of the 2,442 studies identified, 21 met the inclusion criteria. The pooled sensitivity and specificity of AI were 0.94 (confidence interval (CI): ± 0.78-0.99) and 0.91 (CI: ± 0.84-0.95), respectively. The positive predictive value (PPV) ranged from 0.15 to 0.87, indicating a moderate capacity for identifying true positives among decayed teeth. The negative predictive value (NPV) ranged from 0.79 to 1.00, demonstrating a high ability to exclude healthy teeth. The diagnostic odds ratio was high, indicating strong overall diagnostic performance. CONCLUSIONS AI models demonstrate clinically acceptable diagnostic accuracy for approximal caries on bitewing radiographs. Although AI can be valuable for preliminary screening, positive findings should be verified by dental experts to prevent unnecessary treatments and ensure timely diagnosis. AI models are highly reliable in excluding healthy approximal surfaces. CLINICAL SIGNIFICANCE AI can assist dentists in detecting approximal caries on bitewing radiographs. However, expert supervision is required to prevent iatrogenic damage and ensure timely diagnosis.
Collapse
Affiliation(s)
| | - Elias-Leon Nolden
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary
| | - Alexander Schulze Wenning
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary
| | - Szilvia Kiss-Dala
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary
| | - Gergely Agócs
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Biophysics and Radiation Biology, Semmelweis University, Tűzoltó utca 37-47, 1072, Budapest, Hungary
| | - Ivett Róth
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Prosthodontics, Semmelweis University, Szentkirályi utca 47 1088, Budapest, Hungary
| | - Beáta Kerémi
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Restorative Dentistry and Endodontics, Semmelweis University, Szentkirályi utca 47, 1088, Budapest, Hungary
| | - Zoltán Géczi
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Prosthodontics, Semmelweis University, Szentkirályi utca 47 1088, Budapest, Hungary
| | - Péter Hegyi
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Institute of Pancreatic Diseases, Semmelweis University, Tömő utca 25-29 1083, Budapest, Hungary; Institute for Translational Medicine, Medical School, University of Pécs, Szigeti utca 12 7624, Pécs, Hungary
| | - Márton Kivovics
- Centre for Translational Medicine, Semmelweis University, Tűzoltó utca 37-47 1072, Budapest, Hungary; Department of Community Dentistry, Semmelweis University, Szentkirályi utca 40 1088, Budapest, Hungary.
| |
Collapse
|
22
|
Song S, Wang J, Wang Z, Wang H, Su J, Ding X, Dang K. DualStreamFoveaNet: A Dual Stream Fusion Architecture With Anatomical Awareness for Robust Fovea Localization. IEEE J Biomed Health Inform 2024; 28:7217-7229. [PMID: 39150813 DOI: 10.1109/jbhi.2024.3445112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/18/2024]
Abstract
Accurate fovea localization is essential for analyzing retinal diseases to prevent irreversible vision loss. While current deep learning-based methods outperform traditional ones, they still face challenges such as the lack of local anatomical landmarks around the fovea, the inability to robustly handle diseased retinal images, and the variations in image conditions. In this paper, we propose a novel transformer-based architecture called DualStreamFoveaNet (DSFN) for multi-cue fusion. This architecture explicitly incorporates long-range connections and global features using retina and vessel distributions for robust fovea localization. We introduce a spatial attention mechanism in the dual-stream encoder to extract and fuse self-learned anatomical information, focusing more on features distributed along blood vessels and significantly reducing computational costs by decreasing token numbers. Our extensive experiments show that the proposed architecture achieves state-of-the-art performance on two public datasets and one large-scale private dataset. Furthermore, we demonstrate that the DSFN is more robust on both normal and diseased retina images and has better generalization capacity in cross-dataset experiments.
Collapse
|
23
|
Raina D, Chandrashekhara SH, Voyles R, Wachs J, Saha SK. Deep-Learning Model for Quality Assessment of Urinary Bladder Ultrasound Images Using Multiscale and Higher-Order Processing. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2024; 71:1451-1463. [PMID: 38598406 DOI: 10.1109/tuffc.2024.3386919] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/12/2024]
Abstract
Autonomous ultrasound image quality assessment (US-IQA) is a promising tool to aid the interpretation by practicing sonographers and to enable the future robotization of ultrasound procedures. However, autonomous US-IQA has several challenges. Ultrasound images contain many spurious artifacts, such as noise due to handheld probe positioning, errors in the selection of probe parameters, and patient respiration during the procedure. Furthermore, these images are highly variable in appearance with respect to the individual patient's physiology. We propose to use a deep convolutional neural network (CNN), USQNet, which utilizes a multiscale and local-to-global second-order pooling (MS-L2GSoP) classifier to conduct the sonographer-like assessment of image quality. This classifier first extracts features at multiple scales to encode the interpatient anatomical variations, similar to a sonographer's understanding of anatomy. Then, it uses SoP in the intermediate layers (local) and at the end of the network (global) to exploit the second-order statistical dependency of MS structural and multiregion textural features. The L2GSoP will capture the higher-order relationships between different spatial locations and provide the seed for correlating local patches, much like a sonographer prioritizes regions across the image. We experimentally validated the USQNet for a new dataset of the human urinary bladder (UB) ultrasound images. The validation involved first with the subjective assessment by experienced radiologists' annotation, and then with state-of-the-art (SOTA) CNN networks for US-IQA and its ablated counterparts. The results demonstrate that USQNet achieves a remarkable accuracy of 92.4% and outperforms the SOTA models by 3%-14% while requiring comparable computation time.
Collapse
|
24
|
Fantini M, Ciravegna G, Koudounas A, Cerquitelli T, Baralis E, Succo G, Crosetti E. The Rapidly Evolving Scenario of Acoustic Voice Analysis in Otolaryngology. Cureus 2024; 16:e73491. [PMID: 39669823 PMCID: PMC11635181 DOI: 10.7759/cureus.73491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/11/2024] [Indexed: 12/14/2024] Open
Abstract
The field of voice analysis has experienced significant transformations, evolving from basic perceptual assessments to the incorporation of advanced digital signal processing and computational tools. This progression has facilitated a deeper understanding of the complex dynamics of vocal function, particularly through the use of acoustic voice analysis within a multidimensional evaluation framework. Traditionally, voice analysis relied on parameters such as fundamental frequency, jitter, shimmer, and noise-to-harmonic ratio, which, despite their utility, have faced criticism for variability and lack of robustness. Recent developments have led to a shift toward more reliable metrics such as cepstral measures, which offer improved accuracy in voice quality assessments. Furthermore, the integration of multiparametric constructs underscores a comprehensive approach to evaluating vocal quality, blending sustained vowels, and continuous speech analyses. Current trends in clinical practice increasingly favor these advanced measures over traditional parameters due to their greater reliability and clinical utility. Additionally, the emergence of artificial intelligence (AI), particularly deep learning, holds promise for revolutionizing voice analysis by enhancing diagnostic precision and enabling efficient, non-invasive screening methods. This shift toward AI-driven approaches signifies a potential paradigm change in voice health, suggesting a future where AI not only aids in diagnosis but also the early detection and treatment of voice-related pathologies.
Collapse
Affiliation(s)
- Marco Fantini
- Ear, Nose, and Throat Unit, Koelliker Hospital, Turin, ITA
- Ear, Nose, and Throat Unit, San Feliciano Hospital, Rome, Italy
| | - Gabriele Ciravegna
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Alkis Koudounas
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Tania Cerquitelli
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Elena Baralis
- DataBase and Data Mining Group (DBDMG) Department of Control and Computer Engineering (DAUIN), Polytechnic University of Turin, Turin, ITA
| | - Giovanni Succo
- Ear, Nose, and Throat Clinic, Head and Neck Cancer Unit, San Giovanni Bosco Hospital, Turin, ITA
- Department of Oncology, University of Turin, Turin, ITA
| | - Erika Crosetti
- Ear, Nose, and Throat Clinic, Head and Neck Cancer Unit, San Giovanni Bosco Hospital, Turin, ITA
| |
Collapse
|
25
|
Báskay J, Kivovics M, Pénzes D, Kontsek E, Pesti A, Kiss A, Szócska M, Németh O, Pollner P. Reconstructing 3D histological structures using machine learning (artificial intelligence) algorithms. PATHOLOGIE (HEIDELBERG, GERMANY) 2024; 45:67-73. [PMID: 39570395 DOI: 10.1007/s00292-024-01387-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 10/11/2024] [Indexed: 11/22/2024]
Abstract
BACKGROUND Histomorphometry is currently the gold standard for bone microarchitectural examinations. This relies on two-dimensional (2D) sections to deduce the spatial properties of structures. Micromorphometric parameters are calculated from these sections based on the assumption of a plate-like 3D microarchitecture, resulting in the loss of 3D structure due to the destructive nature of classical histological processing. MATERIALS AND METHODS To overcome the limitation of histomorphometry and reconstruct the 3D architecture of bone core biopsy samples from 2D histological sections, bone core biopsy samples were decalcified and embedded in paraffin. Subsequently, 5 µm thick serial sections were stained with hematoxylin and eosin and scanned using a 3DHISTECH PANNORAMIC 1000 Digital Slide Scanner (3DHISTECH, Budapest, Hungary). A modified U‑Net architecture was trained to categorize tissues on the sections. LoFTR feature matching combined with affine transformations was employed to create the histologic reconstruction. Micromorphometric parameters were calculated using Bruker's CTAn software (v. 1.18.8.0, Bruker, Kontich, Belgium) for both histological and microCT datasets. RESULTS Our method achieved an overall accuracy of 95.26% (95% confidence interval (CI): [94.15%, 96.37%]) with an F‑score of 0.9320 (95% CI: [0.9211, 0.9429]) averaged across all classes. Correlation coefficients between micromorphometric parameters measured on microCT imaging and histological reconstruction showed a strong linear relationship, with Spearman's ρ‑values of 0.777, 0.717, 0.705, 0.666, and 0.687 for bone volume/tissue volume (BV/TV), bone surface/TV, trabecular pattern factor, trabecular thickness, and trabecular separation, respectively. Bland-Altman and mountain plots indicated good agreement between the methods for BV/TV measurements. CONCLUSION This method enables examination of tissue microarchitecture in 3D with an even higher resolution than microcomputed tomography (microCT), without losing information on cellularity. However, the limitation of this procedure is its destructive nature, which precludes subsequent mechanical testing of the sample or any further secondary measurements. Furthermore, the number of histological sections that can be created from a single sample is limited.
Collapse
Affiliation(s)
- J Báskay
- Department of Biological Physics, Eötvös Loránd University, Pázmány Péter Sétány 1/a, 1117, Budapest, Hungary
| | - M Kivovics
- Department of Community Dentistry, Semmelweis University, Szentkirályi Utca 40, 1088, Budapest, Hungary
| | - D Pénzes
- Department of Community Dentistry, Semmelweis University, Szentkirályi Utca 40, 1088, Budapest, Hungary
| | - E Kontsek
- Department of Pathology, Forensic and Insurance Medicine, Semmelweis University, Üllői út 93, 1091, Budapest, Hungary.
| | - A Pesti
- Department of Pathology, Forensic and Insurance Medicine, Semmelweis University, Üllői út 93, 1091, Budapest, Hungary
| | - A Kiss
- Department of Pathology, Forensic and Insurance Medicine, Semmelweis University, Üllői út 93, 1091, Budapest, Hungary
| | - M Szócska
- Data-Driven Health Division of National Laboratory for Health Security, Health Services Management Training Centre, Semmelweis University, Kútvölgyi út 2, 1125, Budapest, Hungary
| | - O Németh
- Department of Community Dentistry, Semmelweis University, Szentkirályi Utca 40, 1088, Budapest, Hungary
| | - P Pollner
- Department of Biological Physics, Eötvös Loránd University, Pázmány Péter Sétány 1/a, 1117, Budapest, Hungary
| |
Collapse
|
26
|
Soliman-Aboumarie H, Geers J, Lowcock D, Suji T, Kok K, Cameli M, Galiatsou E. Artificial intelligence-assisted focused cardiac ultrasound training: A survey among undergraduate medical students. ULTRASOUND (LEEDS, ENGLAND) 2024:1742271X241287923. [PMID: 39555149 PMCID: PMC11563526 DOI: 10.1177/1742271x241287923] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/06/2024] [Accepted: 09/05/2024] [Indexed: 11/19/2024]
Abstract
Objectives Focused cardiac ultrasound (FoCUS) is increasingly applied in many specialities, and adequate education and training of physicians is therefore mandatory. This study aimed to assess the impact of artificial intelligence (AI)-assisted interactive focused cardiac ultrasound (FoCUS) teaching session on undergraduate medical students' confidence level and knowledge in cardiac ultrasound. Methods The AI-assisted interactive FoCUS teaching session was held during the 9th National Undergraduate Cardiovascular Conference in London in March 2023 and all undergraduate medical students were invited to attend, and 79 students enrolled and attended the training. Two workshops were conducted each over 3-hour period. Each workshop consisted of a theoretical lecture followed by a supervised hands-on session by experts, first workshop trained 39 students and the second workshop trained 40 students. The students' pre- and post-session knowledge and confidence levels were assessed by Likert-type-scale questionnaires filled in by the students before and immediately after the workshop. Results A total of 61 pre-session and 52 post-session questionnaires were completed. Confidence level in ultrasound skills increased significantly for all six domains after the workshop, with the greatest improvement seen in obtaining basic cardiac views (p < 0.001 for all six domains). Students strongly agreed about the effectiveness of the teaching session and supported the integration of ultrasound training into their medical curriculum. Conclusions AI-assisted interactive FoCUS training can be an effective and powerful tool to increase ultrasound skills and confidence levels of undergraduate medical students. Integration of such ultrasound courses into the medical curriculum should therefore be considered.
Collapse
Affiliation(s)
- Hatem Soliman-Aboumarie
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
- School of Cardiovascular and Metabolic Medicine & Sciences, King’s College London, London, UK
| | - Jolien Geers
- Department of Cardiology, The Brussels University Hospital, Brussels, Belgium
| | - Dominic Lowcock
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
| | - Trisha Suji
- School of Cardiovascular and Metabolic Medicine & Sciences, King’s College London, London, UK
| | - Kimberley Kok
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
| | - Matteo Cameli
- School of Cardiovascular Medicine, University of Siena, Siena, Italy
| | - Eftychia Galiatsou
- Department of Anaesthetics and Critical Care, Harefield Hospital, Royal Brompton and Harefield Hospitals, London, UK
| |
Collapse
|
27
|
Chiu IM, Chen TY, Zheng YC, Lin XH, Cheng FJ, Ouyang D, Cheng CY. Prospective clinical evaluation of deep learning for ultrasonographic screening of abdominal aortic aneurysms. NPJ Digit Med 2024; 7:282. [PMID: 39406888 PMCID: PMC11480325 DOI: 10.1038/s41746-024-01269-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Accepted: 09/23/2024] [Indexed: 10/19/2024] Open
Abstract
Abdominal aortic aneurysm (AAA) often remains undetected until rupture due to limited access to diagnostic ultrasound. This trial evaluated a deep learning (DL) algorithm to guide AAA screening by novice nurses with no prior ultrasonography experience. Ten nurses performed 15 scans each on patients over 65, assisted by a DL object detection algorithm, and compared against physician-performed scans. Ultrasound scan quality, assessed by three blinded expert physicians, was the primary outcome. Among 184 patients, DL-guided novices achieved adequate scan quality in 87.5% of cases, comparable to the 91.3% by physicians (p = 0.310). The DL model predicted AAA with an AUC of 0.975, 100% sensitivity, and 97.8% specificity, with a mean absolute error of 2.8 mm in predicting aortic width compared to physicians. This study demonstrates that DL-guided POCUS has the potential to democratize AAA screening, offering performance comparable to experienced physicians and improving early detection.
Collapse
Affiliation(s)
- I-Min Chiu
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
- Department of Cardiology, Smidt Heart Institute, Cedars-Sinai Medical Center, Los Angeles, CA, USA
- Department of Computer Science and Engineering, National Sun Yat-sen University, Kaohsiung, Taiwan
| | - Tien-Yu Chen
- Division of Cardiology, Department of Internal Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - You-Cheng Zheng
- Division of Cardiology, Department of Internal Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - Xin-Hong Lin
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - Fu-Jen Cheng
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan
| | - David Ouyang
- Department of Cardiology, Smidt Heart Institute, Cedars-Sinai Medical Center, Los Angeles, CA, USA.
| | - Chi-Yung Cheng
- Department of Emergency Medicine, Kaohsiung Chang Gung Memorial Hospital, Kaohsiung, Taiwan.
- Department of Computer Science and Engineering, National Sun Yat-sen University, Kaohsiung, Taiwan.
| |
Collapse
|
28
|
Bargagna F, Zigrino D, De Santi LA, Genovesi D, Scipioni M, Favilli B, Vergaro G, Emdin M, Giorgetti A, Positano V, Santarelli MF. Automated Neural Architecture Search for Cardiac Amyloidosis Classification from [18F]-Florbetaben PET Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01275-8. [PMID: 39356368 DOI: 10.1007/s10278-024-01275-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/18/2024] [Revised: 08/30/2024] [Accepted: 09/08/2024] [Indexed: 10/03/2024]
Abstract
Medical image classification using convolutional neural networks (CNNs) is promising but often requires extensive manual tuning for optimal model definition. Neural architecture search (NAS) automates this process, reducing human intervention significantly. This study applies NAS to [18F]-Florbetaben PET cardiac images for classifying cardiac amyloidosis (CA) sub-types (amyloid light chain (AL) and transthyretin amyloid (ATTR)) and controls. Following data preprocessing and augmentation, an evolutionary cell-based NAS approach with a fixed network macro-structure is employed, automatically deriving cells' micro-structure. The algorithm is executed five times, evaluating 100 mutating architectures per run on an augmented dataset of 4048 images (originally 597), totaling 5000 architectures evaluated. The best network (NAS-Net) achieves 76.95% overall accuracy. K-fold analysis yields mean ± SD percentages of sensitivity, specificity, and accuracy on the test dataset: AL subjects (98.7 ± 2.9, 99.3 ± 1.1, 99.7 ± 0.7), ATTR-CA subjects (93.3 ± 7.8, 78.0 ± 2.9, 70.9 ± 3.7), and controls (35.8 ± 14.6, 77.1 ± 2.0, 96.7 ± 4.4). NAS-derived network performance rivals manually determined networks in the literature while using fewer parameters, validating its automatic approach's efficacy.
Collapse
Affiliation(s)
- Filippo Bargagna
- Department of Information Engineering, University of Pisa, Via G. Caruso 16, 56122, Pisa, Italy.
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy.
| | - Donato Zigrino
- Department of Information Engineering, University of Pisa, Via G. Caruso 16, 56122, Pisa, Italy
| | - Lisa Anita De Santi
- Department of Information Engineering, University of Pisa, Via G. Caruso 16, 56122, Pisa, Italy
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Dario Genovesi
- Nuclear Medicine Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Michele Scipioni
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital and Harvard Medical School, Boston, MA, USA
| | - Brunella Favilli
- Nuclear Medicine Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Giuseppe Vergaro
- Division of Cardiology and Cardiovascular Medicine, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Michele Emdin
- Division of Cardiology and Cardiovascular Medicine, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
- Health Science Interdisciplinary Center, Scuola Universitaria Superiore 'S. Anna", Piazza Martiri della Libertà 33, 56127, Pisa, Italy
| | - Assuero Giorgetti
- Nuclear Medicine Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Vincenzo Positano
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| | - Maria Filomena Santarelli
- Bioengineering Unit, Fondazione Toscana G Monasterio, Via Giuseppe Moruzzi, 56124, Pisa, Italy
- CNR Institute of Clinical Physiology, Via Giuseppe Moruzzi, 56124, Pisa, Italy
| |
Collapse
|
29
|
Wu J, Li R, Gan J, Zheng Q, Wang G, Tao W, Yang M, Li W, Ji G, Li W. Application of artificial intelligence in lung cancer screening: A real-world study in a Chinese physical examination population. Thorac Cancer 2024; 15:2061-2072. [PMID: 39206529 PMCID: PMC11444925 DOI: 10.1111/1759-7714.15428] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2024] [Revised: 07/29/2024] [Accepted: 07/31/2024] [Indexed: 09/04/2024] Open
Abstract
BACKGROUND With the rapid increase of chest computed tomography (CT) images, the workload faced by radiologists has increased dramatically. It is undeniable that the use of artificial intelligence (AI) image-assisted diagnosis system in clinical treatment is a major trend in medical development. Therefore, in order to explore the value and diagnostic accuracy of the current AI system in clinical application, we aim to compare the detection and differentiation of benign and malignant pulmonary nodules between AI system and physicians, so as to provide a theoretical basis for clinical application. METHODS Our study encompassed a cohort of 23 336 patients who underwent chest low-dose spiral CT screening for lung cancer at the Health Management Center of West China Hospital. We conducted a comparative analysis between AI-assisted reading and manual interpretation, focusing on the detection and differentiation of benign and malignant pulmonary nodules. RESULTS The AI-assisted reading exhibited a significantly higher screening positive rate and probability of diagnosing malignant pulmonary nodules compared with manual interpretation (p < 0.001). Moreover, AI scanning demonstrated a markedly superior detection rate of malignant pulmonary nodules compared with manual scanning (97.2% vs. 86.4%, p < 0.001). Additionally, the lung cancer detection rate was substantially higher in the AI reading group compared with the manual reading group (98.9% vs. 90.3%, p < 0.001). CONCLUSIONS Our findings underscore the superior screening positive rate and lung cancer detection rate achieved through AI-assisted reading compared with manual interpretation. Thus, AI exhibits considerable potential as an adjunctive tool in lung cancer screening within clinical practice settings.
Collapse
Affiliation(s)
- Jiaxuan Wu
- Department of Pulmonary and Critical Care Medicine, West China HospitalSichuan UniversityChengduSichuanChina
- State Key Laboratory of Respiratory Health and MultimorbidityWest China HospitalChengduSichuanChina
- Institute of Respiratory Health and Multimorbidity, West China HospitalSichuan UniversityChengduSichuanChina
| | - Ruicen Li
- Health Management Center, General Practice Medical Center, West China HospitalSichuan UniversityChengduChina
| | - Jiadi Gan
- Department of Pulmonary and Critical Care Medicine, West China HospitalSichuan UniversityChengduSichuanChina
- State Key Laboratory of Respiratory Health and MultimorbidityWest China HospitalChengduSichuanChina
- Institute of Respiratory Health and Multimorbidity, West China HospitalSichuan UniversityChengduSichuanChina
| | - Qian Zheng
- West China Clinical Medical CollegeSichuan UniversityChengduChina
| | - Guoqing Wang
- State Key Laboratory of Biotherapy and Cancer Center, West China HospitalSichuan UniversityChengduSichuanChina
| | - Wenjuan Tao
- Institute of Hospital Management, West China HospitalSichuan UniversityChengduChina
| | - Ming Yang
- National Clinical Research Center for Geriatrics (WCH), West China HospitalSichuan UniversityChengduChina
- Center of Gerontology and Geriatrics, West China HospitalSichuan UniversityChengduChina
| | - Wenyu Li
- Health Management Center, General Practice Medical Center, West China HospitalSichuan UniversityChengduChina
| | - Guiyi Ji
- Health Management Center, General Practice Medical Center, West China HospitalSichuan UniversityChengduChina
| | - Weimin Li
- Department of Pulmonary and Critical Care Medicine, West China HospitalSichuan UniversityChengduSichuanChina
- State Key Laboratory of Respiratory Health and MultimorbidityWest China HospitalChengduSichuanChina
- Institute of Respiratory Health and Multimorbidity, West China HospitalSichuan UniversityChengduSichuanChina
- Institute of Respiratory Health, Frontiers Science Center for Disease‐related Molecular Network, West China HospitalSichuan UniversityChengduSichuanChina
- Precision Medicine Center, Precision Medicine Key Laboratory of Sichuan Province, West China HospitalSichuan UniversityChengduSichuanChina
- The Research Units of West China, Chinese Academy of Medical SciencesWest China HospitalChengduSichuanChina
| |
Collapse
|
30
|
Yıldız Potter İ, Rodriguez EK, Wu J, Nazarian A, Vaziri A. An Automated Vertebrae Localization, Segmentation, and Osteoporotic Compression Fracture Detection Pipeline for Computed Tomographic Imaging. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2428-2443. [PMID: 38717516 PMCID: PMC11522205 DOI: 10.1007/s10278-024-01135-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Revised: 04/30/2024] [Accepted: 05/01/2024] [Indexed: 06/29/2024]
Abstract
Osteoporosis is the most common chronic metabolic bone disease worldwide. Vertebral compression fracture (VCF) is the most common type of osteoporotic fracture. Approximately 700,000 osteoporotic VCFs are diagnosed annually in the USA alone, resulting in an annual economic burden of ~$13.8B. With an aging population, the rate of osteoporotic VCFs and their associated burdens are expected to rise. Those burdens include pain, functional impairment, and increased medical expenditure. Therefore, it is of utmost importance to develop an analytical tool to aid in the identification of VCFs. Computed Tomography (CT) imaging is commonly used to detect occult injuries. Unlike the existing VCF detection approaches based on CT, the standard clinical criteria for determining VCF relies on the shape of vertebrae, such as loss of vertebral body height. We developed a novel automated vertebrae localization, segmentation, and osteoporotic VCF detection pipeline for CT scans using state-of-the-art deep learning models to bridge this gap. To do so, we employed a publicly available dataset of spine CT scans with 325 scans annotated for segmentation, 126 of which also graded for VCF (81 with VCFs and 45 without VCFs). Our approach attained 96% sensitivity and 81% specificity in detecting VCF at the vertebral-level, and 100% accuracy at the subject-level, outperforming deep learning counterparts tested for VCF detection without segmentation. Crucially, we showed that adding predicted vertebrae segments as inputs significantly improved VCF detection at both vertebral and subject levels by up to 14% Sensitivity and 20% Specificity (p-value = 0.028).
Collapse
Affiliation(s)
| | - Edward K Rodriguez
- Carl J. Shapiro Department of Orthopedic Surgery, Beth Israel Deaconess Medical Center (BIDMC), Harvard Medical School, 330 Brookline Avenue, Stoneman 10, Boston, MA, 02215, USA
- Musculoskeletal Translational Innovation Initiative, Beth Israel Deaconess Medical Center, Harvard Medical School, 330 Brookline Avenue, RN123, Boston, MA, 02215, USA
| | - Jim Wu
- Department of Radiology, Beth Israel Deaconess Medical Center (BIDMC), Harvard Medical School, 330 Brookline Avenue, Shapiro 4, Boston, MA, 02215, USA
| | - Ara Nazarian
- Carl J. Shapiro Department of Orthopedic Surgery, Beth Israel Deaconess Medical Center (BIDMC), Harvard Medical School, 330 Brookline Avenue, Stoneman 10, Boston, MA, 02215, USA
- Musculoskeletal Translational Innovation Initiative, Beth Israel Deaconess Medical Center, Harvard Medical School, 330 Brookline Avenue, RN123, Boston, MA, 02215, USA
- Department of Orthopaedics Surgery, Yerevan State University, Yerevan, Armenia
| | - Ashkan Vaziri
- BioSensics, LLC, 57 Chapel Street, Newton, MA, 02458, USA
| |
Collapse
|
31
|
Li Y, Cai P, Huang Y, Yu W, Liu Z, Liu P. Deep learning based detection and classification of fetal lip in ultrasound images. J Perinat Med 2024; 52:769-777. [PMID: 39028804 DOI: 10.1515/jpm-2024-0122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/19/2024] [Accepted: 07/07/2024] [Indexed: 07/21/2024]
Abstract
OBJECTIVES Fetal cleft lip is a common congenital defect. Considering the delicacy and difficulty of observing fetal lips, we have utilized deep learning technology to develop a new model aimed at quickly and accurately assessing the development of fetal lips during prenatal examinations. This model can detect ultrasound images of the fetal lips and classify them, aiming to provide a more objective prediction for the development of fetal lips. METHODS This study included 632 pregnant women in their mid-pregnancy stage, who underwent ultrasound examinations of the fetal lips, collecting both normal and abnormal fetal lip ultrasound images. To improve the accuracy of the detection and classification of fetal lips, we proposed and validated the Yolov5-ECA model. RESULTS The experimental results show that, compared with the currently popular 10 models, our model achieved the best results in the detection and classification of fetal lips. In terms of the detection of fetal lips, the mean average precision (mAP) at 0.5 and mAP at 0.5:0.95 were 0.920 and 0.630, respectively. In the classification of fetal lip ultrasound images, the accuracy reached 0.925. CONCLUSIONS The deep learning algorithm has accuracy consistent with manual evaluation in the detection and classification process of fetal lips. This automated recognition technology can provide a powerful tool for inexperienced young doctors, helping them to accurately conduct examinations and diagnoses of fetal lips.
Collapse
Affiliation(s)
- Yapeng Li
- School of Medicine, Huaqiao University, Quanzhou, China
| | - Peiya Cai
- Department of Gynecology and Obstetrics, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, China
| | - Yubing Huang
- Department of Ultrasound, The Second Affiliated Hospital of Fujian Medical University, Quanzhou, China
| | - Weifeng Yu
- Department of Ultrasound, Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou, China
| | - Zhonghua Liu
- Department of Ultrasound, Quanzhou First Hospital Affiliated to Fujian Medical University, Quanzhou, China
| | - Peizhong Liu
- School of Medicine, Huaqiao University, Quanzhou, China
- College of Engineering, Huaqiao University, Quanzhou, China
| |
Collapse
|
32
|
Yarahmadi P, Ahmadpour E, Moradi P, Samadzadehaghdam N. Automatic classification of Giardia infection from stool microscopic images using deep neural networks. BIOIMPACTS : BI 2024; 15:30272. [PMID: 40161947 PMCID: PMC11954735 DOI: 10.34172/bi.30272] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Revised: 04/23/2024] [Accepted: 05/28/2024] [Indexed: 04/02/2025]
Abstract
Introduction Giardiasis is a common intestinal infection caused by the Giardia lamblia parasite, and its rapid and accurate diagnosis is crucial for effective treatment. The automatic classification of Giardia infection from stool microscopic images plays a vital role in this diagnosis process. In this study, we applied deep learning-based models to automatically classify stool microscopic images into three categories, namely, normal, cyst, and trophozoite. Methods Unlike previous studies focused on images acquired from drinking water samples, we specifically targeted stool samples. In this regard, we collected a dataset of 1610 microscopic digital images captured by a smartphone with a resolution of 2340 × 1080 pixels from stool samples under the Nikon YS100 microscope. First, we applied CLAHE (Contrast Limited Adaptive Histogram Equalization) histogram equalization a method to enhance the image quality and contrast. We employed three deep learning models, namely Xception, ResNet-50, and EfficientNet-B0, to evaluate their classification performance. Each model was trained on the dataset of microscopic images and fine-tuned using transfer learning techniques. Results Among the three deep learning models, EfficientNet-B0 demonstrated superior performance in classifying Giardia lamblia parasites from stool microscopic images. The model achieved precision, accuracy, recall, specificity, and F1-score values of 0.9599, 0.9629, 0.9619, 0.9821, and 0.9607, respectively. Conclusion The EfficientNet-B0 showed promising results in accurately identifying normal, cyst, and trophozoite forms of Giardia lamblia parasites. This automated classification approach can provide valuable assistance to laboratory science experts and parasitologists in the rapid and accurate diagnosis of giardiasis, ultimately improving patient care and treatment outcomes.
Collapse
Affiliation(s)
- Pezhman Yarahmadi
- Department of Biomedical Engineering, Faculty of Advanced Medical Sciences, Tabriz University of Medical Sciences Tabriz, Iran
| | - Ehsan Ahmadpour
- Department of Parasitology and Mycology, Faculty of Medicine, Tabriz University of Medical Sciences, Tabriz, Iran
| | - Parham Moradi
- Department of Computer Engineering, University of Kurdistan, Sanandaj, Iran
| | - Nasser Samadzadehaghdam
- Department of Biomedical Engineering, Faculty of Advanced Medical Sciences, Tabriz University of Medical Sciences Tabriz, Iran
| |
Collapse
|
33
|
Shah STH, Shah SAH, Khan II, Imran A, Shah SBH, Mehmood A, Qureshi SA, Raza M, Di Terlizzi A, Cavaglià M, Deriu MA. Data-driven classification and explainable-AI in the field of lung imaging. Front Big Data 2024; 7:1393758. [PMID: 39364222 PMCID: PMC11446784 DOI: 10.3389/fdata.2024.1393758] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Accepted: 09/03/2024] [Indexed: 10/05/2024] Open
Abstract
Detecting lung diseases in medical images can be quite challenging for radiologists. In some cases, even experienced experts may struggle with accurately diagnosing chest diseases, leading to potential inaccuracies due to complex or unseen biomarkers. This review paper delves into various datasets and machine learning techniques employed in recent research for lung disease classification, focusing on pneumonia analysis using chest X-ray images. We explore conventional machine learning methods, pretrained deep learning models, customized convolutional neural networks (CNNs), and ensemble methods. A comprehensive comparison of different classification approaches is presented, encompassing data acquisition, preprocessing, feature extraction, and classification using machine vision, machine and deep learning, and explainable-AI (XAI). Our analysis highlights the superior performance of transfer learning-based methods using CNNs and ensemble models/features for lung disease classification. In addition, our comprehensive review offers insights for researchers in other medical domains too who utilize radiological images. By providing a thorough overview of various techniques, our work enables the establishment of effective strategies and identification of suitable methods for a wide range of challenges. Currently, beyond traditional evaluation metrics, researchers emphasize the importance of XAI techniques in machine and deep learning models and their applications in classification tasks. This incorporation helps in gaining a deeper understanding of their decision-making processes, leading to improved trust, transparency, and overall clinical decision-making. Our comprehensive review serves as a valuable resource for researchers and practitioners seeking not only to advance the field of lung disease detection using machine learning and XAI but also from other diverse domains.
Collapse
Affiliation(s)
- Syed Taimoor Hussain Shah
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
| | - Syed Adil Hussain Shah
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
- Department of Research and Development (R&D), GPI SpA, Trento, Italy
| | - Iqra Iqbal Khan
- Department of Computer Science, Bahauddin Zakariya University, Multan, Pakistan
| | - Atif Imran
- College of Electrical and Mechanical Engineering, National University of Sciences and Technology, Rawalpindi, Pakistan
| | - Syed Baqir Hussain Shah
- Department of Computer Science, Commission on Science and Technology for Sustainable Development in the South (COMSATS) University Islamabad (CUI), Wah Campus, Wah, Pakistan
| | - Atif Mehmood
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua, China
- Zhejiang Institute of Photoelectronics & Zhejiang Institute for Advanced Light Source, Zhejiang Normal University, Jinhua, Zhejiang, China
| | - Shahzad Ahmad Qureshi
- Department of Computer and Information Sciences, Pakistan Institute of Engineering and Applied Sciences (PIEAS), Islamabad, Pakistan
| | - Mudassar Raza
- Department of Computer Science, Namal University Mianwali, Mianwali, Pakistan
- Department of Computer Science, Heavy Industries Taxila Education City (HITEC), University of Taxila, Taxila, Pakistan
| | | | - Marco Cavaglià
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
| | - Marco Agostino Deriu
- PolitoBIOMed Lab, Department of Mechanical and Aerospace Engineering, Politecnico di Torino, Turin, Italy
| |
Collapse
|
34
|
Li F, Xu Y, Lemus OD, Wang TJC, Sisti MB, Wuu CS. Synthetic CT for gamma knife radiosurgery dose calculation: A feasibility study. Phys Med 2024; 125:104504. [PMID: 39197262 DOI: 10.1016/j.ejmp.2024.104504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Revised: 06/24/2024] [Accepted: 08/22/2024] [Indexed: 09/01/2024] Open
Abstract
PURPOSE To determine if MRI-based synthetic CTs (sCT), generated with no predefined pulse sequence, can be used for inhomogeneity correction in routine gamma knife radiosurgery (GKRS) treatment planning dose calculation. METHODS Two sets of sCTs were generated from T1post and T2 images using cycleGAN. Twenty-eight patients (18 training, 10 validation) were retrospectively selected. The image quality of the generated sCTs was compared with the original CT (oCT) regarding the HU value preservation using histogram comparison, RMSE and MAE, and structural integrity. Dosimetric comparisons were also made among GKRS plans from 3 calculation approaches: TMR10 (oCT), and convolution (oCT and sCT), at four locations: original disease site, bone/tissue interface, air/tissue interface, and mid-brain. RESULTS The study showed that sCTs and oCTs' HU were similar, with T2-sCT performing better. TMR10 significantly underdosed the target by a mean of 5.4% compared to the convolution algorithm. There was no significant difference in convolution algorithm shot time between the oCT and sCT generated with T2. The highest and lowest dosimetric differences between the two CTs were observed in the bone and air interface, respectively. Dosimetric differences of 3.3% were observed in sCT predicted from MRI with stereotactic frames, which was not included in the training sets. CONCLUSIONS MRI-based sCT can be utilized for GKRS convolution dose calculation without the unnecessary radiation dose, and sCT without metal artifacts could be generated in framed cases. Larger datasets inclusive of all pulse sequences can improve the training set. Further investigation and validation studies are needed before clinical implementation.
Collapse
Affiliation(s)
- Fiona Li
- Department of Radiation Oncology, Columbia University, New York, NY, USA.
| | - Yuanguang Xu
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| | - Olga D Lemus
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| | - Tony J C Wang
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| | - Michael B Sisti
- Department of Neurological Surgery, Columbia University, New York, NY, USA
| | - Cheng-Shie Wuu
- Department of Radiation Oncology, Columbia University, New York, NY, USA
| |
Collapse
|
35
|
Cai M, Zhao L, Qiang Y, Wang L, Zhao J. CHNet: A multi-task global-local Collaborative Hybrid Network for KRAS mutation status prediction in colorectal cancer. Artif Intell Med 2024; 155:102931. [PMID: 39094228 DOI: 10.1016/j.artmed.2024.102931] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 06/29/2024] [Accepted: 07/03/2024] [Indexed: 08/04/2024]
Abstract
Accurate prediction of Kirsten rat sarcoma (KRAS) mutation status is crucial for personalized treatment of advanced colorectal cancer patients. However, despite the excellent performance of deep learning models in certain aspects, they often overlook the synergistic promotion among multiple tasks and the consideration of both global and local information, which can significantly reduce prediction accuracy. To address these issues, this paper proposes an innovative method called the Multi-task Global-Local Collaborative Hybrid Network (CHNet) aimed at more accurately predicting patients' KRAS mutation status. CHNet consists of two branches that can extract global and local features from segmentation and classification tasks, respectively, and exchange complementary information to collaborate in executing these tasks. Within the two branches, we have designed a Channel-wise Hybrid Transformer (CHT) and a Spatial-wise Hybrid Transformer (SHT). These transformers integrate the advantages of both Transformer and CNN, employing cascaded hybrid attention and convolution to capture global and local information from the two tasks. Additionally, we have created an Adaptive Collaborative Attention (ACA) module to facilitate the collaborative fusion of segmentation and classification features through guidance. Furthermore, we introduce a novel Class Activation Map (CAM) loss to encourage CHNet to learn complementary information between the two tasks. We evaluate CHNet on the T2-weighted MRI dataset, and achieve an accuracy of 88.93% in KRAS mutation status prediction, which outperforms the performance of representative KRAS mutation status prediction methods. The results suggest that our CHNet can more accurately predict KRAS mutation status in patients via a multi-task collaborative facilitation and considering global-local information way, which can assist doctors in formulating more personalized treatment strategies for patients.
Collapse
Affiliation(s)
- Meiling Cai
- College of computer science and technology (College of data science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China.
| | - Lin Zhao
- Southeast University, Nanjing, 210037, Jiangsu, China
| | - Yan Qiang
- College of computer science and technology (College of data science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China
| | - Long Wang
- Jinzhong College of Information, Jinzhong, 030800, Shanxi, China
| | - Juanjuan Zhao
- College of computer science and technology (College of data science), Taiyuan University of Technology, Taiyuan, 030024, Shanxi, China.
| |
Collapse
|
36
|
Li K, Han X, Meng Y, Li J, Hong Y, Chen X, You JY, Yao L, Hu W, Xia Z, Ke G, Zhang L, Zhang J, Zhao X. Single-Image-Based Deep Learning for Precise Atomic Defect Identification. NANO LETTERS 2024; 24:10275-10283. [PMID: 39106329 DOI: 10.1021/acs.nanolett.4c02654] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/09/2024]
Abstract
Defect engineering is widely used to impart the desired functionalities on materials. Despite the widespread application of atomic-resolution scanning transmission electron microscopy (STEM), traditional methods for defect analysis are highly sensitive to random noise and human bias. While deep learning (DL) presents a viable alternative, it requires extensive amounts of training data with labeled ground truth. Herein, employing cycle generative adversarial networks (CycleGAN) and U-Nets, we propose a method based on a single experimental STEM image to tackle high annotation costs and image noise for defect detection. Not only atomic defects but also oxygen dopants in monolayer MoS2 are visualized. The method can be readily extended to other two-dimensional systems, as the training is based on unit-cell-level images. Therefore, our results outline novel ways to train the model with minimal data sets, offering great opportunities to fully exploit the power of DL in the materials science community.
Collapse
Affiliation(s)
- Kangshu Li
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Xiaocang Han
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Yuan Meng
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Junxian Li
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | | | - Xiang Chen
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Jing-Yang You
- Department of Physics, National University of Singapore, Singapore 117551
| | - Lin Yao
- DP Technology, Beijing 100080, China
| | - Wenchao Hu
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Zhiyi Xia
- DP Technology, Beijing 100080, China
| | - Guolin Ke
- DP Technology, Beijing 100080, China
| | - Linfeng Zhang
- DP Technology, Beijing 100080, China
- AI for Science Institute, Beijing 100084, China
| | - Jin Zhang
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
| | - Xiaoxu Zhao
- School of Materials Science and Engineering, Peking University, Beijing 100871, China
- AI for Science Institute, Beijing 100084, China
| |
Collapse
|
37
|
Stripelis D, Gupta U, Saleem H, Dhinagar N, Ghai T, Anastasiou C, Sánchez R, Steeg GV, Ravi S, Naveed M, Thompson PM, Ambite JL. A federated learning architecture for secure and private neuroimaging analysis. PATTERNS (NEW YORK, N.Y.) 2024; 5:101031. [PMID: 39233693 PMCID: PMC11368680 DOI: 10.1016/j.patter.2024.101031] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/10/2023] [Revised: 04/04/2024] [Accepted: 06/06/2024] [Indexed: 09/06/2024]
Abstract
The amount of biomedical data continues to grow rapidly. However, collecting data from multiple sites for joint analysis remains challenging due to security, privacy, and regulatory concerns. To overcome this challenge, we use federated learning, which enables distributed training of neural network models over multiple data sources without sharing data. Each site trains the neural network over its private data for some time and then shares the neural network parameters (i.e., weights and/or gradients) with a federation controller, which in turn aggregates the local models and sends the resulting community model back to each site, and the process repeats. Our federated learning architecture, MetisFL, provides strong security and privacy. First, sample data never leave a site. Second, neural network parameters are encrypted before transmission and the global neural model is computed under fully homomorphic encryption. Finally, we use information-theoretic methods to limit information leakage from the neural model to prevent a "curious" site from performing model inversion or membership attacks. We present a thorough evaluation of the performance of secure, private federated learning in neuroimaging tasks, including for predicting Alzheimer's disease and for brain age gap estimation (BrainAGE) from magnetic resonance imaging (MRI) studies in challenging, heterogeneous federated environments where sites have different amounts of data and statistical distributions.
Collapse
Affiliation(s)
- Dimitris Stripelis
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Umang Gupta
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Hamza Saleem
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Nikhil Dhinagar
- University of Southern California, Imaging Genetics Center, Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, Marina del Rey, CA 90292, USA
| | - Tanmay Ghai
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | | | - Rafael Sánchez
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Greg Ver Steeg
- University of California, Riverside, Riverside, CA 92521, USA
| | - Srivatsan Ravi
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Muhammad Naveed
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| | - Paul M. Thompson
- University of Southern California, Imaging Genetics Center, Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, Marina del Rey, CA 90292, USA
| | - José Luis Ambite
- University of Southern California, Information Sciences Institute, Marina del Rey, CA 90292, USA
- University of Southern California, Computer Science Department, Los Angeles, CA 90089, USA
| |
Collapse
|
38
|
Zhao Y, Dohi O, Ishida T, Yoshida N, Ochiai T, Mukai H, Seya M, Yamauchi K, Miyazaki H, Fukui H, Yasuda T, Iwai N, Inoue K, Itoh Y, Liu X, Zhang R, Zhu X. Linked Color Imaging with Artificial Intelligence Improves the Detection of Early Gastric Cancer. Dig Dis 2024; 42:503-511. [PMID: 39102801 DOI: 10.1159/000540728] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/22/2024] [Accepted: 07/31/2024] [Indexed: 08/07/2024]
Abstract
INTRODUCTION Esophagogastroduodenoscopy is the most important tool to detect gastric cancer (GC). In this study, we developed a computer-aided detection (CADe) system to detect GC with white light imaging (WLI) and linked color imaging (LCI) modes and aimed to compare the performance of CADe with that of endoscopists. METHODS The system was developed based on the deep learning framework from 9,021 images in 385 patients between 2017 and 2020. A total of 116 LCI and WLI videos from 110 patients between 2017 and 2023 were used to evaluate per-case sensitivity and per-frame specificity. RESULTS The per-case sensitivity and per-frame specificity of CADe with a confidence level of 0.5 in detecting GC were 78.6% and 93.4% for WLI and 94.0% and 93.3% for LCI, respectively (p < 0.001). The per-case sensitivities of nonexpert endoscopists for WLI and LCI were 45.8% and 80.4%, whereas those of expert endoscopists were 66.7% and 90.6%, respectively. Regarding detectability between CADe and endoscopists, the per-case sensitivities for WLI and LCI were 78.6% and 94.0% in CADe, respectively, which were significantly higher than those for LCI in experts (90.6%, p = 0.004) and those for WLI and LCI in nonexperts (45.8% and 80.4%, respectively, p < 0.001); however, no significant difference for WLI was observed between CADe and experts (p = 0.134). CONCLUSIONS Our CADe system showed significantly better sensitivity in detecting GC when used in LCI compared with WLI mode. Moreover, the sensitivity of CADe using LCI is significantly higher than those of expert endoscopists using LCI to detect GC.
Collapse
Affiliation(s)
- Youshen Zhao
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| | - Osamu Dohi
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Tsugitaka Ishida
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Naohisa Yoshida
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Tomoko Ochiai
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Hiroki Mukai
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Mayuko Seya
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Katsuma Yamauchi
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Hajime Miyazaki
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Hayato Fukui
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Takeshi Yasuda
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Naoto Iwai
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Ken Inoue
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Yoshito Itoh
- Molecular Gastroenterology and Hepatology, Graduate School of Medical Science, Kyoto Prefectural University of Medicine, Kyoto, Japan
| | - Xinkai Liu
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| | - Ruiyao Zhang
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| | - Xin Zhu
- Graduate School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu, Fukushima, Japan
| |
Collapse
|
39
|
Dunenova G, Kalmataeva Z, Kaidarova D, Dauletbaev N, Semenova Y, Mansurova M, Grjibovski A, Kassymbekova F, Sarsembayev A, Semenov D, Glushkova N. The Performance and Clinical Applicability of HER2 Digital Image Analysis in Breast Cancer: A Systematic Review. Cancers (Basel) 2024; 16:2761. [PMID: 39123488 PMCID: PMC11311684 DOI: 10.3390/cancers16152761] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Revised: 07/28/2024] [Accepted: 07/30/2024] [Indexed: 08/12/2024] Open
Abstract
This systematic review aims to address the research gap in the performance of computational algorithms for the digital image analysis of HER2 images in clinical settings. While numerous studies have explored various aspects of these algorithms, there is a lack of comprehensive evaluation regarding their effectiveness in real-world clinical applications. We conducted a search of the Web of Science and PubMed databases for studies published from 31 December 2013 to 30 June 2024, focusing on performance effectiveness and components such as dataset size, diversity and source, ground truth, annotation, and validation methods. The study was registered with PROSPERO (CRD42024525404). Key questions guiding this review include the following: How effective are current computational algorithms at detecting HER2 status in digital images? What are the common validation methods and dataset characteristics used in these studies? Is there standardization of algorithm evaluations of clinical applications that can improve the clinical utility and reliability of computational tools for HER2 detection in digital image analysis? We identified 6833 publications, with 25 meeting the inclusion criteria. The accuracy rate with clinical datasets varied from 84.19% to 97.9%. The highest accuracy was achieved on the publicly available Warwick dataset at 98.8% in synthesized datasets. Only 12% of studies used separate datasets for external validation; 64% of studies used a combination of accuracy, precision, recall, and F1 as a set of performance measures. Despite the high accuracy rates reported in these studies, there is a notable absence of direct evidence supporting their clinical application. To facilitate the integration of these technologies into clinical practice, there is an urgent need to address real-world challenges and overreliance on internal validation. Standardizing study designs on real clinical datasets can enhance the reliability and clinical applicability of computational algorithms in improving the detection of HER2 cancer.
Collapse
Affiliation(s)
- Gauhar Dunenova
- Department of Epidemiology, Biostatistics and Evidence-Based Medicine, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
| | - Zhanna Kalmataeva
- Rector Office, Asfendiyarov Kazakh National Medical University, Almaty 050000, Kazakhstan;
| | - Dilyara Kaidarova
- Kazakh Research Institute of Oncology and Radiology, Almaty 050022, Kazakhstan;
| | - Nurlan Dauletbaev
- Department of Internal, Respiratory and Critical Care Medicine, Philipps University of Marburg, 35037 Marburg, Germany;
- Department of Pediatrics, Faculty of Medicine and Health Sciences, McGill University, Montreal, QC H4A 3J1, Canada
- Faculty of Medicine and Health Care, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
| | - Yuliya Semenova
- School of Medicine, Nazarbayev University, Astana 010000, Kazakhstan;
| | - Madina Mansurova
- Department of Artificial Intelligence and Big Data, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan;
| | - Andrej Grjibovski
- Central Scientific Research Laboratory, Northern State Medical University, Arkhangelsk 163000, Russia;
- Department of Epidemiology and Modern Vaccination Technologies, I.M. Sechenov First Moscow State Medical University, Moscow 105064, Russia
- Department of Biology, Ecology and Biotechnology, Northern (Arctic) Federal University, Arkhangelsk 163000, Russia
- Department of Health Policy and Management, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
| | - Fatima Kassymbekova
- Department of Public Health and Social Sciences, Kazakhstan Medical University “KSPH”, Almaty 050060, Kazakhstan;
| | - Aidos Sarsembayev
- School of Digital Technologies, Almaty Management University, Almaty 050060, Kazakhstan;
- Health Research Institute, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan;
| | - Daniil Semenov
- Computer Science and Engineering Program, Astana IT University, Astana 020000, Kazakhstan;
| | - Natalya Glushkova
- Department of Epidemiology, Biostatistics and Evidence-Based Medicine, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan
- Health Research Institute, Al-Farabi Kazakh National University, Almaty 050040, Kazakhstan;
| |
Collapse
|
40
|
Badkul A, Vamsi I, Sudha R. Comparative study of DCNN and image processing based classification of chest X-rays for identification of COVID-19 patients using fine-tuning. J Med Eng Technol 2024; 48:213-222. [PMID: 39648993 DOI: 10.1080/03091902.2024.2438158] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 11/20/2024] [Accepted: 11/30/2024] [Indexed: 12/10/2024]
Abstract
The conventional detection of COVID-19 by evaluating the CT scan images is tiresome, often experiences high inter-observer variability and uncertainty issues. This work proposes the automatic detection and classification of COVID-19 by analysing the chest X-ray images (CXR) with the deep convolutional neural network (DCNN) models through a fine-tuning and pre-training approach. CXR images pertaining to four health scenarios, namely, healthy, COVID-19, bacterial pneumonia and viral pneumonia, are considered and subjected to data augmentation. Two types of input datasets are prepared; in which dataset I contains the original image dataset categorised under four classes, whereas the original CXR images are subjected to image pre-processing via Contrast Limited Adaptive Histogram Equalisation (CLAHE) algorithm and Blackhat Morphological Operation (BMO) for devising the input dataset II. Both datasets are supplied as input to various DCNN models such as DenseNet, MobileNet, ResNet, VGG16, and Xception for achieving multi-class classification. It is observed that the classification accuracies are improved, and the classification errors are reduced with the image pre-processing. Overall, the VGG16 model resulted in better classification accuracies and reduced classification errors while accomplishing multi-class classification. Thus, the proposed work would assist the clinical diagnosis, and reduce the workload of the front-line healthcare workforce and medical professionals.
Collapse
Affiliation(s)
- Amitesh Badkul
- Department of Electrical and Electronics, Birla Institute of Technology and Science-Pilani, Hyderabad, India
| | - Inturi Vamsi
- Mechanical Engineering Department, Chaitanya Bharathi Institute of Technology (A), Hyderabad, India
| | - Radhika Sudha
- Department of Electrical and Electronics, Birla Institute of Technology and Science-Pilani, Hyderabad, India
| |
Collapse
|
41
|
Rozhyna A, Somfai GM, Atzori M, DeBuc DC, Saad A, Zoellin J, Müller H. Exploring Publicly Accessible Optical Coherence Tomography Datasets: A Comprehensive Overview. Diagnostics (Basel) 2024; 14:1668. [PMID: 39125544 PMCID: PMC11312046 DOI: 10.3390/diagnostics14151668] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Revised: 07/15/2024] [Accepted: 07/25/2024] [Indexed: 08/12/2024] Open
Abstract
Artificial intelligence has transformed medical diagnostic capabilities, particularly through medical image analysis. AI algorithms perform well in detecting abnormalities with a strong performance, enabling computer-aided diagnosis by analyzing the extensive amounts of patient data. The data serve as a foundation upon which algorithms learn and make predictions. Thus, the importance of data cannot be underestimated, and clinically corresponding datasets are required. Many researchers face a lack of medical data due to limited access, privacy concerns, or the absence of available annotations. One of the most widely used diagnostic tools in ophthalmology is Optical Coherence Tomography (OCT). Addressing the data availability issue is crucial for enhancing AI applications in the field of OCT diagnostics. This review aims to provide a comprehensive analysis of all publicly accessible retinal OCT datasets. Our main objective is to compile a list of OCT datasets and their properties, which can serve as an accessible reference, facilitating data curation for medical image analysis tasks. For this review, we searched through the Zenodo repository, Mendeley Data repository, MEDLINE database, and Google Dataset search engine. We systematically evaluated all the identified datasets and found 23 open-access datasets containing OCT images, which significantly vary in terms of size, scope, and ground-truth labels. Our findings indicate the need for improvement in data-sharing practices and standardized documentation. Enhancing the availability and quality of OCT datasets will support the development of AI algorithms and ultimately improve diagnostic capabilities in ophthalmology. By providing a comprehensive list of accessible OCT datasets, this review aims to facilitate better utilization and development of AI in medical image analysis.
Collapse
Affiliation(s)
- Anastasiia Rozhyna
- Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO), 3960 Sierre, Switzerland
- Medical Informatics, University of Geneva, 1205 Geneva, Switzerland
| | - Gábor Márk Somfai
- Department of Ophthalmology, Stadtspital Zürich, 8063 Zurich, Switzerland
- Spross Research Institute, 8063 Zurich, Switzerland
| | - Manfredo Atzori
- Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO), 3960 Sierre, Switzerland
- Department of Neuroscience, University of Padua, 35121 Padova, Italy
| | - Delia Cabrera DeBuc
- Bascom Palmer Eye Institute, Miller School of Medicine, University of Miami, Miami, FL 33136, USA
| | - Amr Saad
- Department of Ophthalmology, Stadtspital Zürich, 8063 Zurich, Switzerland
- Spross Research Institute, 8063 Zurich, Switzerland
| | - Jay Zoellin
- Department of Ophthalmology, Stadtspital Zürich, 8063 Zurich, Switzerland
- Spross Research Institute, 8063 Zurich, Switzerland
| | - Henning Müller
- Informatics Institute, University of Applied Sciences Western Switzerland (HES-SO), 3960 Sierre, Switzerland
- Medical Informatics, University of Geneva, 1205 Geneva, Switzerland
- The Sense Research and Innovation Center, 1007 Lausanne, Switzerland
| |
Collapse
|
42
|
Sureshkumar V, Prasad RSN, Balasubramaniam S, Jagannathan D, Daniel J, Dhanasekaran S. Breast Cancer Detection and Analytics Using Hybrid CNN and Extreme Learning Machine. J Pers Med 2024; 14:792. [PMID: 39201984 PMCID: PMC11355507 DOI: 10.3390/jpm14080792] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2024] [Revised: 07/08/2024] [Accepted: 07/15/2024] [Indexed: 09/03/2024] Open
Abstract
Early detection of breast cancer is essential for increasing survival rates, as it is one of the primary causes of death for women globally. Mammograms are extensively used by physicians for diagnosis, but selecting appropriate algorithms for image enhancement, segmentation, feature extraction, and classification remains a significant research challenge. This paper presents a computer-aided diagnosis (CAD)-based hybrid model combining convolutional neural networks (CNN) with a pruned ensembled extreme learning machine (HCPELM) to enhance breast cancer detection, segmentation, feature extraction, and classification. The model employs the rectified linear unit (ReLU) activation function to enhance data analytics after removing artifacts and pectoral muscles, and the HCPELM hybridized with the CNN model improves feature extraction. The hybrid elements are convolutional and fully connected layers. Convolutional layers extract spatial features like edges, textures, and more complex features in deeper layers. The fully connected layers take these features and combine them in a non-linear manner to perform the final classification. ELM performs classification and recognition tasks, aiming for state-of-the-art performance. This hybrid classifier is used for transfer learning by freezing certain layers and modifying the architecture to reduce parameters, easing cancer detection. The HCPELM classifier was trained using the MIAS database and evaluated against benchmark methods. It achieved a breast image recognition accuracy of 86%, outperforming benchmark deep learning models. HCPELM is demonstrating superior performance in early detection and diagnosis, thus aiding healthcare practitioners in breast cancer diagnosis.
Collapse
Affiliation(s)
- Vidhushavarshini Sureshkumar
- Department of Computer Science and Engineering, SRM Institute of Science and Technology, Vadapalani, Chennai 600026, India
| | | | | | - Dhayanithi Jagannathan
- Department of Computer Science and Engineering, Sona College of Technology, Salem 636005, India; (S.B.); (D.J.)
| | - Jayanthi Daniel
- Department of Electronics and Communication Engineering, Rajalakshmi Engineering College, Chennai 602105, India;
| | | |
Collapse
|
43
|
Ahmad J, Akram S, Jaffar A, Ali Z, Bhatti SM, Ahmad A, Rehman SU. Deep learning empowered breast cancer diagnosis: Advancements in detection and classification. PLoS One 2024; 19:e0304757. [PMID: 38990817 PMCID: PMC11239011 DOI: 10.1371/journal.pone.0304757] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Accepted: 05/18/2024] [Indexed: 07/13/2024] Open
Abstract
Recent advancements in AI, driven by big data technologies, have reshaped various industries, with a strong focus on data-driven approaches. This has resulted in remarkable progress in fields like computer vision, e-commerce, cybersecurity, and healthcare, primarily fueled by the integration of machine learning and deep learning models. Notably, the intersection of oncology and computer science has given rise to Computer-Aided Diagnosis (CAD) systems, offering vital tools to aid medical professionals in tumor detection, classification, recurrence tracking, and prognosis prediction. Breast cancer, a significant global health concern, is particularly prevalent in Asia due to diverse factors like lifestyle, genetics, environmental exposures, and healthcare accessibility. Early detection through mammography screening is critical, but the accuracy of mammograms can vary due to factors like breast composition and tumor characteristics, leading to potential misdiagnoses. To address this, an innovative CAD system leveraging deep learning and computer vision techniques was introduced. This system enhances breast cancer diagnosis by independently identifying and categorizing breast lesions, segmenting mass lesions, and classifying them based on pathology. Thorough validation using the Curated Breast Imaging Subset of Digital Database for Screening Mammography (CBIS-DDSM) demonstrated the CAD system's exceptional performance, with a 99% success rate in detecting and classifying breast masses. While the accuracy of detection is 98.5%, when segmenting breast masses into separate groups for examination, the method's performance was approximately 95.39%. Upon completing all the analysis, the system's classification phase yielded an overall accuracy of 99.16% for classification. The potential for this integrated framework to outperform current deep learning techniques is proposed, despite potential challenges related to the high number of trainable parameters. Ultimately, this recommended framework offers valuable support to researchers and physicians in breast cancer diagnosis by harnessing cutting-edge AI and image processing technologies, extending recent advances in deep learning to the medical domain.
Collapse
Affiliation(s)
- Jawad Ahmad
- Faculty of Computer Science & Information Technology, The Superior University, Lahore, Pakistan
- Intelligent Data Visual Computing Research (IDVCR), Lahore, Pakistan
| | - Sheeraz Akram
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - Arfan Jaffar
- Faculty of Computer Science & Information Technology, The Superior University, Lahore, Pakistan
- Intelligent Data Visual Computing Research (IDVCR), Lahore, Pakistan
| | - Zulfiqar Ali
- School of Computer Science and Electronic Engineering (CSEE), University of Essex, Wivenhoe Park, Colchester, United Kingdom
| | - Sohail Masood Bhatti
- Faculty of Computer Science & Information Technology, The Superior University, Lahore, Pakistan
- Intelligent Data Visual Computing Research (IDVCR), Lahore, Pakistan
| | - Awais Ahmad
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| | - Shafiq Ur Rehman
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh, Saudi Arabia
| |
Collapse
|
44
|
Li C, Zhang F, Du Y, Li H. Classification of brain tumor types through MRIs using parallel CNNs and firefly optimization. Sci Rep 2024; 14:15057. [PMID: 38956224 PMCID: PMC11219740 DOI: 10.1038/s41598-024-65714-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2024] [Accepted: 06/24/2024] [Indexed: 07/04/2024] Open
Abstract
Image segmentation is a critical and challenging endeavor in the field of medicine. A magnetic resonance imaging (MRI) scan is a helpful method for locating any abnormal brain tissue these days. It is a difficult undertaking for radiologists to diagnose and classify the tumor from several pictures. This work develops an intelligent method for accurately identifying brain tumors. This research investigates the identification of brain tumor types from MRI data using convolutional neural networks and optimization strategies. Two novel approaches are presented: the first is a novel segmentation technique based on firefly optimization (FFO) that assesses segmentation quality based on many parameters, and the other is a combination of two types of convolutional neural networks to categorize tumor traits and identify the kind of tumor. These upgrades are intended to raise the general efficacy of the MRI scan technique and increase identification accuracy. Using MRI scans from BBRATS2018, the testing is carried out, and the suggested approach has shown improved performance with an average accuracy of 98.6%.
Collapse
Affiliation(s)
- Chen Li
- Department of Neurosurgery, Shandong Provincial Third Hospital, Shandong University, No.12 Wuyingshan Middle Road, Jinan, 250031, Shandong, China
| | - Faxue Zhang
- Department of Neurosurgery, Shandong Provincial Third Hospital, Shandong University, No.12 Wuyingshan Middle Road, Jinan, 250031, Shandong, China
| | - Yongjian Du
- Department of Neurosurgery, The Fifth People's Hospital of Jinan, No.24297, Jingshi Road, Jinan, 250022, Shandong, China
| | - Huachao Li
- Department of Neurosurgery, Shandong Provincial Third Hospital, Shandong University, No.12 Wuyingshan Middle Road, Jinan, 250031, Shandong, China.
| |
Collapse
|
45
|
Jiang C, Ji T, Qiao Q. Application and progress of artificial intelligence in radiation therapy dose prediction. Clin Transl Radiat Oncol 2024; 47:100792. [PMID: 38779524 PMCID: PMC11109740 DOI: 10.1016/j.ctro.2024.100792] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2024] [Accepted: 05/07/2024] [Indexed: 05/25/2024] Open
Abstract
Radiation therapy (RT) nowadays is a main treatment modality of cancer. To ensure the therapeutic efficacy of patients, accurate dose distribution is often required, which is a time-consuming and labor-intensive process. In addition, due to the differences in knowledge and experience among participants and diverse institutions, the predicted dose are often inconsistent. In last several decades, artificial intelligence (AI) has been applied in various aspects of RT, several products have been implemented in clinical practice and confirmed superiority. In this paper, we will review the research of AI in dose prediction, focusing on the progress in deep learning (DL).
Collapse
Affiliation(s)
- Chen Jiang
- Department of Radiation Oncology, The First Hospital of China Medical University, Shenyang, China
| | - Tianlong Ji
- Department of Radiation Oncology, The First Hospital of China Medical University, Shenyang, China
| | - Qiao Qiao
- Department of Radiation Oncology, The First Hospital of China Medical University, Shenyang, China
| |
Collapse
|
46
|
Zhu K, Shen Z, Wang M, Jiang L, Zhang Y, Yang T, Zhang H, Zhang M. Visual Knowledge Domain of Artificial Intelligence in Computed Tomography: A Review Based on Bibliometric Analysis. J Comput Assist Tomogr 2024; 48:652-662. [PMID: 38271538 DOI: 10.1097/rct.0000000000001585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2024]
Abstract
ABSTRACT Artificial intelligence (AI)-assisted medical imaging technology is a new research area of great interest that has developed rapidly over the last decade. However, there has been no bibliometric analysis of published studies in this field. The present review focuses on AI-related studies on computed tomography imaging in the Web of Science database and uses CiteSpace and VOSviewer to generate a knowledge map and conduct the basic information analysis, co-word analysis, and co-citation analysis. A total of 7265 documents were included and the number of documents published had an overall upward trend. Scholars from the United States and China have made outstanding achievements, and there is a general lack of extensive cooperation in this field. In recent years, the research areas of great interest and difficulty have been the optimization and upgrading of algorithms, and the application of theoretical models to practical clinical applications. This review will help researchers understand the developments, research areas of great interest, and research frontiers in this field and provide reference and guidance for future studies.
Collapse
|
47
|
Yilmaz S, Tasyurek M, Amuk M, Celik M, Canger EM. Developing deep learning methods for classification of teeth in dental panoramic radiography. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 138:118-127. [PMID: 37316425 DOI: 10.1016/j.oooo.2023.02.021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2022] [Revised: 09/13/2022] [Accepted: 02/10/2023] [Indexed: 06/16/2023]
Abstract
OBJECTIVES We aimed to develop an artificial intelligence-based clinical dental decision-support system using deep-learning methods to reduce diagnostic interpretation error and time and increase the effectiveness of dental treatment and classification. STUDY DESIGN We compared the performance of 2 deep-learning methods, You Only Look Once V4 (YOLO-V4) and Faster Regions with the Convolutional Neural Networks (R-CNN), for tooth classification in dental panoramic radiography for tooth classification in dental panoramic radiography to determine which is more successful in terms of accuracy, time, and detection ability. Using a method based on deep-learning models trained on a semantic segmentation task, we analyzed 1200 panoramic radiographs selected retrospectively. In the classification process, our model identified 36 classes, including 32 teeth and 4 impacted teeth. RESULTS The YOLO-V4 method achieved a mean 99.90% precision, 99.18% recall, and 99.54% F1 score. The Faster R-CNN method achieved a mean 93.67% precision, 90.79% recall, and 92.21% F1 score. Experimental evaluations showed that the YOLO-V4 method outperformed the Faster R-CNN method in terms of accuracy of predicted teeth in the tooth classification process, speed of tooth classification, and ability to detect impacted and erupted third molars. CONCLUSIONS The YOLO-V4 method outperforms the Faster R-CNN method in terms of accuracy of tooth prediction, speed of detection, and ability to detect impacted third molars and erupted third molars. The proposed deep learning based methods can assist dentists in clinical decision making, save time, and reduce the negative effects of stress and fatigue in daily practice.
Collapse
Affiliation(s)
- Serkan Yilmaz
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Erciyes University, Kayseri, Turkey
| | - Murat Tasyurek
- Department of Computer Engineering, Kayseri University, Kayseri, Turkey
| | - Mehmet Amuk
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Erciyes University, Kayseri, Turkey
| | - Mete Celik
- Department of Computer Engineering, Erciyes University, Kayseri, Turkey
| | - Emin Murat Canger
- Faculty of Dentistry, Department of Oral and Maxillofacial Radiology, Erciyes University, Kayseri, Turkey.
| |
Collapse
|
48
|
Altındağ A, Bahrilli S, Çelik Ö, Bayrakdar İŞ, Orhan K. Tooth numbering and classification on bitewing radiographs: an artificial intelligence pilot study. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 137:679-689. [PMID: 38632035 DOI: 10.1016/j.oooo.2024.02.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Revised: 01/13/2024] [Accepted: 02/08/2024] [Indexed: 04/19/2024]
Abstract
OBJECTIVE The aim of this study is to assess the efficacy of employing a deep learning methodology for the automated identification and enumeration of permanent teeth in bitewing radiographs. The experimental procedures and techniques employed in this study are described in the following section. STUDY DESIGN A total of 1248 bitewing radiography images were annotated using the CranioCatch labeling program, developed in Eskişehir, Turkey. The dataset has been partitioned into 3 subsets: training (n = 1000, 80% of the total), validation (n = 124, 10% of the total), and test (n = 124, 10% of the total) sets. The images were subjected to a 3 × 3 clash operation in order to enhance the clarity of the labeled regions. RESULTS The F1, sensitivity and precision results of the artificial intelligence model obtained using the Yolov5 architecture in the test dataset were found to be 0.9913, 0.9954, and 0.9873, respectively. CONCLUSION The utilization of numerical identification for teeth within deep learning-based artificial intelligence algorithms applied to bitewing radiographs has demonstrated notable efficacy. The utilization of clinical decision support system software, which is augmented by artificial intelligence, has the potential to enhance the efficiency and effectiveness of dental practitioners.
Collapse
Affiliation(s)
- Ali Altındağ
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Necmettin Erbakan University, Konya, Turkey.
| | - Serkan Bahrilli
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Necmettin Erbakan University, Konya, Turkey
| | - Özer Çelik
- Department of Mathematics-Computer, Eskisehir Osmangazi University Faculty of Science, Eskisehir, Turkey
| | - İbrahim Şevki Bayrakdar
- Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Eskişehir Osmangazi University, Eskişehir, Turkey
| | - Kaan Orhan
- Department of Oral and Maxillofacial Radiology, Faculty of Dentistry, Ankara University, Ankara, Turkey
| |
Collapse
|
49
|
Botnari A, Kadar M, Patrascu JM. A Comprehensive Evaluation of Deep Learning Models on Knee MRIs for the Diagnosis and Classification of Meniscal Tears: A Systematic Review and Meta-Analysis. Diagnostics (Basel) 2024; 14:1090. [PMID: 38893617 PMCID: PMC11172202 DOI: 10.3390/diagnostics14111090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2024] [Revised: 05/19/2024] [Accepted: 05/21/2024] [Indexed: 06/21/2024] Open
Abstract
OBJECTIVES This study delves into the cutting-edge field of deep learning techniques, particularly deep convolutional neural networks (DCNNs), which have demonstrated unprecedented potential in assisting radiologists and orthopedic surgeons in precisely identifying meniscal tears. This research aims to evaluate the effectiveness of deep learning models in recognizing, localizing, describing, and categorizing meniscal tears in magnetic resonance images (MRIs). MATERIALS AND METHODS This systematic review was rigorously conducted, strictly following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. Extensive searches were conducted on MEDLINE (PubMed), Web of Science, Cochrane Library, and Google Scholar. All identified articles underwent a comprehensive risk of bias analysis. Predictive performance values were either extracted or calculated for quantitative analysis, including sensitivity and specificity. The meta-analysis was performed for all prediction models that identified the presence and location of meniscus tears. RESULTS This study's findings underscore that a range of deep learning models exhibit robust performance in detecting and classifying meniscal tears, in one case surpassing the expertise of musculoskeletal radiologists. Most studies in this review concentrated on identifying tears in the medial or lateral meniscus and even precisely locating tears-whether in the anterior or posterior horn-with exceptional accuracy, as demonstrated by AUC values ranging from 0.83 to 0.94. CONCLUSIONS Based on these findings, deep learning models have showcased significant potential in analyzing knee MR images by learning intricate details within images. They offer precise outcomes across diverse tasks, including segmenting specific anatomical structures and identifying pathological regions. Contributions: This study focused exclusively on DL models for identifying and localizing meniscus tears. It presents a meta-analysis that includes eight studies for detecting the presence of a torn meniscus and a meta-analysis of three studies with low heterogeneity that localize and classify the menisci. Another novelty is the analysis of arthroscopic surgery as ground truth. The quality of the studies was assessed against the CLAIM checklist, and the risk of bias was determined using the QUADAS-2 tool.
Collapse
Affiliation(s)
- Alexei Botnari
- Department of Orthopedics, Faculty of Medicine, “Victor Babes” University of Medicine and Pharmacy, 300041 Timisoara, Romania
| | - Manuella Kadar
- Department of Computer Science, Faculty of Informatics and Engineering, “1 Decembrie 1918” University of Alba Iulia, 510009 Alba Iulia, Romania
| | - Jenel Marian Patrascu
- Department of Orthopedics-Traumatology, Faculty of Medicine, “Victor Babes” University of Medicine and Pharmacy, 300041 Timisoara, Romania;
| |
Collapse
|
50
|
Jiao Z, Liang Z, Liao Q, Chen S, Yang H, Hong G, Gui H. Deep learning for automatic detection of cephalometric landmarks on lateral cephalometric radiographs using the Mask Region-based Convolutional Neural Network: a pilot study. Oral Surg Oral Med Oral Pathol Oral Radiol 2024; 137:554-562. [PMID: 38480069 DOI: 10.1016/j.oooo.2024.02.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/10/2023] [Revised: 01/24/2024] [Accepted: 02/04/2024] [Indexed: 06/20/2024]
Abstract
OBJECTIVE We examined the effectiveness and feasibility of the Mask Region-based Convolutional Neural Network (Mask R-CNN) for automatic detection of cephalometric landmarks on lateral cephalometric radiographs (LCRs). STUDY DESIGN In total, 400 LCRs, each with 19 manually identified landmarks, were collected. Of this total, 320 images were randomly selected as the training dataset for Mask R-CNN, and the remaining 80 images were used for testing the automatic detection of the 19 cephalometric landmarks, for a total of 1520 landmarks. Detection rate, average error, and detection accuracy rate were calculated to assess Mask R-CNN performance. RESULTS Of the 1520 landmarks, 1494 were detected, for a detection rate of 98.29%. The average error, or linear deviation distance between the detected points and the originally marked points of each detected landmark, ranged from 0.56 to 9.51 mm, with an average of 2.19 mm. For detection accuracy rate, 649 landmarks (43.44%) had a linear deviation distance less than 1 mm, 1020 (68.27%) less than 2 mm, and 1281 (85.74%) less than 4 mm in deviation from the manually marked point. The average detection time was 1.48 seconds per image. CONCLUSIONS Deep learning Mask R-CNN shows promise in enhancing cephalometric analysis by automating landmark detection on LCRs, addressing the limitations of manual analysis, and demonstrating effectiveness and feasibility.
Collapse
Affiliation(s)
- Zhentao Jiao
- Department of Oral and Maxillofacial Surgery, Dalian Stomatological Hospital, Dalian, China; Division for Globalization Initiative, Liaison Center for Innovative Dentistry, Graduate School of Dentistry, Tohoku University, Sendai, Japan
| | - Zhuangzhuang Liang
- School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Qian Liao
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China
| | - Sheng Chen
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China
| | - Hui Yang
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China
| | - Guang Hong
- Division for Globalization Initiative, Liaison Center for Innovative Dentistry, Graduate School of Dentistry, Tohoku University, Sendai, Japan.
| | - Haijun Gui
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai, China.
| |
Collapse
|