1
|
Gan M, Xie W, Tan X, Wang W. Coronary artery segmentation framework based on three types of U-Net and voting ensembles. Health Inf Sci Syst 2025; 13:6. [PMID: 39687033 PMCID: PMC11646233 DOI: 10.1007/s13755-024-00322-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Accepted: 11/23/2024] [Indexed: 12/18/2024] Open
Abstract
Coronary artery (CA) segmentation is critical for enabling disease diagnosis. However, the structural complexity and extensive branching of CAs may cause the segmentation outcomes of existing methods to exhibit discontinuities and considerable pseudo-CA regions. Therefore, we propose a voting-based ensemble segmentation framework based on three U-Net types to capture CA structural features from global and local perspectives. The lightweight U-Net performs direct segmentation on CAs, helping to eliminate interferences from small connected regions during segmentation and preserve global information. Patch-based and multi-slice U-Nets provide superior local partition information. Finally, a voting-based strategy is adopted to ensemble the segmentation results for the three models to obtain the final result. Our proposed segmentation framework performs well, attaining a Dice score of 82.31% on a large dataset.
Collapse
Affiliation(s)
- Mengkun Gan
- Information and Data Center, Guangzhou First People’s Hospital, Guangzhou Medical University, Guangzhou, 510180 China
- Information and Data Center The Second Affiliated Hospital School of Medicine, South China University of Technology, Guangzhou, 510180 China
| | - Weijie Xie
- Information and Data Center, Guangzhou First People’s Hospital, Guangzhou Medical University, Guangzhou, 510180 China
- Information and Data Center The Second Affiliated Hospital School of Medicine, South China University of Technology, Guangzhou, 510180 China
| | - Xiaocong Tan
- Information and Data Center, Guangzhou First People’s Hospital, Guangzhou Medical University, Guangzhou, 510180 China
- Information and Data Center The Second Affiliated Hospital School of Medicine, South China University of Technology, Guangzhou, 510180 China
| | - Wenhui Wang
- Information and Data Center, Guangzhou First People’s Hospital, Guangzhou Medical University, Guangzhou, 510180 China
- Information and Data Center The Second Affiliated Hospital School of Medicine, South China University of Technology, Guangzhou, 510180 China
| |
Collapse
|
2
|
Kunekar P, Dadheech P, Gupta MK. ECn-MultiBSTM: multiclass epileptic seizure classification using electro cetacean optimized bidirectional long short-term memory model. Cogn Neurodyn 2025; 19:83. [PMID: 40443913 PMCID: PMC12116414 DOI: 10.1007/s11571-025-10268-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 03/13/2025] [Accepted: 04/24/2025] [Indexed: 06/02/2025] Open
Abstract
Multiclass epileptic seizureclassification aims to identify and categorize different epileptic seizure types like a non-epileptic seizure, epileptic interictal seizure, and epileptic ictal seizurein individuals based on Electroencephalography (EEG) signal characteristics. Multi-class seizure classification requires recognizing various seizure forms and patterns, which can be challenging due to noise and high variability patterns in EEG signals. Existing models face limitations such as difficulty in handling the complex and dynamic nature of seizure patterns, poor generalization to unseen data, and sensitivity to noise and artifacts, all of which impact classification accuracy and reliability. To address these issues, the Electro Cetacean Optimization based Multi Bidirectional Long Short-Term Memory (ECn-MultiBSTM) model is proposed. The BiLSTM modelis utilized for feature extraction, which captures sequential data by processing data in both forward and backward directions. This bidirectional approach enables the model to identify subtle patterns that distinguish various seizure types with higher accuracy. The ECn-MultiBSTM model also incorporates advanced Electro Cetacean optimizationtechniques that enhance its ability to search efficiently for optimal solutions.Through dynamic social coordination and rapid search strategies, the model fine-tunes its hyperparameters, ensuring improved performance and adaptability.The proposed ECn-MultiBSTM model significantly enhances multiclassseizure classification performance, achieving impressive metrics of 95.84% accuracy, 95.30% precision, 95.54% F1-score,0.94% MCC, 95.79% sensitivity, and 95.88% specificity when evaluated on the CHB-MIT SCALP EEG dataset.
Collapse
Affiliation(s)
- Pankaj Kunekar
- Department of Computer Science & Engineering, Swami Keshvanand Institute of Technology, Management & Gramothan (SKIT), Ramnagaria, Jagatpura, Jaipur, Rajasthan 302017 India
| | - Pankaj Dadheech
- Department of Computer Science & Engineering, Swami Keshvanand Institute of Technology, Management & Gramothan (SKIT), Ramnagaria, Jagatpura, Jaipur, Rajasthan 302017 India
| | - Mukesh Kumar Gupta
- Digital Data Governance Group, National Informatics Centre (NIC), A-Block, CGO Complex, Lodhi Road, New Delhi, 110003 India
| |
Collapse
|
3
|
Agrawal KK, Kumar G. LiDSCUNet++: A lightweight depth separable convolutional UNet++ for vertebral column segmentation and spondylosis detection. Res Vet Sci 2025; 192:105703. [PMID: 40460622 DOI: 10.1016/j.rvsc.2025.105703] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2025] [Revised: 05/10/2025] [Accepted: 05/17/2025] [Indexed: 06/11/2025]
Abstract
Accurate computer-aided diagnosis systems rely on precise segmentation of the vertebral column to assist physicians in diagnosing various disorders. However, segmenting spinal disks and bones becomes challenging in the presence of abnormalities and complex anatomical structures. While Deep Convolutional Neural Networks (DCNNs) achieve remarkable results in medical image segmentation, their performance is limited by data insufficiency and the high computational complexity of existing solutions. This paper introduces LiDSCUNet++, a lightweight deep learning framework based on depthwise-separable and pointwise convolutions integrated with UNet++ for vertebral column segmentation. The model segments vertebral anomalies from dog radiographs, and the results are further processed by YOLOv8 for automated detection of Spondylosis Deformans. LiDSCUNet++ delivers comparable segmentation performance while significantly reducing trainable parameters, memory usage, energy consumption, and computational time, making it an efficient and practical solution for medical image analysis.
Collapse
Affiliation(s)
- Krishna K Agrawal
- Department of Computer Science & Engineering, National Institute of Technology Delhi, Plot No. FA7, Zone P1, GT Karnal Road, Delhi-110036, India.
| | - Gautam Kumar
- Department of Computer Science & Engineering, National Institute of Technology Delhi, Plot No. FA7, Zone P1, GT Karnal Road, Delhi-110036, India.
| |
Collapse
|
4
|
Zhu N, Niu F, Fan S, Meng X, Hu Y, Han J, Wang Z. Predicting progression-free survival in sarcoma using MRI-based automatic segmentation models and radiomics nomograms: a preliminary multicenter study. Skeletal Radiol 2025; 54:1417-1427. [PMID: 39630238 DOI: 10.1007/s00256-024-04837-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/01/2024] [Revised: 11/17/2024] [Accepted: 11/18/2024] [Indexed: 05/16/2025]
Abstract
OBJECTIVES Some sarcomas are highly malignant, associated with high recurrence despite treatment. This multicenter study aimed to develop and validate a radiomics signature to estimate sarcoma progression-free survival (PFS). MATERIALS AND METHODS The study retrospectively enrolled 202 consecutive patients with pathologically diagnosed sarcoma, who had pre-treatment axial fat-suppressed T2-weighted images (FS-T2WI), and included them in the ROI-Net model for training. Among them, 120 patients were included in the radiomics analysis, all of whom had pre-treatment axial T1-weighted and transverse FS-T2WI images, and were randomly divided into a development group (n = 96) and a validation group (n = 24). In the development cohort, Least Absolute Shrinkage and Selection Operator (LASSO) Cox regression was used to develop the radiomics features for PFS prediction. By combining significant clinical features with radiomics features, a nomogram was constructed using Cox regression. RESULTS The proposed ROI-Net framework achieved a Dice coefficient of 0.820 (0.791-0.848). The radiomics signature based on 21 features could distinguish high-risk patients with poor PFS. Univariate Cox analysis revealed that peritumoral edema, metastases, and the radiomics score were associated with poor PFS and were included in the construction of the nomogram. The Radiomics-T1WI-Clinical model exhibited the best performance, with AUC values of 0.947, 0.907, and 0.924 at 300 days, 600 days, and 900 days, respectively. CONCLUSION The proposed ROI-Net framework demonstrated high consistency between its segmentation results and expert annotations. The radiomics features and the combined nomogram have the potential to aid in predicting PFS for patients with sarcoma.
Collapse
Affiliation(s)
- Nana Zhu
- Graduate School, Tianjin Medical University, Tianjin, China
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, 300211, China
| | - Feige Niu
- Graduate School, Tianjin Medical University, Tianjin, China
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, 300211, China
| | - Shuxuan Fan
- The Department of Radiology, Tianjin Medical University Cancer Hospital, Tianjin, China
| | - Xianghong Meng
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, 300211, China
| | - Yongcheng Hu
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, 300211, China
| | - Jun Han
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, 300211, China
- Graduate School, Tianjin University, Tianjin, China
| | - Zhi Wang
- The Department of Radiology, Tianjin University Tianjin Hospital, 406 Jiefang Southern Road, Tianjin, 300211, China.
| |
Collapse
|
5
|
Deniz HA, Bayrakdar İŞ, Nalçacı R, Orhan K. Segmentation of the nasopalatine canal and detection of canal furcation status with artificial intelligence on cone-beam computed tomography images. Oral Radiol 2025; 41:403-413. [PMID: 40021578 DOI: 10.1007/s11282-025-00812-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2024] [Accepted: 02/12/2025] [Indexed: 03/03/2025]
Abstract
OBJECTIVES The nasopalatine canal (NPC) is an anatomical formation with varying morphology. NPC can be visualized using the cone-beam computed tomography (CBCT). Also, CBCT has been used in many studies on artificial intelligence (AI). The "You only look once" (YOLO) is an AI framework that stands out with its speed. This study compared the observer and AI regarding the NPC segmentation and assessment of the NPC furcation status in CBCT images. METHODS In this study, axial sections of 200 CBCT images were used. These images were labeled and evaluated for the absence or presence of the NPC furcation. These images were then divided into three; 160 images were used as the training dataset, 20 as the validation dataset, and 20 as the test dataset. The training was performed by making 800 epochs using the YOLOv5x-seg model. RESULTS Sensitivity, Precision, F1 score, IoU, mAP, and AUC values were determined for NPC detection, segmentation, and classification of the YOLOv5x-seg model. The values were found to be 0.9680, 0.9953, 0.9815, 0.9636, 0.7930, and 0.8841, respectively, for the group with the absence of the NPC furcation; and 0.9827, 0.9975, 0.9900, 0.9803, 0.9637, and 0.9510, for the group with the presence of the NPC furcation. CONCLUSIONS Our results showed that even when the YOLOv5x-seg model is trained with the NPC furcation and fewer datasets, it achieves sufficient prediction accuracy. The segmentation feature of the YOLOv5 algorithm, which is based on an object detection algorithm, has achieved quite successful results despite its recent development.
Collapse
Affiliation(s)
- Hatice Ahsen Deniz
- Department of Oral and Maxillofacial Radiology, Aksaray University Faculty of Dentistry, Aksaray, Turkey.
| | - İbrahim Şevki Bayrakdar
- Department of Oral and Maxillofacial Radiology, Osmangazi University Faculty of Dentistry, Eskişehir, Turkey
| | - Rana Nalçacı
- Department of Oral and Maxillofacial Radiology, Ankara University Faculty of Dentistry, Ankara, Turkey
| | - Kaan Orhan
- Department of Oral and Maxillofacial Radiology, Ankara University Faculty of Dentistry, Ankara, Turkey
| |
Collapse
|
6
|
Zbinden L, Erb S, Catucci D, Doorenbos L, Hulbert L, Berzigotti A, Brönimann M, Ebner L, Christe A, Obmann VC, Sznitman R, Huber AT. Automated quantification of T1 and T2 relaxation times in liver mpMRI using deep learning: a sequence-adaptive approach. Eur Radiol Exp 2025; 9:58. [PMID: 40515936 DOI: 10.1186/s41747-025-00596-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2024] [Accepted: 05/19/2025] [Indexed: 06/16/2025] Open
Abstract
OBJECTIVES To evaluate a deep learning sequence-adaptive liver multiparametric MRI (mpMRI) assessment with validation in different populations using total and segmental T1 and T2 relaxation time maps. METHODS A neural network was trained to label liver segmental parenchyma and its vessels on noncontrast T1-weighted gradient-echo Dixon in-phase acquisitions on 200 liver mpMRI examinations. Then, 120 unseen liver mpMRI examinations of patients with primary sclerosing cholangitis or healthy controls were assessed by coregistering the labels to noncontrast and contrast-enhanced T1 and T2 relaxation time maps for optimization and internal testing. The algorithm was externally tested in a segmental and total liver analysis of previously unseen 65 patients with biopsy-proven liver fibrosis and 25 healthy volunteers. Measured relaxation times were compared to manual measurements using intraclass correlation coefficient (ICC) and Wilcoxon test. RESULTS Comparison of manual and deep learning-generated segmental areas on different T1 and T2 maps was excellent for segmental (ICC = 0.95 ± 0.1; p < 0.001) and total liver assessment (0.97 ± 0.02, p < 0.001). The resulting median of the differences between automated and manual measurements among all testing populations and liver segments was 1.8 ms for noncontrast T1 (median 835 versus 842 ms), 2.0 ms for contrast-enhanced T1 (median 518 versus 519 ms), and 0.3 ms for T2 (median 37 versus 37 ms). CONCLUSION Automated quantification of liver mpMRI is highly effective across different patient populations, offering excellent reliability for total and segmental T1 and T2 maps. Its scalable, sequence-adaptive design could foster comprehensive clinical decision-making. RELEVANCE STATEMENT The proposed automated, sequence-adaptive algorithm for total and segmental analysis of liver mpMRI may be co-registered to any combination of parametric sequences, enabling comprehensive quantitative analysis of liver mpMRI without sequence-specific training. KEY POINTS A deep learning-based algorithm automatically quantified segmental T1 and T2 relaxation times in liver mpMRI. The two-step approach of segmentation and co-registration allowed to assess arbitrary sequences. The algorithm demonstrated high reliability with manual reader quantification. No additional sequence-specific training is required to assess other parametric sequences. The DL algorithm has the potential to enhance individual liver phenotyping.
Collapse
Affiliation(s)
- Lukas Zbinden
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Samuel Erb
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Damiano Catucci
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
- Graduate School for Health Sciences, University of Bern, Bern, Switzerland
| | - Lars Doorenbos
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
| | - Leona Hulbert
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Annalisa Berzigotti
- Hepatology, Department of Visceral Surgery and Medicine, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Michael Brönimann
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Lukas Ebner
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
- Department of Radiology and Nuclear Medicine, Luzerner Kantonsspital, University Teaching and Research Hospital, University of Lucerne, Lucerne, Switzerland
| | - Andreas Christe
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Verena Carola Obmann
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland
| | - Raphael Sznitman
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
| | - Adrian Thomas Huber
- Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, Bern, Switzerland.
- Department of Radiology and Nuclear Medicine, Luzerner Kantonsspital, University Teaching and Research Hospital, University of Lucerne, Lucerne, Switzerland.
- Department of Radiology, Beau-Site, Hirslanden, Bern, Switzerland.
| |
Collapse
|
7
|
Kong X, Zhang A, Zhou X, Zhao M, Liu J, Zhang X, Zhang W, Meng X, Li N, Yang Z. A strategy for the automatic diagnostic pipeline towards feature-based models: a primer with pleural invasion prediction from preoperative PET/CT images. EJNMMI Res 2025; 15:70. [PMID: 40506669 PMCID: PMC12162436 DOI: 10.1186/s13550-025-01264-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2025] [Accepted: 05/23/2025] [Indexed: 06/16/2025] Open
Abstract
BACKGROUND This study aims to explore the feasibility to automate the application process of nomograms in clinical medicine, demonstrated through the task of preoperative pleural invasion prediction in non-small cell lung cancer patients using PET/CT imaging. RESULTS The automatic pipeline involves multimodal segmentation, feature extraction, and model prediction. It is validated on a cohort of 1116 patients from two medical centers. The performance of the feature-based diagnostic model outperformed both the radiomics model and individual machine learning models. The segmentation models for CT and PET images achieved mean dice similarity coefficients of 0.85 and 0.89, respectively, and the segmented lung contours showed high consistency with the actual contours. The automatic diagnostic system achieved an accuracy of 0.87 in the internal test set and 0.82 in the external test set, demonstrating comparable overall diagnostic performance to the human-based diagnostic model. In comparative analysis, the automatic diagnostic system showed superior performance relative to other segmentation and diagnostic pipelines. CONCLUSIONS The proposed automatic diagnostic system provides an interpretable, automated solution for predicting pleural invasion in non-small cell lung cancer.
Collapse
Affiliation(s)
- Xiangxing Kong
- Institution of Medical Technology, Peking University Health Science Center, Beijing, 100191, China
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Beijing Key Laboratory of Research, Investigation and Evaluation of Radiopharmaceuticals, NMPA Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Peking University Cancer Hospital and Institute, No. 52 Fucheng Rd., Haidian District, Beijing, 100142, China
| | - Annan Zhang
- Department of Nuclear Medicine, Peking University Third Hospital, Beijing, 100191, China
| | - Xin Zhou
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Beijing Key Laboratory of Research, Investigation and Evaluation of Radiopharmaceuticals, NMPA Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Peking University Cancer Hospital and Institute, No. 52 Fucheng Rd., Haidian District, Beijing, 100142, China
| | - Meixin Zhao
- Department of Nuclear Medicine, Peking University Third Hospital, Beijing, 100191, China
| | - Jiayue Liu
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Beijing Key Laboratory of Research, Investigation and Evaluation of Radiopharmaceuticals, NMPA Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Peking University Cancer Hospital and Institute, No. 52 Fucheng Rd., Haidian District, Beijing, 100142, China
| | - Xinliang Zhang
- Institution of Medical Technology, Peking University Health Science Center, Beijing, 100191, China
| | - Weifang Zhang
- Department of Nuclear Medicine, Peking University Third Hospital, Beijing, 100191, China
| | - Xiangxi Meng
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Beijing Key Laboratory of Research, Investigation and Evaluation of Radiopharmaceuticals, NMPA Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Peking University Cancer Hospital and Institute, No. 52 Fucheng Rd., Haidian District, Beijing, 100142, China.
| | - Nan Li
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Beijing Key Laboratory of Research, Investigation and Evaluation of Radiopharmaceuticals, NMPA Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Peking University Cancer Hospital and Institute, No. 52 Fucheng Rd., Haidian District, Beijing, 100142, China.
| | - Zhi Yang
- Institution of Medical Technology, Peking University Health Science Center, Beijing, 100191, China.
- State Key Laboratory of Holistic Integrative Management of Gastrointestinal Cancers, Beijing Key Laboratory of Research, Investigation and Evaluation of Radiopharmaceuticals, NMPA Key Laboratory for Research and Evaluation of Radiopharmaceuticals (National Medical Products Administration), Department of Nuclear Medicine, Peking University Cancer Hospital and Institute, No. 52 Fucheng Rd., Haidian District, Beijing, 100142, China.
| |
Collapse
|
8
|
Siyasari A, Qalanou MS, Dehghanmehr S. Commentary on "Photodiagnosis with deep learning: A GAN and autoencoder-based approach for diabetic retinopathy detection" by Gencer et al., 2025. Photodiagnosis Photodyn Ther 2025:104666. [PMID: 40516582 DOI: 10.1016/j.pdpdt.2025.104666] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2025] [Revised: 05/29/2025] [Accepted: 06/12/2025] [Indexed: 06/16/2025]
Affiliation(s)
- Ahmadreza Siyasari
- Master of Science in Medical Surgical Nursing, Zabol University of Medical Sciences, Zabol, Iran
| | - Motahareh Sabaghi Qalanou
- Paramedical Faculty of Iranshahr, Iranshahr University of Medical Sciences Iranshahr, Iran; Tropical and Communicable Diseases Research Center, Iranshahr University of Medical Sciences Iranshahr, Iran
| | - Sadegh Dehghanmehr
- Tropical and Communicable Diseases Research Center, Iranshahr University of Medical Sciences Iranshahr, Iran; Instructor of Medical Surgical Nursing, Department of Nursing, School of Nursing and Midwifery, Iranshahr University of Medical Sciences, Iranshahr, Iran.
| |
Collapse
|
9
|
Yousefzamani M, Babapour Mofrad F. Deep learning without borders: recent advances in ultrasound image classification for liver diseases diagnosis. Expert Rev Med Devices 2025:1-17. [PMID: 40445166 DOI: 10.1080/17434440.2025.2514764] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2024] [Accepted: 05/29/2025] [Indexed: 06/11/2025]
Abstract
INTRODUCTION Liver diseases are among the top global health burdens. Recently, there has been an increasing significance of diagnostics without discomfort to the patient; among them, ultrasound is the most used. Deep learning, in particular convolutional neural networks, has revolutionized the classification of liver diseases by automatically performing some specific analyses of difficult images. AREAS COVERED This review summarizes the progress that has been made in deep learning techniques for the classification of liver diseases using ultrasound imaging. It evaluates various models from CNNs to their hybrid versions, such as CNN-Transformer, for detecting fatty liver, fibrosis, and liver cancer, among others. Several challenges in the generalization of data and models across a different clinical environment are also discussed. EXPERT OPINION Deep learning has great prospects for automatic diagnosis of liver diseases. Most of the models have performed with high accuracy in different clinical studies. Despite this promise, challenges relating to generalization have remained. Future hardware developments and access to quality clinical data continue to further improve the performance of these models and ensure their vital role in the diagnosis of liver diseases.
Collapse
Affiliation(s)
- Midya Yousefzamani
- Department of Medical Radiation Engineering SR.C., Islamic Azad University, Tehran, Iran
| | | |
Collapse
|
10
|
Lal R, Singh RK, Nishad DK, Khalid S. Quantum-enhanced intelligent system for personalized adaptive radiotherapy dose estimation. Sci Rep 2025; 15:19919. [PMID: 40481168 PMCID: PMC12144112 DOI: 10.1038/s41598-025-05673-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2025] [Accepted: 06/03/2025] [Indexed: 06/11/2025] Open
Abstract
This research introduces a novel quantum-enhanced intelligent system tailored for personalized adaptive radiotherapy dose estimation. The system efficiently models radiation transport and predicts patient-specific dose distributions by integrating quantum algorithms, deep learning, and Monte Carlo simulations. Quantum-enhanced Monte Carlo simulations, employing algorithms such as Harrow-Hassidim-Lloyd (HHL) and Variational Quantum Eigensolver (VQE), achieve computational speedups of 8-15 times compared to classical methods while maintaining high accuracy. The deep learning architecture leverages convolutional and recurrent neural networks to capture complex anatomical and dosimetric patterns. Validation on simulated datasets demonstrates a 50-70% reduction in mean absolute error and 2-3% improvements in gamma index metrics compared to conventional approaches. Dose-volume histogram analysis further highlights enhanced Dice coefficients and reduced Hausdorff distances. These advancements underscore the potential for precise, efficient, and clinically relevant dose estimations, paving the way for improved outcomes in personalized adaptive radiotherapy.
Collapse
Affiliation(s)
- Radhey Lal
- Dr. APJ Abdul, Kalam Technical University Lucknow, 226021, Lucknow, India
| | | | - Dinesh Kumar Nishad
- Department of Electrical Engineering, Dr. Shakuntala Misra National Rehabilitation University, Lucknow, India
| | | |
Collapse
|
11
|
Robinson D, Khatib M, Eissa M, Yassin M. Potential Use of a New Energy Vision (NEV) Camera for Diagnostic Support of Carpal Tunnel Syndrome: Development of a Decision-Making Algorithm to Differentiate Carpal Tunnel-Affected Hands from Controls. Diagnostics (Basel) 2025; 15:1417. [PMID: 40506989 PMCID: PMC12154292 DOI: 10.3390/diagnostics15111417] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2025] [Revised: 05/19/2025] [Accepted: 05/28/2025] [Indexed: 06/16/2025] Open
Abstract
Introduction: Carpal Tunnel Syndrome (CTS) is a prevalent neuropathy requiring accurate, non-invasive diagnostics to minimize patient burden. This study evaluates the New Energy Vision (NEV) camera, an RGB-based multispectral imaging tool, to detect CTS through skin texture and color analysis, developing a machine learning algorithm to distinguish CTS-affected hands from controls. Methods: A two-part observational study included 103 participants (50 controls, 53 CTS patients) in Part 1, using NEV camera images to train a Support Vector Machine (SVM) classifier. Part 2 compared median nerve-damaged (MED) and ulnar nerve-normal (ULN) palm areas in 32 CTS patients. Validations included nerve conduction tests (NCT), Semmes-Weinstein monofilament testing (SWMT), and Boston Carpal Tunnel Questionnaire (BCTQ). Results: The SVM classifier achieved 93.33% accuracy (confusion matrix: [[14, 1], [1, 14]]), with 81.79% cross-validation accuracy. Part 2 identified significant differences (p < 0.05) in color proportions (e.g., red_proportion) and Haralick texture features between MED and ULN areas, corroborated by BCTQ and SWMT. Conclusions: The NEV camera, leveraging multispectral imaging, offers a promising non-invasive CTS diagnostic tool using detection of nerve-related skin changes. Further validation is needed for clinical adoption.
Collapse
Affiliation(s)
- Dror Robinson
- Department of Orthopedics, Hasharon Hospital, Rabin Medical Center, Affiliated to Tel Aviv University, Tel Aviv 6997801, Israel; (M.K.); (M.E.); (M.Y.)
| | | | | | | |
Collapse
|
12
|
Zsarnoczay E, Rapaka S, Schoepf UJ, Gnasso C, Vecsey-Nagy M, Todoran TM, Hagar MT, Kravchenko D, Tremamunno G, Griffith JP, Fink N, Derrick S, Bowman M, Sam H, Tiller M, Godoy K, Condrea F, Sharma P, O'Doherty J, Maurovich-Horvat P, Emrich T, Varga-Szemes A. Accuracy of a deep neural network for automated pulmonary embolism detection on dedicated CT pulmonary angiograms. Eur J Radiol 2025; 187:112077. [PMID: 40187199 DOI: 10.1016/j.ejrad.2025.112077] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2024] [Revised: 03/11/2025] [Accepted: 03/26/2025] [Indexed: 04/07/2025]
Abstract
PURPOSE To assess the performance of a Deep Neural Network (DNN)-based prototype algorithm for automated PE detection on CTPA scans. METHODS Patients who had previously undergone CTPA with three different systems (SOMATOM Force, go.Top, and Definition AS; Siemens Healthineers, Forchheim, Germany) because of suspected PE from September 2022 to January 2023 were retrospectively enrolled in this study (n = 1,000, 58.8 % women). For detailed evaluation, all PE were divided into three location-based subgroups: central arteries, lobar branches, and peripheral regions. Clinical reports served as ground truth. Sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and accuracy were determined to evaluate the performance of DNN-based PE detection. RESULTS Cases were excluded due to incomplete data (n = 32), inconclusive report (n = 17), insufficient contrast detected in the pulmonary trunk (n = 40), or failure of the preprocessing algorithms (n = 8). Therefore, the final cohort included 903 cases with a PE prevalence of 12 % (n = 110). The model achieved a sensitivity, specificity, PPV, and NPV of 84.6, 95.1, 70.5, and 97.8 %, respectively, and delivered an overall accuracy of 93.8 %. Among the false positive cases (n = 39), common sources of error included lung masses, pneumonia, and contrast flow artifacts. Common sources of false negatives (n = 17) included chronic and subsegmental PEs. CONCLUSION The proposed DNN-based algorithm provides excellent performance for the detection of PE, suggesting its potential utility to support radiologists in clinical reading and exam prioritization.
Collapse
Affiliation(s)
- Emese Zsarnoczay
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Radiology Department, Medical Imaging Centre, Semmelweis University, Korányi Sándor utca 2, 1083 Budapest, Hungary.
| | - Saikiran Rapaka
- Siemens Healthineers, 755 College Rd E, Princeton, NJ 08540, USA.
| | - U Joseph Schoepf
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Chiara Gnasso
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Experimental Imaging Center, IRCCS San Raffaele Scientific Institute, Via Olgettina, 60, 20132 Milan, Italy.
| | - Milan Vecsey-Nagy
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Heart and Vascular Center, Semmelweis University, Gaál József út 9, 1122 Budapest, Hungary.
| | - Thomas M Todoran
- Division of Cardiology, Department of Medicine, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Muhammad Taha Hagar
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Department of Diagnostic and Interventional Radiology, Medical Center, Faculty of Medicine, University of Freiburg, Hugstetter Straße 55, Freiburg im Breisgau 79106, Germany.
| | - Dmitrij Kravchenko
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Department of Diagnostic and Interventional Radiology, University Hospital Bonn, Venusberg-Campus 1, 53127 Bonn, Germany.
| | - Giuseppe Tremamunno
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Department of Medical Surgical Sciences and Translational Medicine, Sapienza University of Rome - Radiology Unit - Sant'Andrea University Hospital, Via di Grottarossa, 1035-1039, 00189 Rome, Italy.
| | - Joseph Parkwood Griffith
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Nicola Fink
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Department of Radiology, University Hospital, LMU Munich, Marchioninistraße 15, 81377 Munich, Germany.
| | - Sydney Derrick
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Meredith Bowman
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Henry Sam
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Mikayla Tiller
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Kathleen Godoy
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| | - Florin Condrea
- Siemens Healthineers, Nine, Bulevardul Gării 13A, Brașov 500227, Romania.
| | - Puneet Sharma
- Siemens Healthineers, 755 College Rd E, Princeton, NJ 08540, USA.
| | - Jim O'Doherty
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Siemens Medical Solutions, 40 Liberty Blvd, Malvern, PA 19355, USA.
| | - Pal Maurovich-Horvat
- Radiology Department, Medical Imaging Centre, Semmelweis University, Korányi Sándor utca 2, 1083 Budapest, Hungary.
| | - Tilman Emrich
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA; Department of Diagnostic and Interventional Radiology, University Medical Center of the Johannes Gutenberg-University, Langenbeckstraße 1, Mainz 55131, Germany; German Centre for Cardiovascular Research, Mainz, Germany.
| | - Akos Varga-Szemes
- Department of Radiology and Radiological Science, Medical University of South Carolina, 25 Courtenay Drive, Charleston, SC 29425, USA.
| |
Collapse
|
13
|
Kulathilake CD, Udupihille J, Abeysundara SP, Senoo A. Deep learning-driven multi-class classification of brain strokes using computed tomography: A step towards enhanced diagnostic precision. Eur J Radiol 2025; 187:112109. [PMID: 40252282 DOI: 10.1016/j.ejrad.2025.112109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2025] [Revised: 04/08/2025] [Accepted: 04/09/2025] [Indexed: 04/21/2025]
Abstract
OBJECTIVE To develop and validate deep learning models leveraging CT imaging for the prediction and classification of brain stroke conditions, with the potential to enhance accuracy and support clinical decision-making. MATERIALS AND METHODS This retrospective, bi-center study included data from 250 patients, with a dataset of 8186 CT images collected from 2017 to 2022. Two AI models were developed using the Expanded ResNet101 deep learning framework as a two-step model. Model performance was evaluated using confusion matrices, supplemented by external validation with an independent dataset. External validation was conducted by an expert and two external members. Overall accuracy, confidence intervals, Cohen's Kappa value, and McNemar's test P-values were calculated. RESULTS A total of 8186 CT images were incorporated, with 6386 images used for the training and 900 datasets for testing and validation in Model 01. Further, 1619 CT images were used for training and 600 datasets for testing and validation in Model 02. The average accuracy, precision, and F1 score for both models were assessed: Model 01 achieved 99.6 %, 99.4 %, and 99.6 % respectively, whereas Model 02 achieved 99.2 %, 98.8 %, and 99.1 %. The external validation accuracies were 78.6 % (95 % CI: 0.73,0.83; P < 0.001) and 60.2 % (95 % CI: 0.48,0.70; P < 0.001) for Models 01 and 02 respectively, as evaluated by the expert. CONCLUSION Deep learning models demonstrated high accuracy, precision, and F1 scores in predicting outcomes for brain stroke patients. With larger cohort and diverse radiologic mimics, these models could support clinicians in prognosis and decision-making.
Collapse
Affiliation(s)
- Chathura D Kulathilake
- Department of Radiological Sciences, School of Human Health Sciences, Tokyo Metropolitan University, Japan
| | - Jeevani Udupihille
- Department of Radiology, Faculty of Medicine, University of Peradeniya, Sri Lanka
| | - Sachith P Abeysundara
- Department of Statistics and Computer Science, Faculty of Science, University of Peradeniya, Sri Lanka
| | - Atsushi Senoo
- Department of Radiological Sciences, School of Human Health Sciences, Tokyo Metropolitan University, Japan.
| |
Collapse
|
14
|
Khalil RU, Sajjad M, Dhahbi S, Bourouis S, Hijji M, Muhammad K. Mitosis detection and classification for breast cancer diagnosis: What we know and what is next. Comput Biol Med 2025; 191:110057. [PMID: 40209577 DOI: 10.1016/j.compbiomed.2025.110057] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Revised: 02/22/2025] [Accepted: 03/18/2025] [Indexed: 04/12/2025]
Abstract
Breast cancer is the second most deadly malignancy in women, behind lung cancer. Despite significant improvements in medical research, breast cancer is still accurately diagnosed with histological analysis. During this procedure, pathologists examine a physical sample for the presence of mitotic cells, or dividing cells. However, the high resolution of histopathology images and the difficulty of manually detecting tiny mitotic nuclei make it particularly challenging to differentiate mitotic cells from other types of cells. Numerous studies have addressed the detection and classification of mitosis, owing to increasing capacity and developments in automated approaches. The combination of machine learning and deep learning techniques has greatly revolutionized the process of identifying mitotic cells by offering automated, precise, and efficient solutions. In the last ten years, several pioneering methods have been presented, advancing towards practical applications in clinical settings. Unlike other forms of cancer, breast cancer and gliomas are categorized according to the number of mitotic divisions. Numerous papers have been published on techniques for identifying mitosis due to easy access to datasets and open competitions. Convolutional neural networks and other deep learning architectures can precisely identify mitotic cells, significantly decreasing the amount of labor that pathologists must perform. This article examines the techniques used over the past decade to identify and classify mitotic cells in histologically stained breast cancer hematoxylin and eosin images. Furthermore, we examine the benefits of current research techniques and predict forthcoming developments in the investigation of breast cancer mitosis, specifically highlighting machine learning and deep learning.
Collapse
Affiliation(s)
- Rafi Ullah Khalil
- Digital Image Processing Lab, Department of Computer Science, Islamia College Peshawar, Peshawar, 25000, Pakistan.
| | - Muhammad Sajjad
- Digital Image Processing Lab, Department of Computer Science, Islamia College Peshawar, Peshawar, 25000, Pakistan.
| | - Sami Dhahbi
- Applied College of Mahail Aseer, King Khalid University, Muhayil Aseer, 62529, Saudi Arabia.
| | - Sami Bourouis
- Department of Information Technology, College of Computers and Information Technology, Taif University, Taif, 21944, Saudi Arabia.
| | - Mohammad Hijji
- Faculty of Computers and Information Technology, University of Tabuk, Tabuk, 71491 Saudi Arabia.
| | - Khan Muhammad
- Visual Analytics for Knowledge Laboratory (VIS2KNOW Lab), Department of Applied AI, School of Convergence, College of Computing and Informatics, Sungkyunkwan University, Seoul, 03063, South Korea.
| |
Collapse
|
15
|
Xin J, Yu Y, Shen Q, Zhang S, Su N, Wang Z. BCT-Net: semantic-guided breast cancer segmentation on BUS. Med Biol Eng Comput 2025; 63:1809-1820. [PMID: 39883373 DOI: 10.1007/s11517-025-03304-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2024] [Accepted: 01/17/2025] [Indexed: 01/31/2025]
Abstract
Accurately and swiftly segmenting breast tumors is significant for cancer diagnosis and treatment. Ultrasound imaging stands as one of the widely employed methods in clinical practice. However, due to challenges such as low contrast, blurred boundaries, and prevalent shadows in ultrasound images, tumor segmentation remains a daunting task. In this study, we propose BCT-Net, a network amalgamating CNN and transformer components for breast tumor segmentation. BCT-Net integrates a dual-level attention mechanism to capture more features and redefines the skip connection module. We introduce the utilization of a classification task as an auxiliary task to impart additional semantic information to the segmentation network, employing supervised contrastive learning. A hybrid objective loss function is proposed, which combines pixel-wise cross-entropy, binary cross-entropy, and supervised contrastive learning loss. Experimental results demonstrate that BCT-Net achieves high precision, with Pre and DSC indices of 86.12% and 88.70%, respectively. Experiments conducted on the BUSI dataset of breast ultrasound images manifest that this approach exhibits high accuracy in breast tumor segmentation.
Collapse
Affiliation(s)
- Junchang Xin
- School of Computer Science and Engineering, Northeastern University, Shenyang, 110169, China
| | - Yaqi Yu
- School of Computer Science and Engineering, Northeastern University, Shenyang, 110169, China
| | - Qi Shen
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, 110169, China
| | - Shudi Zhang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, 110169, China
| | - Na Su
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, 110169, China
| | - Zhiqiong Wang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, 110169, China.
| |
Collapse
|
16
|
Aravazhi PS, Gunasekaran P, Benjamin NZY, Thai A, Chandrasekar KK, Kolanu ND, Prajjwal P, Tekuru Y, Brito LV, Inban P. The integration of artificial intelligence into clinical medicine: Trends, challenges, and future directions. Dis Mon 2025; 71:101882. [PMID: 40140300 DOI: 10.1016/j.disamonth.2025.101882] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/28/2025]
Abstract
BACKGROUND AND OBJECTIVES AI has emerged as a transformative force in clinical medicine, changing the diagnosis, treatment, and management of patients. Tools have been derived for working with ML, DL, and NLP algorithms to analyze large complex medical datasets with unprecedented accuracy and speed, thereby improving diagnostic precision, treatment personalization, and patient care outcomes. For example, CNNs have dramatically improved the accuracy of medical imaging diagnoses, and NLP algorithms have greatly helped extract insights from unstructured data, including EHRs. However, there are still numerous challenges that face AI integration into clinical workflows, including data privacy, algorithmic bias, ethical dilemmas, and problems with the interpretability of "black-box" AI models. These barriers have thus far prevented the widespread application of AI in health care, and its possible trends, obstacles, and future implications are necessary to be systematically explored. The purpose of this paper is, therefore, to assess the current trends in AI applications in clinical medicine, identify those obstacles that are hindering adoption, and identify possible future directions. This research hopes to synthesize evidence from other peer-reviewed articles to provide a more comprehensive understanding of the role that AI plays to advance clinical practices, improve patient outcomes, or enhance decision-making. METHODS A systematic review was done according to the PRISMA guidelines to explore the integration of Artificial Intelligence in clinical medicine, including trends, challenges, and future directions. PubMed, Cochrane Library, Web of Science, and Scopus databases were searched for peer-reviewed articles from 2014 to 2024 with keywords such as "Artificial Intelligence in Medicine," "AI in Clinical Practice," "Machine Learning in Healthcare," and "Ethical Implications of AI in Medicine." Studies focusing on AI application in diagnostics, treatment planning, and patient care reporting measurable clinical outcomes were included. Non-clinical AI applications and articles published before 2014 were excluded. Selected studies were screened for relevance, and then their quality was critically appraised to synthesize data reliably and rigorously. RESULTS This systematic review includes the findings of 8 studies that pointed out the transformational role of AI in clinical medicine. AI tools, such as CNNs, had diagnostic accuracy more than the traditional methods, particularly in radiology and pathology. Predictive models efficiently supported risk stratification, early disease detection, and personalized medicine. Despite these improvements, significant hurdles, including data privacy, algorithmic bias, and resistance from clinicians regarding the "black-box" nature of AI, had yet to be surmounted. XAI has emerged as an attractive solution that offers the promise to enhance interpretability and trust. As a whole, AI appeared promising in enhancing diagnostics, treatment personalization, and clinical workflows by dealing with systemic inefficiencies. CONCLUSION The transformation potential of AI in clinical medicine can transform diagnostics, treatment strategies, and efficiency. Overcoming obstacles such as concerns about data privacy, the danger of algorithmic bias, and difficulties with interpretability may pave the way for broader use and facilitate improvement in patient outcomes while transforming clinical workflows to bring sustainability into healthcare delivery.
Collapse
Affiliation(s)
| | | | | | - Andy Thai
- Internal Medicine, Alameda Health System, Highland Hospital, Oakland, USA
| | | | | | | | - Yogesh Tekuru
- RVM Institute of Medical Sciences and Research Center, Laxmakkapally, India
| | | | - Pugazhendi Inban
- Internal Medicine, St. Mary's General Hospital and Saint Clare's Health, NY, USA.
| |
Collapse
|
17
|
Liu J, Bhadra S, Shafaat O, Mukherjee P, Parnell C, Summers RM. A unified approach to medical image segmentation by leveraging mixed supervision and self and transfer learning (MIST). Comput Med Imaging Graph 2025; 122:102517. [PMID: 40088573 PMCID: PMC12007390 DOI: 10.1016/j.compmedimag.2025.102517] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2024] [Revised: 01/15/2025] [Accepted: 02/22/2025] [Indexed: 03/17/2025]
Abstract
Medical image segmentation is important for quantitative disease diagnosis and treatment but relies on accurate pixel-wise labels, which are costly, time-consuming, and require domain expertise. This work introduces MIST (MIxed supervision, Self, and Transfer learning) to reduce manual labeling in medical image segmentation. A small set of cases was manually annotated ("strong labels"), while the rest used automated, less accurate labels ("weak labels"). Both label types trained a dual-branch network with a shared encoder and two decoders. Self-training iteratively refined weak labels, and transfer learning reduced computational costs by freezing the encoder and fine-tuning the decoders. Applied to segmenting muscle, subcutaneous, and visceral adipose tissue, MIST used only 100 manually labeled slices from 20 CT scans to generate accurate labels for all slices of 102 internal scans, which were then used to train a 3D nnU-Net model. Using MIST to update weak labels significantly improved nnU-Net segmentation accuracy compared to training directly on strong and weak labels. Dice similarity coefficient (DSC) increased for muscle (89.2 ± 4.3% to 93.2 ± 2.1%), subcutaneous (75.1 ± 14.4% to 94.2 ± 2.8%), and visceral adipose tissue (66.6 ± 16.4% to 77.1 ± 19.0% ) on an internal dataset (p<.05). DSC improved for muscle (80.5 ± 6.9% to 86.6 ± 3.9%) and subcutaneous adipose tissue (61.8 ± 12.5% to 82.7 ± 11.1%) on an external dataset (p<.05). MIST reduced the annotation burden by 99%, enabling efficient, accurate pixel-wise labeling for medical image segmentation. Code is available at https://github.com/rsummers11/NIH_CADLab_Body_Composition.
Collapse
Affiliation(s)
- Jianfei Liu
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA.
| | - Sayantan Bhadra
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| | - Omid Shafaat
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| | - Pritam Mukherjee
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| | - Christopher Parnell
- Walter Reed National Military Medical Center, 4494 Palmer Rd N, Bethesda, 20814, MD, USA
| | - Ronald M Summers
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, 10 Center Dr, Bethesda, 20892, MD, USA
| |
Collapse
|
18
|
Dai C, Yang Q, Zhou J, Zhu L, Lin L, Wang J, Cai C, Cai S. Single-shot T 2 mapping via multi-slice information sharing based on switching modulation patterns multiple overlapping-echo detachment imaging. Med Phys 2025; 52:4464-4479. [PMID: 40129089 DOI: 10.1002/mp.17778] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2024] [Revised: 02/18/2025] [Accepted: 03/06/2025] [Indexed: 03/26/2025] Open
Abstract
BACKGROUND Quantitative magnetic resonance imaging (qMRI) offers reliable biomarkers in clinic. Nevertheless, most qMRI methods are time-consuming and sensitive to motion. Single-shot multiple overlapping-echo detachment (MOLED) magnetic resonance imaging can deliver robust T2 mapping in about 100 ms with high motion tolerance. However, its spatial resolution is relatively low due to the limitations of signal-to-noise ratio (SNR) and echo-train length. At the mean time, the number of echoes with different evolution times collected is usually limited, which is not conducive to T2 mapping in high accuracy. PURPOSE To propose a novel method to improve the spatial resolution and quantification accuracy of single-shot MOLED T2 mapping. METHODS A new method called switching modulation patterns multiple overlapping-echo detachment imaging (SWP-MOLED) was designed for multi-slice information sharing via switching the k-space modulation pattern of MOLED imaging. In the SWP-MOLED pulse sequence, three different k-space modulation patterns were devised, making the 12 main echoes of any three adjacent slices symmetrically and uniformly distributed around their k-space centers to obtain diverse contrast weighting information. A multi-slice fusion three-dimensional spatial attention context-guided U-Net was trained with 3000/7000 synthetic data with geometric/brain patterns to efficiently learn the mapping relationship between SWP-MOLED signals and T2 maps. Experiments on numerical human brains, a phantom containing MnCl2 solutions with different concentrations, three healthy volunteers, and three patients diagnosed with meningioma or glioblastoma were performed. The effectiveness of the new method was quantitatively assessed using the structure similarity index measure (SSIM) and root mean square error (RMSE). Multiple statistical analyses were utilized to evaluate the accuracy and significance of the method, including linear regression, Bland-Altman analysis, Mann-Whitney test, Wilcoxon signed rank test, and Friedman test with Bonferroni correction, with the p-value significance level of 0.05. RESULTS The results from numerical human brain (The average SSIM of the reconstructed T2 maps was 0.9742/0.9782/0.9826 for MOLED/MS-MOLED/SWP-MOLED) and phantom (The slope of linear fitting of the predicted T2 values vs. reference values was 0.9934/9942/0.9972 for MOLED/MS-MOLED/SWP-MOLED) demonstrated that more accurate T2 maps were delivered by the proposed method, closely resembling the reference maps. From the Friedman test performed on the results of the test data set after the multi-comparison correction, we found that the pairwise performance differences among different reconstruction networks were all statistically significant (p < 0.001). In healthy human brain experiments, the comparison of SWP-MOLED reconstruction with reference measurements indicated no significant difference (p = 0.4504). SWP-MOLED was quite repeatable (average coefficient of variation [CV] = 4.17%) and was not corrupted by motion (average CV = 7.49%). Moreover, the proposed method exhibited clearer lesion contours in clinical cases, demonstrating the potential of the proposed method for clinical applications. CONCLUSIONS SWP-MOLED can efficiently exploit the structural similarity and parameter-weighted information diversity of adjacent slices to improve the spatial resolution and quantification accuracy of MOLED T2 mapping. It also exhibits excellent motion robustness. This technique would extend the application of MOLED imaging.
Collapse
Affiliation(s)
- Chenyang Dai
- Department of Electronic Science, Fujian Provincial Key Laboratory of Plasma and Magnetic Resonance, Xiamen University, Xiamen, China
- Shenzhen Research Institute of Xiamen University, Shenzhen, China
| | - Qinqin Yang
- Department of Electronic Science, Fujian Provincial Key Laboratory of Plasma and Magnetic Resonance, Xiamen University, Xiamen, China
| | - Jianjun Zhou
- Department of Radiology, Zhongshan Hospital (Xiamen), Fudan University, Fujian Province Key Clinical Specialty Construction Project (Medical Imaging Department), Xiamen Key Laboratory of Clinical Transformation of Imaging Big Data and Artificial Intelligence, Xiamen, China
| | - Liuhong Zhu
- Department of Electronic Science, Fujian Provincial Key Laboratory of Plasma and Magnetic Resonance, Xiamen University, Xiamen, China
- Department of Radiology, Zhongshan Hospital (Xiamen), Fudan University, Fujian Province Key Clinical Specialty Construction Project (Medical Imaging Department), Xiamen Key Laboratory of Clinical Transformation of Imaging Big Data and Artificial Intelligence, Xiamen, China
| | - Liangjie Lin
- Department of Clinical & Technical Support, Philips Healthcare, Beijing, China
| | - Jiazheng Wang
- Department of Clinical & Technical Support, Philips Healthcare, Beijing, China
| | - Congbo Cai
- Department of Electronic Science, Fujian Provincial Key Laboratory of Plasma and Magnetic Resonance, Xiamen University, Xiamen, China
| | - Shuhui Cai
- Department of Electronic Science, Fujian Provincial Key Laboratory of Plasma and Magnetic Resonance, Xiamen University, Xiamen, China
- Shenzhen Research Institute of Xiamen University, Shenzhen, China
| |
Collapse
|
19
|
Alnageeb MHO, M H S. Real-time brain tumour diagnoses using a novel lightweight deep learning model. Comput Biol Med 2025; 192:110242. [PMID: 40334297 DOI: 10.1016/j.compbiomed.2025.110242] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2024] [Revised: 04/10/2025] [Accepted: 04/21/2025] [Indexed: 05/09/2025]
Abstract
Brain tumours continue to be a primary cause of worldwide death, highlighting the critical need for effective and accurate diagnostic tools. This article presents MK-YOLOv8, an innovative lightweight deep learning framework developed for the real-time detection and categorization of brain tumours from MRI images. Based on the YOLOv8 architecture, the proposed model incorporates Ghost Convolution, the C3Ghost module, and the SPPELAN module to improve feature extraction and substantially decrease computational complexity. An x-small object detection layer has been added, supporting precise detection of small and x-small tumours, which is crucial for early diagnosis. Trained on the Figshare Brain Tumour (FBT) dataset comprising (3,064) MRI images, MK-YOLOv8 achieved a mean Average Precision (mAP) of 99.1% at IoU (0.50) and 88.4% at IoU (0.50-0.95), outperforming YOLOv8 (98% and 78.8%, respectively). Glioma recall improved by 26%, underscoring the enhanced sensitivity to challenging tumour types. With a computational footprint of only 96.9 GFLOPs (representing 37.5% of YOYOLOv8x'sFLOPs) and utilizing 12.6 million parameters, a mere 18.5% of YOYOLOv8's parameters, MK-YOLOv8 delivers high efficiency with reduced resource demands. Also, it trained on the Br35H dataset (801 images) to guarantee the model's robustness and generalization; it achieved a mAP of 98.6% at IoU (0.50). The suggested model operates at 62 frames per second (FPS) and is suited for real-time clinical processes. These developments establish MK-YOLOv8 as an innovative framework, overcoming challenges in tiny tumour identification and providing a generalizable, adaptable, and precise detection approach for brain tumour diagnostics in clinical settings.
Collapse
Affiliation(s)
| | - Supriya M H
- Cochin University of Science and Technology, Department of Electronics, CUSAT, Kochi, 682022, Kerala, India
| |
Collapse
|
20
|
Cabezas E, Toro-Tobon D, Johnson T, Álvarez M, Azadi JR, Gonzalez-Velasquez C, Singh Ospina N, Ponce OJ, Branda ME, Brito JP. ChatGPT-4's Accuracy in Estimating Thyroid Nodule Features and Cancer Risk From Ultrasound Images. Endocr Pract 2025; 31:716-723. [PMID: 40139461 DOI: 10.1016/j.eprac.2025.03.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/24/2025] [Revised: 03/06/2025] [Accepted: 03/18/2025] [Indexed: 03/29/2025]
Abstract
OBJECTIVE To evaluate the performance of GPT-4 and GPT-4o in accurately identifying features and categories from thyroid nodule ultrasound images following the American College of Radiology Thyroid Imaging Reporting and Data System (TI-RADS). METHODS This comparative validation study, conducted between October 2023 and May 2024, utilized 202 thyroid ultrasound images sourced from 3 open-access databases. Both complete and cropped versions of each image were independently evaluated by expert radiologists to establish a reference standard for TI-RADS features and categories. GPT-4 and GPT-4o were prompted to analyze each image, and their generated TI-RADS outputs were compared to the reference standard. RESULTS GPT-4 demonstrated high specificity but low sensitivity when assessing complete thyroid ultrasound images across most TI-RADS categories, resulting in mixed overall accuracy. For low-risk nodules (benign), GPT-4 achieved 25.0% sensitivity, 99.5% specificity, and 93.6% accuracy. In contrast, in the higher risk moderately suspicious category GPT-4 showed 75% sensitivity, 22.2% specificity, and 42.1% accuracy. While GPT-4 effectively identified features like smooth margins (73% vs 65% the reference standard), it struggled to identify other features like isoechoic echogenicity (5% vs 46%), and echogenic foci (3% vs 27%). The assessment of cropped images using both GPT-4 and GPT-4o followed similar patterns, though with slight deviations indicating a decrease in performance compared to GPT-4's assessment of complete images. CONCLUSION While GPT-4 and GPT-4o models show potential for improving the efficiency of thyroid nodule triage, their performance remains suboptimal, particularly in higher-risk categories. Further refinement and validation of these models are necessary before clinical implementation.
Collapse
Affiliation(s)
- Esteban Cabezas
- Knowledge and Evaluation Research Unit, Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Department of Medicine, Mayo Clinic, Rochester, Minnesota
| | - David Toro-Tobon
- Knowledge and Evaluation Research Unit, Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Department of Medicine, Mayo Clinic, Rochester, Minnesota; Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Mayo Clinic, Rochester, Minnesota
| | - Thomas Johnson
- Division of Endocrinology, Mercy Hospital, Springfield, Missouri
| | - Marco Álvarez
- Universidad Politécnica de Sinaloa en Mazatlán, Sinaloa, México
| | - Javad R Azadi
- Johns Hopkins University School of Medicine, Department of Radiology, Baltimore, Maryland
| | | | - Naykky Singh Ospina
- Division of Endocrinology, Department of Medicine, University of Florida, Gainesville, Florida
| | - Oscar J Ponce
- Knowledge and Evaluation Research Unit, Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Department of Medicine, Mayo Clinic, Rochester, Minnesota; University Hospitals Plymouth NHS Trust, Plymouth, Devon, the United Kingdom
| | - Megan E Branda
- Knowledge and Evaluation Research Unit, Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Department of Medicine, Mayo Clinic, Rochester, Minnesota; Division of Biomedical Statistics and Informatics, Department of Health Sciences Research, Mayo Clinic, Rochester, Minnesota
| | - Juan P Brito
- Knowledge and Evaluation Research Unit, Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Department of Medicine, Mayo Clinic, Rochester, Minnesota; Division of Endocrinology, Diabetes, Metabolism, and Nutrition, Mayo Clinic, Rochester, Minnesota.
| |
Collapse
|
21
|
Hussain J, Båth M, Ivarsson J. Generative adversarial networks in medical image reconstruction: A systematic literature review. Comput Biol Med 2025; 191:110094. [PMID: 40198987 DOI: 10.1016/j.compbiomed.2025.110094] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Revised: 01/12/2025] [Accepted: 03/25/2025] [Indexed: 04/10/2025]
Abstract
PURPOSE Recent advancements in generative adversarial networks (GANs) have demonstrated substantial potential in medical image processing. Despite this progress, reconstructing images from incomplete data remains a challenge, impacting image quality. This systematic literature review explores the use of GANs in enhancing and reconstructing medical imaging data. METHOD A document survey of computing literature was conducted using the ACM Digital Library to identify relevant articles from journals and conference proceedings using keyword combinations, such as "generative adversarial networks or generative adversarial network," "medical image or medical imaging," and "image reconstruction." RESULTS Across the reviewed articles, there were 122 datasets used in 175 instances, 89 top metrics employed 335 times, 10 different tasks with a total count of 173, 31 distinct organs featured in 119 instances, and 18 modalities utilized in 121 instances, collectively depicting significant utilization of GANs in medical imaging. The adaptability and efficacy of GANs were showcased across diverse medical tasks, organs, and modalities, utilizing top public as well as private/synthetic datasets for disease diagnosis, including the identification of conditions like cancer in different anatomical regions. The study emphasized GAN's increasing integration and adaptability in diverse radiology modalities, showcasing their transformative impact on diagnostic techniques, including cross-modality tasks. The intricate interplay between network size, batch size, and loss function refinement significantly impacts GAN's performance, although challenges in training persist. CONCLUSIONS The study underscores GANs as dynamic tools shaping medical imaging, contributing significantly to image quality, training methodologies, and overall medical advancements, positioning them as substantial components driving medical advancements.
Collapse
Affiliation(s)
- Jabbar Hussain
- Dept. of Applied IT, University of Gothenburg, Forskningsgången 6, 417 56, Sweden.
| | - Magnus Båth
- Department of Medical Radiation Sciences, University of Gothenburg, Sweden
| | - Jonas Ivarsson
- Dept. of Applied IT, University of Gothenburg, Forskningsgången 6, 417 56, Sweden
| |
Collapse
|
22
|
Harris CE, Liu L, Almeida L, Kassick C, Makrogiannis S. Artificial intelligence in pediatric osteopenia diagnosis: evaluating deep network classification and model interpretability using wrist X-rays. Bone Rep 2025; 25:101845. [PMID: 40343188 PMCID: PMC12059325 DOI: 10.1016/j.bonr.2025.101845] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/16/2024] [Revised: 04/11/2025] [Accepted: 04/21/2025] [Indexed: 05/11/2025] Open
Abstract
Osteopenia is a bone disorder that causes low bone density and affects millions of people worldwide. Diagnosis of this condition is commonly achieved through clinical assessment of bone mineral density (BMD). State of the art machine learning (ML) techniques, such as convolutional neural networks (CNNs) and transformer models, have gained increasing popularity in medicine. In this work, we employ six deep networks for osteopenia vs. healthy bone classification using X-ray imaging from the pediatric wrist dataset GRAZPEDWRI-DX. We apply two explainable AI techniques to analyze and interpret visual explanations for network decisions. Experimental results show that deep networks are able to effectively learn osteopenic and healthy bone features, achieving high classification accuracy rates. Among the six evaluated networks, DenseNet201 with transfer learning yielded the top classification accuracy at 95.2 %. Furthermore, visual explanations of CNN decisions provide valuable insight into the blackbox inner workings and present interpretable results. Our evaluation of deep network classification results highlights their capability to accurately differentiate between osteopenic and healthy bones in pediatric wrist X-rays. The combination of high classification accuracy and interpretable visual explanations underscores the promise of incorporating machine learning techniques into clinical workflows for the early and accurate diagnosis of osteopenia.
Collapse
Affiliation(s)
- Chelsea E. Harris
- Division of Physics, Engineering, Mathematics, and Computer Science, Delaware State University, 1200 N. Dupont Hwy., Dover, 19901, DE, USA
| | - Lingling Liu
- Division of Physics, Engineering, Mathematics, and Computer Science, Delaware State University, 1200 N. Dupont Hwy., Dover, 19901, DE, USA
| | - Luiz Almeida
- Department of Orthopaedic Surgery, Duke University, 2080 Duke University Road, Durham, 27710, NC, USA
| | - Carolina Kassick
- Division of Physics, Engineering, Mathematics, and Computer Science, Delaware State University, 1200 N. Dupont Hwy., Dover, 19901, DE, USA
| | - Sokratis Makrogiannis
- Division of Physics, Engineering, Mathematics, and Computer Science, Delaware State University, 1200 N. Dupont Hwy., Dover, 19901, DE, USA
| |
Collapse
|
23
|
Levital MF, Khawaled S, Kennedy JA, Freiman M. Non-parametric Bayesian deep learning approach for whole-body low-dose PET reconstruction and uncertainty assessment. Med Biol Eng Comput 2025; 63:1715-1730. [PMID: 39847156 DOI: 10.1007/s11517-025-03296-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2024] [Accepted: 01/12/2025] [Indexed: 01/24/2025]
Abstract
Positron emission tomography (PET) imaging plays a pivotal role in oncology for the early detection of metastatic tumors and response to therapy assessment due to its high sensitivity compared to anatomical imaging modalities. The balance between image quality and radiation exposure is critical, as reducing the administered dose results in a lower signal-to-noise ratio (SNR) and information loss, which may significantly affect clinical diagnosis. Deep learning (DL) algorithms have recently made significant progress in low-dose (LD) PET reconstruction. Nevertheless, a successful clinical application requires a thorough evaluation of uncertainty to ensure informed clinical judgment. We propose NPB-LDPET, a DL-based non-parametric Bayesian framework for LD PET reconstruction and uncertainty assessment. Our framework utilizes an Adam optimizer with stochastic gradient Langevin dynamics (SGLD) to sample from the underlying posterior distribution. We employed the Ultra-low-dose PET Challenge dataset to assess our framework's performance relative to the Monte Carlo dropout benchmark. We evaluated global reconstruction accuracy utilizing SSIM, PSNR, and NRMSE, local lesion conspicuity using mean absolute error (MAE) and local contrast, and the clinical relevance of uncertainty maps employing correlation between the uncertainty measures and the dose reduction factor (DRF). Our NPB-LDPET reconstruction method exhibits a significantly superior global reconstruction accuracy for various DRFs (paired t-test, p < 0.0001 , N=10, 631). Moreover, we demonstrate a 21% reduction in MAE (573.54 vs. 723.70, paired t-test, p < 0.0001 , N=28) and an 8.3% improvement in local lesion contrast (2.077 vs. 1.916, paired t-test, p < 0.0001 , N=28). Furthermore, our framework exhibits a stronger correlation between the predicted uncertainty 95th percentile score and the DRF (r 2 = 0.9174 vs.r 2 = 0.6144 , N=10, 631). The proposed framework has the potential to improve clinical decision-making for LD PET imaging by providing a more accurate and informative reconstruction while reducing radiation exposure.
Collapse
Affiliation(s)
- Maya Fichmann Levital
- The Interdisciplinary Program for Robotics and Autonomous Systems, Technion - Israel Institute of Technology, Haifa, Israel
| | - Samah Khawaled
- The Interdisciplinary Program in Applied Mathematics, Faculty of Mathematics, Technion - Israel Institute of Technology, Haifa, Israel
| | - John A Kennedy
- Department of Nuclear Medicine, Rambam Health Care Campus, Haifa, Israel
- B. Rappaport Faculty of Medicine, Technion - Israel Institute of Technology, Haifa, Israel
| | - Moti Freiman
- Faculty of Biomedical Engineering, Technion - Israel Institute of Technology, Haifa, Israel.
| |
Collapse
|
24
|
Sobhi N, Sadeghi-Bazargani Y, Mirzaei M, Abdollahi M, Jafarizadeh A, Pedrammehr S, Alizadehsani R, Tan RS, Islam SMS, Acharya UR. Artificial intelligence for early detection of diabetes mellitus complications via retinal imaging. J Diabetes Metab Disord 2025; 24:104. [PMID: 40224528 PMCID: PMC11993533 DOI: 10.1007/s40200-025-01596-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/03/2024] [Accepted: 02/23/2025] [Indexed: 04/15/2025]
Abstract
Background Diabetes mellitus (DM) increases the risk of vascular complications, and retinal vasculature imaging serves as a valuable indicator of both microvascular and macrovascular health. Moreover, artificial intelligence (AI)-enabled systems developed for high-throughput detection of diabetic retinopathy (DR) using digitized retinal images have become clinically adopted. This study reviews AI applications using retinal images for DM-related complications, highlighting advancements beyond DR screening, diagnosis, and prognosis, and addresses implementation challenges, such as ethics, data privacy, equitable access, and explainability. Methods We conducted a thorough literature search across several databases, including PubMed, Scopus, and Web of Science, focusing on studies involving diabetes, the retina, and artificial intelligence. We reviewed the original research based on their methodology, AI algorithms, data processing techniques, and validation procedures to ensure a detailed analysis of AI applications in diabetic retinal imaging. Results Retinal images can be used to diagnose DM complications including DR, neuropathy, nephropathy, and atherosclerotic cardiovascular disease, as well as to predict the risk of cardiovascular events. Beyond DR screening, AI integration also offers significant potential to address the challenges in the comprehensive care of patients with DM. Conclusion With the ability to evaluate the patient's health status in relation to DM complications as well as risk prognostication of future cardiovascular complications, AI-assisted retinal image analysis has the potential to become a central tool for modern personalized medicine in patients with DM.
Collapse
Affiliation(s)
- Navid Sobhi
- Nikookari Eye Center, Tabriz University of Medical Sciences, Tabriz, Iran
| | | | - Majid Mirzaei
- Student Research Committee, Tabriz University of Medical Sciences, Tabriz, Iran
| | - Mirsaeed Abdollahi
- Nikookari Eye Center, Tabriz University of Medical Sciences, Tabriz, Iran
| | - Ali Jafarizadeh
- Nikookari Eye Center, Tabriz University of Medical Sciences, Tabriz, Iran
| | - Siamak Pedrammehr
- Institute for Intelligent Systems Research and Innovation (IISRI), Deakin University, 75 Pigdons Rd, Waurn Ponds, VIC 3216 Australia
- Faculty of Design, Tabriz Islamic Art University, Tabriz, Iran
| | - Roohallah Alizadehsani
- Institute for Intelligent Systems Research and Innovation (IISRI), Deakin University, 75 Pigdons Rd, Waurn Ponds, VIC 3216 Australia
| | - Ru-San Tan
- National Heart Centre Singapore, Singapore, Singapore
- Duke-NUS Medical School, Singapore, Singapore
| | - Sheikh Mohammed Shariful Islam
- Institute for Physical Activity and Nutrition, School of Exercise and Nutrition Sciences, Deakin University, Melbourne, VIC Australia
- Cardiovascular Division, The George Institute for Global Health, Newtown, Australia
- Sydney Medical School, University of Sydney, Camperdown, Australia
| | - U. Rajendra Acharya
- School of Mathematics, Physics, and Computing, University of Southern Queensland, Springfield, QLD 4300 Australia
- Centre for Health Research, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
25
|
Qian X, Shao HC, Li Y, Lu W, Zhang Y. Histogram matching-enhanced adversarial learning for unsupervised domain adaptation in medical image segmentation. Med Phys 2025; 52:4299-4317. [PMID: 40102198 DOI: 10.1002/mp.17757] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2024] [Revised: 02/20/2025] [Accepted: 02/26/2025] [Indexed: 03/20/2025] Open
Abstract
BACKGROUND Unsupervised domain adaptation (UDA) seeks to mitigate the performance degradation of deep neural networks when applied to new, unlabeled domains by leveraging knowledge from source domains. In medical image segmentation, prevailing UDA techniques often utilize adversarial learning to address domain shifts for cross-modality adaptation. Current research on adversarial learning tends to adopt increasingly complex models and loss functions, making the training process highly intricate and less stable/robust. Furthermore, most methods primarily focused on segmentation accuracy while neglecting the associated confidence levels and uncertainties. PURPOSE To develop a simple yet effective UDA method based on histogram matching-enhanced adversarial learning (HMeAL-UDA), and provide comprehensive uncertainty estimations of the model predictions. METHODS Aiming to bridge the domain gap while reducing the model complexity, we developed a novel adversarial learning approach to align multi-modality features. The method, termed HMeAL-UDA, integrates a plug-and-play histogram matching strategy to mitigate domain-specific image style biases across modalities. We employed adversarial learning to constrain the model in the prediction space, enabling it to focus on domain-invariant features during segmentation. Moreover, we quantified the model's prediction confidence using Monte Carlo (MC) dropouts to assess two voxel-level uncertainty estimates of the segmentation results, which were subsequently aggregated into a volume-level uncertainty score, providing an overall measure of the model's reliability. The proposed method was evaluated on three public datasets (Combined Healthy Abdominal Organ Segmentation [CHAOS], Beyond the Cranial Vault [BTCV], and Abdominal Multi-Organ Segmentation Challenge [AMOS]) and one in-house clinical dataset (UTSW). We used 30 MRI scans (20 from the CHAOS dataset and 10 from the in-house dataset) and 30 CT scans from the BTCV dataset for UDA-based, cross-modality liver segmentation. Additionally, 240 CT scans and 60 MRI scans from the AMOS dataset were utilized for cross-modality multi-organ segmentation. The training and testing sets for each modality were split with ratios of approximately 4:1-3:1. RESULTS Extensive experiments on cross-modality medical image segmentation demonstrated the superiority of HMeAL-UDA over two state-of-the-art approaches. HMeAL-UDA achieved a mean (± s.d.) Dice similarity coefficient (DSC) of 91.34% ± 1.23% and an HD95 of 6.18 ± 2.93 mm for cross-modality (from CT to MRI) adaptation of abdominal multi-organ segmentation, and a DSC of 87.13% ± 3.67% with an HD95 of 2.48 ± 1.56 mm for segmentation adaptation in the opposite direction (MRI to CT). The results are approaching or even outperforming those of supervised methods trained with "ground-truth" labels in the target domain. In addition, we provide a comprehensive assessment of the model's uncertainty, which can help with the understanding of segmentation reliability to guide clinical decisions. CONCLUSION HMeAL-UDA provides a powerful segmentation tool to address cross-modality domain shifts, with the potential to generalize to other deep learning applications in medical imaging.
Collapse
Affiliation(s)
- Xiaoxue Qian
- Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - Hua-Chieh Shao
- Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - Yunxiang Li
- Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - Weiguo Lu
- Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| | - You Zhang
- Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, Texas, USA
| |
Collapse
|
26
|
Yu J, Liu Q, Xu C, Zhou Q, Xu J, Zhu L, Chen C, Zhou Y, Xiao B, Zheng L, Zhou X, Zhang F, Ye Y, Mi H, Zhang D, Yang L, Wu Z, Wang J, Chen M, Zhou Z, Wang H, Wang VY, Wang E, Xu D. Deep learning-driven modality imputation and subregion segmentation to enhance high-grade glioma grading. BMC Med Inform Decis Mak 2025; 25:200. [PMID: 40448035 DOI: 10.1186/s12911-025-03029-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2024] [Accepted: 05/12/2025] [Indexed: 06/02/2025] Open
Abstract
PURPOSE This study aims to develop a deep learning framework that leverages modality imputation and subregion segmentation to improve grading accuracy in high-grade gliomas. MATERIALS AND METHODS A retrospective analysis was conducted using data from 1,251 patients in the BraTS2021 dataset as the main cohort and 181 clinical cases collected from a medical center between April 2013 and June 2018 (51 years ± 17; 104 males) as the external test set. We propose a PatchGAN-based modality imputation network with an Aggregated Residual Transformer (ART) module combining Transformer self-attention and CNN feature extraction via residual links, paired with a U-Net variant for segmentation. Generative accuracy used PSNR and SSIM for modality conversions, while segmentation performance was measured with DSC and HD95 across necrotic core (NCR), edema (ED), and enhancing tumor (ET) regions. Senior radiologists conducted a comprehensive Likert-based assessment, with diagnostic accuracy evaluated by AUC. Statistical analysis was performed using the Wilcoxon signed-rank test and the DeLong test. RESULTS The best source-target modality pairs for imputation were T1 to T1ce and T1ce to T2 (p < 0.001). In subregion segmentation, the overall DSC was 0.878 and HD95 was 19.491, with the ET region showing the highest segmentation accuracy (DSC: 0.877, HD95: 12.149). Clinical validation revealed an improvement in grading accuracy by the senior radiologist, with the AUC increasing from 0.718 to 0.913 (P < 0.001) when using the combined imputation and segmentation models. CONCLUSION The proposed deep learning framework improves high-grade glioma grading by modality imputation and segmentation, aiding the senior radiologist and offering potential to advance clinical decision-making.
Collapse
Affiliation(s)
- Jiabin Yu
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
- State Key Laboratory of Cardiology and Medical Innovation Center, Frontier Science Center for Stem Cell Research, Bioinformatics Department, School of Life Sciences and Technology, Shanghai East Hospital, Tongji University, Shanghai, 200092, China
- Taizhou Key Laboratory of Minimally Invasive Interventional Therapy & Artificial Intelligence, Department of Radiology Imaging, Taizhou Campus of Zhejiang Cancer Hospital (Taizhou Cancer Hospital), No. 50 Zhenxin Road, Xinhe Town, Wenling, Zhejiang, 317502, China
| | - Qi Liu
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China
| | - Chenjie Xu
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Qinli Zhou
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Jiajun Xu
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Lingying Zhu
- State Key Laboratory of Cardiology and Medical Innovation Center, Frontier Science Center for Stem Cell Research, Bioinformatics Department, School of Life Sciences and Technology, Shanghai East Hospital, Tongji University, Shanghai, 200092, China
| | - Chen Chen
- Taizhou Key Laboratory of Minimally Invasive Interventional Therapy & Artificial Intelligence, Department of Radiology Imaging, Taizhou Campus of Zhejiang Cancer Hospital (Taizhou Cancer Hospital), No. 50 Zhenxin Road, Xinhe Town, Wenling, Zhejiang, 317502, China
- Department of Diagnostic Ultrasound Imaging & Interventional Therapy, Hangzhou Institute of Medicine (HIM), Zhejiang Cancer Hospital, Chinese Academy of Sciences, No.1 East Banshan Road, Gongshu District, Hangzhou, Zhejiang, 310022, China
- Wenling Institute of Big Data and Artificial Intelligence in Medicine, Level 24, Building 1, XinShang Building, Xinhe Town, Wenling, Zhejiang, 317502, China
- Center of Intelligent Diagnosis and Therapy (Taizhou), Hangzhou Institute of Medicine(HIM), Chinese Academy of Sciences, Taizhou, Zhejiang, 317502, China
| | - Yahan Zhou
- Taizhou Key Laboratory of Minimally Invasive Interventional Therapy & Artificial Intelligence, Department of Radiology Imaging, Taizhou Campus of Zhejiang Cancer Hospital (Taizhou Cancer Hospital), No. 50 Zhenxin Road, Xinhe Town, Wenling, Zhejiang, 317502, China
- Wenling Institute of Big Data and Artificial Intelligence in Medicine, Level 24, Building 1, XinShang Building, Xinhe Town, Wenling, Zhejiang, 317502, China
- Center of Intelligent Diagnosis and Therapy (Taizhou), Hangzhou Institute of Medicine(HIM), Chinese Academy of Sciences, Taizhou, Zhejiang, 317502, China
| | - Binggang Xiao
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Lin Zheng
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China
| | - Xiaofeng Zhou
- Department of Radiation Oncology, The Second Affiliated Hospital, National Ministry of Education Key Laboratory of Cancer Prevention and Intervention, Zhejiang University School of Medicine, Hangzhou, Zhejiang, 310009, China
| | - Fengming Zhang
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China
| | - Yuhang Ye
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China
| | - Hongmei Mi
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Dongping Zhang
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Li Yang
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Zhiwei Wu
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Jiayi Wang
- College of Information Engineering, China Jiliang University, Hangzhou, Zhejiang, 310018, China
| | - Ming Chen
- Department of Ultrasound, Zhangzhou Affiliated Hospital of Fujian Medical University, Fujian, 363000, China
| | - Zhirui Zhou
- Radiation Oncology Center, Shanghai Medical College, Huashan Hospital, Fudan University, No.12 Wulumuqi Middle Road, Shanghai, 201107, China
| | - Haoyang Wang
- Department of Computer Science and Technology (Sino-American Joint Program), School of Information Science and Technology, Northeast Normal University, Changchun, Jilin, 130117, China
| | - Vicky Y Wang
- Taizhou Key Laboratory of Minimally Invasive Interventional Therapy & Artificial Intelligence, Department of Radiology Imaging, Taizhou Campus of Zhejiang Cancer Hospital (Taizhou Cancer Hospital), No. 50 Zhenxin Road, Xinhe Town, Wenling, Zhejiang, 317502, China.
- Wenling Institute of Big Data and Artificial Intelligence in Medicine, Level 24, Building 1, XinShang Building, Xinhe Town, Wenling, Zhejiang, 317502, China.
- Center of Intelligent Diagnosis and Therapy (Taizhou), Hangzhou Institute of Medicine(HIM), Chinese Academy of Sciences, Taizhou, Zhejiang, 317502, China.
| | - Enyu Wang
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China.
- Taizhou Key Laboratory of Minimally Invasive Interventional Therapy & Artificial Intelligence, Department of Radiology Imaging, Taizhou Campus of Zhejiang Cancer Hospital (Taizhou Cancer Hospital), No. 50 Zhenxin Road, Xinhe Town, Wenling, Zhejiang, 317502, China.
| | - Dong Xu
- Taizhou Campus, Zhejiang Cancer Hospital (Taizhou Cancer Hospital), Taizhou, Zhejiang, 317502, China.
- Taizhou Key Laboratory of Minimally Invasive Interventional Therapy & Artificial Intelligence, Department of Radiology Imaging, Taizhou Campus of Zhejiang Cancer Hospital (Taizhou Cancer Hospital), No. 50 Zhenxin Road, Xinhe Town, Wenling, Zhejiang, 317502, China.
- Department of Diagnostic Ultrasound Imaging & Interventional Therapy, Hangzhou Institute of Medicine (HIM), Zhejiang Cancer Hospital, Chinese Academy of Sciences, No.1 East Banshan Road, Gongshu District, Hangzhou, Zhejiang, 310022, China.
- Wenling Institute of Big Data and Artificial Intelligence in Medicine, Level 24, Building 1, XinShang Building, Xinhe Town, Wenling, Zhejiang, 317502, China.
| |
Collapse
|
27
|
Jim JR, Rayed ME, Mridha M, Nur K. XLLC-Net: A lightweight and explainable CNN for accurate lung cancer classification using histopathological images. PLoS One 2025; 20:e0322488. [PMID: 40445896 PMCID: PMC12124586 DOI: 10.1371/journal.pone.0322488] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2024] [Accepted: 03/22/2025] [Indexed: 06/02/2025] Open
Abstract
Lung cancer imaging plays a crucial role in early diagnosis and treatment, where machine learning and deep learning have significantly advanced the accuracy and efficiency of disease classification. This study introduces the Explainable and Lightweight Lung Cancer Net (XLLC-Net), a streamlined convolutional neural network designed for classifying lung cancer from histopathological images. Using the LC25000 dataset, which includes three lung cancer classes and two colon cancer classes, we focused solely on the three lung cancer classes for this study. XLLC-Net effectively discerns complex disease patterns within these classes. The model consists of four convolutional layers and contains merely 3 million parameters, considerably reducing its computational footprint compared to existing deep learning models. This compact architecture facilitates efficient training, completing each epoch in just 60 seconds. Remarkably, XLLC-Net achieves a classification accuracy of 99.62% [Formula: see text] 0.16%, with precision, recall, and F1 score of 99.33% [Formula: see text] 0.30%, 99.67% [Formula: see text] 0.30%, and 99.70% [Formula: see text] 0.30%, respectively. Furthermore, the integration of Explainable AI techniques, such as Saliency Map and GRAD-CAM, enhances the interpretability of the model, offering clear visual insights into its decision-making process. Our results underscore the potential of lightweight DL models in medical imaging, providing high accuracy and rapid training while ensuring model transparency and reliability.
Collapse
Affiliation(s)
- Jamin Rahman Jim
- Department of Computer Science, American International University-Bangladesh, Dhaka 1229, Bangladesh
| | - Md. Eshmam Rayed
- Department of Computer Science, American International University-Bangladesh, Dhaka 1229, Bangladesh
| | - M.F. Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka 1229, Bangladesh
| | - Kamruddin Nur
- Department of Computer Science, American International University-Bangladesh, Dhaka 1229, Bangladesh
| |
Collapse
|
28
|
Daniel Huang Y, Jadav N, Rutter G, Szymanski L, Bostina M, Harland DP. A generalist deep-learning volume segmentation tool for volume electron microscopy of biological samples. J Struct Biol 2025:108214. [PMID: 40449855 DOI: 10.1016/j.jsb.2025.108214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2024] [Revised: 05/12/2025] [Accepted: 05/27/2025] [Indexed: 06/03/2025]
Abstract
We present the Volume Segmentation Tool (VST), a deep learning software tool that implements volumetric image segmentation in volume electron microscopy image stack data from a wide range of biological sample types. VST automates the handling of data preprocessing, data augmentation, and network building, as well as the configuration for model training, while adapting to the specific dataset. We have tried to make VST more accessible by designing it to operate entirely on local hardware and have provided a browser-based interface with additional features for visualizations of the networks and augmented datasets. VST can utilise contour map prediction to support instance segmentation on top of semantic segmentation. Through examples from various resin-embedded sample derived transmission electron microscopy and scanning electron microscopy datasets, we demonstrate that VST achieves state of the art performance compared to existing approaches.
Collapse
Affiliation(s)
- Yuyao Daniel Huang
- Department of Microbiology & Immunology, University of Otago, Dunedin 9016, New Zealand
| | - Nickhil Jadav
- Department of Microbiology & Immunology, University of Otago, Dunedin 9016, New Zealand
| | - Georgia Rutter
- Department of Microbiology & Immunology, University of Otago, Dunedin 9016, New Zealand
| | - Lech Szymanski
- School of Computing, University of Otago, Dunedin 9016, New Zealand
| | - Mihnea Bostina
- Department of Microbiology & Immunology, University of Otago, Dunedin 9016, New Zealand.
| | - Duane P Harland
- Smart Foods & Bioproducts Science Group, AgResearch, Christchurch 7608, New Zealand; Biomolecular Interaction Centre (BIC), Te Pokapū Taunekeneke Rāpoi Ngota, New Zealand.
| |
Collapse
|
29
|
Jha D, Susladkar OK, Gorade V, Keles E, Antalek M, Seyithanoglu D, Cebeci T, Aktas HE, Kartal GD, Kaymakoglu S, Erturk SM, Velichko Y, Ladner DP, Borhani AA, Medetalibeyoglu A, Durak G, Bagci U. Large Scale MRI Collection and Segmentation of Cirrhotic Liver. Sci Data 2025; 12:896. [PMID: 40436863 PMCID: PMC12119857 DOI: 10.1038/s41597-025-05201-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2024] [Accepted: 05/14/2025] [Indexed: 06/01/2025] Open
Abstract
Liver cirrhosis represents the end stage of chronic liver disease, characterized by extensive fibrosis and nodular regeneration that significantly increases mortality risk. While magnetic resonance imaging (MRI) offers a non-invasive assessment, accurately segmenting cirrhotic livers presents substantial challenges due to morphological alterations and heterogeneous signal characteristics. Deep learning approaches show promise for automating these tasks, but progress has been limited by the absence of large-scale, annotated datasets. Here, we present CirrMRI600+, the first comprehensive dataset comprising 628 high-resolution abdominal MRI scans (310 T1-weighted and 318 T2-weighted sequences, totaling nearly 40,000 annotated slices) with expert-validated segmentation labels for cirrhotic livers. The dataset includes demographic information, clinical parameters, and histopathological validation where available. Additionally, we provide benchmark results from 11 state-of-the-art deep learning experiments to establish performance standards. CirrMRI600+ enables the development and validation of advanced computational methods for cirrhotic liver analysis, potentially accelerating progress toward automated Cirrhosis visual staging and personalized treatment planning.
Collapse
Affiliation(s)
- Debesh Jha
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Onkar Kishor Susladkar
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Vandan Gorade
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Elif Keles
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Matthew Antalek
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | | | - Timurhan Cebeci
- Istanbul University, School of Medicine (Capa), Istanbul, Turkey
| | - Halil Ertugrul Aktas
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | | | | | | | - Yuri Velichko
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Daniela P Ladner
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Amir A Borhani
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | | | - Gorkem Durak
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA
| | - Ulas Bagci
- Machine & Hybrid Intelligence Lab, Department of Radiology, Northwestern University, Chicago, IL, 60611, USA.
| |
Collapse
|
30
|
Rizzo M. AI in Neurology: Everything, Everywhere, all at Once PART 2: Speech, Sentience, Scruples, and Service. Ann Neurol 2025. [PMID: 40421866 DOI: 10.1002/ana.27229] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2024] [Revised: 02/10/2025] [Accepted: 02/17/2025] [Indexed: 05/28/2025]
Abstract
Artificial intelligence (AI) applications are finding use in real-world neurological settings. Whereas part 1 of this 3-part review series focused on the birth of AI and its foundational principles, this part 2 review shifts gears to explore more practical aspects of neurological care. The review details how large language models, generative AI, and robotics are supporting diagnostic accuracy, patient interaction, and treatment personalization. Special attention is given to ethical and philosophical facets of AI that nonetheless impact practical aspects of care and patient safety, such as accountability for AI-driven decisions and the "black box" nature of many algorithms. We will discuss whether AI systems can develop sentience, and the implications for human-AI collaboration. By examining human-robot interactions in neurology, this part 2 review highlights the profound impact AI could have on patient care and, as covered in the ensuing part 3, on global health care delivery and data analytics, while maintaining ethical oversight and human control. ANN NEUROL 2025.
Collapse
Affiliation(s)
- Matthew Rizzo
- Department of Neurological Sciences, University of Nebraska Medical Center, Omaha, NE
| |
Collapse
|
31
|
Zhou P, Liu Z, Dai J, Yang M, Sui H, Huang Z, Li Y, Song L. KNN algorithm for accurate identification of IFP lesions in the knee joint: a multimodal MRI study. Sci Rep 2025; 15:18163. [PMID: 40414927 DOI: 10.1038/s41598-025-02786-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2024] [Accepted: 05/15/2025] [Indexed: 05/27/2025] Open
Abstract
Knee-related disorders represent a major global health concern and are a leading cause of pain and mobility impairment, particularly in older adults. In clinical medicine, the precise identification and classification of knee joint diseases are essential for early diagnosis and effective treatment. This study presents a novel approach for identifying infrapatellar fat pad (IFP) lesions using the K-Nearest Neighbor (KNN) algorithm in combination with multimodal Magnetic Resonance Imaging (MRI) techniques, specifically mDxion-Quant (mDQ) and T2 mapping (T2m). These imaging methods provide quantitative parameters such as fat fraction (FF), T2*, and T2 values. A set of derived features was constructed through feature engineering to better capture variations within the IFP. These features were used to train the KNN model for classifying knee joint conditions. The proposed method achieved classification accuracies of 94.736% and 92.857% on the training and testing datasets, respectively, outperforming the CNN-Class8 benchmark. This technique holds substantial clinical potential for the early detection of knee joint pathologies, monitoring disease progression, and evaluating post-surgical outcomes.
Collapse
Affiliation(s)
- Peng Zhou
- Department of Imaging, The Affiliated Hospital of Guizhou Medical University, No.28, Beijing Road, Yunyan District, Guiyang, 550000, Guizhou Province, China
- Department of Imaging, Dejiang County People's Hospital, Tongren, 565299, China
| | - Zhenyan Liu
- Department of Imaging, Dejiang County People's Hospital, Tongren, 565299, China
| | - Jiang Dai
- Department of Imaging, Dejiang County People's Hospital, Tongren, 565299, China
| | - Ming Yang
- Department of Imaging, College of Electronic Engineering, Guizhou University, Guiyang, 550025, China
| | - He Sui
- Department of Imaging, The Affiliated Hospital of Guizhou Medical University, No.28, Beijing Road, Yunyan District, Guiyang, 550000, Guizhou Province, China
| | - Zhaoshu Huang
- Department of Imaging, The Affiliated Hospital of Guizhou Medical University, No.28, Beijing Road, Yunyan District, Guiyang, 550000, Guizhou Province, China
| | - Yu Li
- Department of Imaging, The Affiliated Hospital of Guizhou Medical University, No.28, Beijing Road, Yunyan District, Guiyang, 550000, Guizhou Province, China
| | - Lingling Song
- Department of Imaging, The Affiliated Hospital of Guizhou Medical University, No.28, Beijing Road, Yunyan District, Guiyang, 550000, Guizhou Province, China.
| |
Collapse
|
32
|
Kirichenko E, Bachurin S, Lisovin A, Nabiullina R, Kaplya M, Romanov A, Nwosu C, Rodkin S. The Role of Hydrogen Sulfide in the Localization and Structural-Functional Organization of p53 Following Traumatic Brain Injury: Development of a YOLO Model for Detection and Quantification of Apoptotic Nuclei. Int J Mol Sci 2025; 26:5066. [PMID: 40507878 PMCID: PMC12154982 DOI: 10.3390/ijms26115066] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2025] [Revised: 05/17/2025] [Accepted: 05/22/2025] [Indexed: 06/16/2025] Open
Abstract
Traumatic brain injury (TBI) triggers a cascade of molecular and cellular disturbances, including apoptosis, inflammation, and destabilization of neuronal connections. The transcription factor p53 plays a pivotal role in regulating cell fate following brain injury by initiating pro-apoptotic signaling cascades. Hydrogen sulfide (H2S) may significantly contribute to the regulation of p53. Using scanning laser confocal microscopy, we found that after TBI, p53 accumulates extensively in the damaged cerebral cortex, showing distinct subcellular localization in neurons and astrocytes. In neurons, p53 predominantly localizes to the cytoplasm, suggesting involvement in mitochondria-dependent apoptosis, whereas in astrocytes, p53 is found in both the nucleus and cytoplasm, indicating possible activation of transcription-dependent apoptotic pathways. Quantitative analysis confirmed a correlation between p53 localization and morphological signs of cell death, as revealed by Sytox Green and Hoechst nuclear staining. Modulating H2S levels exerted a marked influence on p53 expression and distribution. Administration of the H2S donor sodium thiosulfate (Na2S2O3) reduced the overall number of p53-positive cells, decreased nuclear localization, and lowered the level of apoptosis. Conversely, inhibition of H2S synthesis using aminooxyacetic acid (AOAA) led to enhanced p53 expression, increased numbers of cells exhibiting nuclear fragmentation, and a more pronounced apoptotic response. These findings highlight a neuroprotective role for H2S, likely mediated through the suppression of p53-dependent cell death pathways. To improve analytical accuracy, we developed a YOLO-based deep-learning model for the automated detection of fragmented nuclei. Additionally, evolutionary and molecular dynamics analysis revealed a high degree of p53 conservation among vertebrates and indicated that, although H2S does not form stable complexes with p53, it may modulate its conformational dynamics.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | - Stanislav Rodkin
- Research Laboratory “Medical Digital Images Based on the Basic Model”, Department of Bioengineering, Faculty of Bioengineering and Veterinary Medicine, Don State Technical University, 344000 Rostov-on-Don, Russia
| |
Collapse
|
33
|
Dhiyanesh B, Vijayalakshmi M, Saranya P, Viji D. EnsembleEdgeFusion: advancing semantic segmentation in microvascular decompression imaging with innovative ensemble techniques. Sci Rep 2025; 15:17892. [PMID: 40410312 PMCID: PMC12102392 DOI: 10.1038/s41598-025-02470-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2025] [Accepted: 05/13/2025] [Indexed: 05/25/2025] Open
Abstract
Semantic segmentation involves an imminent part in the investigation of medical images, particularly in the domain of microvascular decompression, where publicly available datasets are scarce, and expert annotation is demanding. In response to this challenge, this study presents a meticulously curated dataset comprising 2003 RGB microvascular decompression images, each intricately paired with annotated masks. Extensive data preprocessing and augmentation strategies were employed to fortify the training dataset, enhancing the robustness of proposed deep learning model. Numerous up-to-date semantic segmentation approaches, including DeepLabv3+, U-Net, DilatedFastFCN with JPU, DANet, and a custom Vanilla architecture, were trained and evaluated using diverse performance metrics. Among these models, DeepLabv3 + emerged as a strong contender, notably excelling in F1 score. Innovatively, ensemble techniques, such as stacking and bagging, were introduced to further elevate segmentation performance. Bagging, notably with the Naïve Bayes approach, exhibited significant improvements, underscoring the potential of ensemble methods in medical image segmentation. The proposed EnsembleEdgeFusion technique exhibited superior loss reduction during training compared to DeepLabv3 + and achieved maximum Mean Intersection over Union (MIoU) scores of 77.73%, surpassing other models. Category-wise analysis affirmed its superiority in accurately delineating various categories within the test dataset.
Collapse
Affiliation(s)
- B Dhiyanesh
- Department of Computer Science and Engineering (ETech), SRM Institute of Science and Technology, Vadapalani Campus, Chennai, Tamil Nadu, India.
| | - M Vijayalakshmi
- Department of Computing Technologies, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| | - P Saranya
- Department of Computational Intelligence, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| | - D Viji
- Department of Computing Technologies, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| |
Collapse
|
34
|
Lo CM, Sung SF. Recognizing artery segments on carotid ultrasonography using embedding concatenation of deep image and vision-language models. Phys Med Biol 2025; 70:115008. [PMID: 40367970 DOI: 10.1088/1361-6560/add8db] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2024] [Accepted: 05/14/2025] [Indexed: 05/16/2025]
Abstract
Objective.Evaluating large artery atherosclerosis is critical for predicting and preventing ischemic strokes. Ultrasonographic assessment of the carotid arteries is the preferred first-line examination due to its ease of use, noninvasive, and absence of radiation exposure. This study proposed an automated classification model for the common carotid artery (CCA), carotid bulb, internal carotid artery (ICA), and external carotid artery (ECA) to enhance the quantification of carotid artery examinations.Approach. A total of 2943 B-mode ultrasound images (CCA: 1563; bulb: 611; ICA: 476; ECA: 293) from 288 patients were collected. Three distinct sets of embedding features were extracted from artificial intelligence networks including pre-trained DenseNet201, vision transformer, and echo contrastive language-image pre-training models using deep learning architectures for pattern recognition. These features were then combined in a support vector machine classifier to interpret the anatomical structures in B-mode images.Main results. After ten-fold cross-validation, the model achieved an accuracy of 82.3%, which was significantly better than using individual feature sets, with ap-value of <0.001.Significance.The proposed model could make carotid artery examinations more accurate and consistent with the achieved classification accuracy. The source code is available athttps://github.com/buddykeywordw/Artery-Segments-Recognition.
Collapse
Affiliation(s)
- Chung-Ming Lo
- Graduate Institute of Library, Information and Archival Studies, National Chengchi University, Taipei, Taiwan
| | - Sheng-Feng Sung
- Division of Neurology, Department of Internal Medicine, Ditmanson Medical Foundation Chia-Yi Christian Hospital, Chiayi City, Taiwan
- Department of Nursing, Fooyin University, Kaohsiung, Taiwan
| |
Collapse
|
35
|
Slika B, Dornaika F, Hammoudi K. Parallel VMamba and Attention-Based Pneumonia Severity Prediction from CXRs: A Robust Model with Segmented Lung Replacement Augmentation. Diagnostics (Basel) 2025; 15:1301. [PMID: 40506873 PMCID: PMC12155560 DOI: 10.3390/diagnostics15111301] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2025] [Revised: 05/14/2025] [Accepted: 05/16/2025] [Indexed: 06/16/2025] Open
Abstract
Background/Objectives: Rapid and accurate assessment of lung diseases, like pneumonia, is critical for effective clinical decision-making, particularly during pandemics when disease progression can be severe. Early diagnosis plays a crucial role in preventing complications, necessitating the development of fast and efficient AI-based models for automated severity assessment. Methods: In this study, we introduce a novel approach that leverages VMamba, a state-of-the-art vision model based on the VisualStateSpace (VSS) framework and 2D-Selective-Scan (SS2D) spatial scanning, to enhance lung severity prediction. Integrated in a parallel multi-image regions approach, VMamba effectively captures global and local contextual features through structured state-space modeling, improving feature representation and robustness in medical image analysis. Additionally, we integrate a segmented lung replacement augmentation strategy to enhance data diversity and improve model generalization. The proposed method is trained on the RALO and COVID-19 datasets and compared against state-of-the-art models. Results: Experimental results demonstrate that our approach achieves superior performance, outperforming existing techniques in prediction accuracy and robustness. Key evaluation metrics, including Mean Absolute Error (MAE) and Pearson Correlation (PC), confirm the model's effectiveness, while the incorporation of segmented lung replacement augmentation further enhances adaptability to diverse lung conditions. Conclusions: These findings highlight the potential of our method for reliable and immediate clinical applications in lung infection assessment.
Collapse
Affiliation(s)
- Bouthaina Slika
- Department of Computer Science and Artificial Intelligence, University of the Basque Country UPV/EHU, 200018 San Sebastian, Spain;
- Faculty of Information Technology, Ho Chi Minh City Open University, Ho Chi Minh City 722000, Vietnam
| | - Fadi Dornaika
- Department of Computer Science and Artificial Intelligence, University of the Basque Country UPV/EHU, 200018 San Sebastian, Spain;
- IKERBASQUE, Basque Foundation for Science, 48009 Bilbao, Spain
| | - Karim Hammoudi
- Department of Computer Science, Institut de Recherche en Informatique, Mathématiques, Automatique et Signal, Université de Haute-Alsace, 68093 Mulhouse, France;
| |
Collapse
|
36
|
Larroza A, Pérez-Benito FJ, Tendero R, Perez-Cortes JC, Román M, Llobet R. Three-Blind Validation Strategy of Deep Learning Models for Image Segmentation. J Imaging 2025; 11:170. [PMID: 40423027 DOI: 10.3390/jimaging11050170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2025] [Revised: 05/07/2025] [Accepted: 05/19/2025] [Indexed: 05/28/2025] Open
Abstract
Image segmentation plays a central role in computer vision applications such as medical imaging, industrial inspection, and environmental monitoring. However, evaluating segmentation performance can be particularly challenging when ground truth is not clearly defined, as is often the case in tasks involving subjective interpretation. These challenges are amplified by inter- and intra-observer variability, which complicates the use of human annotations as a reliable reference. To address this, we propose a novel validation framework-referred to as the three-blind validation strategy-that enables rigorous assessment of segmentation models in contexts where subjectivity and label variability are significant. The core idea is to have a third independent expert, blind to the labeler identities, assess a shuffled set of segmentations produced by multiple human annotators and/or automated models. This allows for the unbiased evaluation of model performance and helps uncover patterns of disagreement that may indicate systematic issues with either human or machine annotations. The primary objective of this study is to introduce and demonstrate this validation strategy as a generalizable framework for robust model evaluation in subjective segmentation tasks. We illustrate its practical implementation in a mammography use case involving dense tissue segmentation while emphasizing its potential applicability to a broad range of segmentation scenarios.
Collapse
Affiliation(s)
- Andrés Larroza
- Instituto Tecnológico de la Informática, Universitat Politècnica de València, Camino de Vera, s/n, 46022 València, Spain
| | - Francisco Javier Pérez-Benito
- Instituto Tecnológico de la Informática, Universitat Politècnica de València, Camino de Vera, s/n, 46022 València, Spain
| | - Raquel Tendero
- Instituto Tecnológico de la Informática, Universitat Politècnica de València, Camino de Vera, s/n, 46022 València, Spain
| | - Juan Carlos Perez-Cortes
- Instituto Tecnológico de la Informática, Universitat Politècnica de València, Camino de Vera, s/n, 46022 València, Spain
| | - Marta Román
- Department of Epidemiology and Evaluation, IMIM (Hospital del Mar Research Institute), Passeig Marítim 25-29, 08003 Barcelona, Spain
| | - Rafael Llobet
- Instituto Tecnológico de la Informática, Universitat Politècnica de València, Camino de Vera, s/n, 46022 València, Spain
| |
Collapse
|
37
|
Amiot V, Jimenez-Del-Toro O, Guex-Crosier Y, Ott M, Bogaciu TE, Banerjee S, Howell J, Amstutz C, Chiquet C, Bergin C, Meloni I, Tomasoni M, Hoogewoud F, Anjos A. Automatic transformer-based grading of multiple retinal inflammatory signs in uveitis on fluorescein angiography. Comput Biol Med 2025; 193:110327. [PMID: 40403640 DOI: 10.1016/j.compbiomed.2025.110327] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2025] [Revised: 04/04/2025] [Accepted: 04/30/2025] [Indexed: 05/24/2025]
Abstract
BACKGROUND Grading fluorescein angiography (FA) for uveitis is complex, often leading to the oversight of retinal inflammation in clinical studies. This study aims to develop an automated method for grading retinal inflammation. METHODS Patients from Jules-Gonin Eye Hospital with active or resolved uveitis who underwent FA between 2018 and 2021 were included. FAs were acquired using a standardized protocol, anonymized, and annotated following the Angiography Scoring for Uveitis Working Group criteria, for four inflammatory signs of the posterior pole. Intergrader agreement was assessed by four independent graders. Four deep learning transformer models were developed, and performance was evaluated using the Ordinal Classification Index, accuracy, F1 scores, and Kappa scores. Saliency analysis was employed to visualize model predictions. FINDINGS A total of 543 patients (1042 eyes, 40987 images) were included in the study. The models closely matched expert graders in detecting vascular leakage (F1-score = 0·87, 1-OCI = 0·89), capillary leakage (F1-score = 0·86, 1-OCI = 0·89), macular edema (F1-score = 0·82, 1-OCI = 0·86), and optic disc hyperfluorescence (F1-score = 0·72, 1-OCI = 0·85). Saliency analysis confirmed that the models focused on relevant retinal structures. The mean intergrader agreement across all inflammatory signs was F1-score = 0·79 and 1-OCI = 0·83. INTERPRETATION We developed a vision transformer-based model for the automatic grading of retinal inflammation in uveitis, utilizing the largest dataset of FAs in uveitis to date. This approach provides significant clinical benefits for the evaluation of uveitis and paves the way for future advancements, including the identification of novel biomarkers through the integration of clinical data and other modalities.
Collapse
Affiliation(s)
- Victor Amiot
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland; Platform for Research in Ocular Imaging, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland
| | | | - Yan Guex-Crosier
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland; Ocular Immune-infectiology Unit Jules-Gonin Eye Hospital, FAA, University of Lausanne, Switzerland
| | - Muriel Ott
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland; Ocular Immune-infectiology Unit Jules-Gonin Eye Hospital, FAA, University of Lausanne, Switzerland
| | - Teodora-Elena Bogaciu
- Grenoble Alpes University, Grenoble, France; Department of Ophthalmology, Grenoble Alpes University Hospital, Grenoble, France
| | - Shalini Banerjee
- Department of Ophthalmology, Cantonal Hospital Lucerne, Lucerne, Switzerland; Faculty of Health Sciences and Medicine, University of Lucerne, Lucerne, Switzerland
| | - Jeremy Howell
- Department of Ophthalmology, Cantonal Hospital Lucerne, Lucerne, Switzerland; Faculty of Health Sciences and Medicine, University of Lucerne, Lucerne, Switzerland
| | - Christoph Amstutz
- Department of Ophthalmology, Cantonal Hospital Lucerne, Lucerne, Switzerland; Faculty of Health Sciences and Medicine, University of Lucerne, Lucerne, Switzerland
| | - Christophe Chiquet
- Grenoble Alpes University, Grenoble, France; Department of Ophthalmology, Grenoble Alpes University Hospital, Grenoble, France
| | - Ciara Bergin
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland
| | - Ilenia Meloni
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland; Platform for Research in Ocular Imaging, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland
| | - Mattia Tomasoni
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland; Platform for Research in Ocular Imaging, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland
| | - Florence Hoogewoud
- Department of Ophthalmology, University of Lausanne, Fondation Asile des Aveugles, Jules Gonin Eye Hospital, Lausanne, Switzerland; Ocular Immune-infectiology Unit Jules-Gonin Eye Hospital, FAA, University of Lausanne, Switzerland
| | - André Anjos
- Idiap Research Institute, Martigny, Switzerland.
| |
Collapse
|
38
|
Shephard AJ, Mahmood H, Raza SEA, Araújo ALD, Santos-Silva AR, Lopes MA, Vargas PA, McCombe KD, Craig SG, James J, Brooks J, Nankivell P, Mehanna H, Khurram SA, Rajpoot NM. Development and validation of an artificial intelligence-based pipeline for predicting oral epithelial dysplasia malignant transformation. COMMUNICATIONS MEDICINE 2025; 5:186. [PMID: 40394272 DOI: 10.1038/s43856-025-00873-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Accepted: 04/14/2025] [Indexed: 05/22/2025] Open
Abstract
BACKGROUND Oral epithelial dysplasia (OED) is a potentially malignant histopathological diagnosis given to lesions of the oral cavity that are at risk of progression to malignancy. Manual grading of OED is subject to substantial variability and does not reliably predict prognosis, potentially resulting in sub-optimal treatment decisions. METHOD We developed a Transformer-based artificial intelligence (AI) pipeline for the prediction of malignant transformation from whole-slide images (WSIs) of Haematoxylin and Eosin (H&E) stained OED tissue slides, named ODYN (Oral Dysplasia Network). ODYN can simultaneously classify OED and assign a predictive score (ODYN-score) to quantify the risk of malignant transformation. The model was trained on a large cohort using three different scanners (Sheffield, 358 OED WSIs, 105 control WSIs) and externally validated on cases from three independent centres (Birmingham and Belfast, UK, and Piracicaba, Brazil; 108 OED WSIs). RESULTS Model testing yielded an F1-score of 0.96 for classification of dysplastic vs non-dysplastic slides, and an AUROC of 0.73 for malignancy prediction, gaining comparable results to clinical grading systems. CONCLUSIONS With further large-scale prospective validation, ODYN promises to offer an objective and reliable solution for assessing OED cases, ultimately improving early detection and treatment of oral cancer.
Collapse
Affiliation(s)
- Adam J Shephard
- Tissue Image Analytics Centre, Department of Computer Science, University of Warwick, Coventry, UK
| | - Hanya Mahmood
- School of Clinical Dentistry, University of Sheffield, Sheffield, UK
| | - Shan E Ahmed Raza
- Tissue Image Analytics Centre, Department of Computer Science, University of Warwick, Coventry, UK
| | - Anna Luíza Damaceno Araújo
- Head and Neck Surgery Department and LIM 28, University of São Paulo Medical School, São Paulo, State of São Paulo, Brazil
- Hospital Israelita Albert Einstein, São Paulo, State of São Paulo, Brazil
| | - Alan Roger Santos-Silva
- Faculdade de Odontologia de Piracicaba, Universidade Estadual de Campinas (FOP-UNICAMP), Piracicaba, State of São Paulo, Brazil
| | - Marcio Ajudarte Lopes
- Faculdade de Odontologia de Piracicaba, Universidade Estadual de Campinas (FOP-UNICAMP), Piracicaba, State of São Paulo, Brazil
| | - Pablo Agustin Vargas
- Faculdade de Odontologia de Piracicaba, Universidade Estadual de Campinas (FOP-UNICAMP), Piracicaba, State of São Paulo, Brazil
| | - Kris D McCombe
- Precision Medicine Centre, Patrick G. Johnston Centre for Cancer Research, Queen's University Belfast, Belfast, UK
| | - Stephanie G Craig
- Precision Medicine Centre, Patrick G. Johnston Centre for Cancer Research, Queen's University Belfast, Belfast, UK
| | - Jacqueline James
- Precision Medicine Centre, Patrick G. Johnston Centre for Cancer Research, Queen's University Belfast, Belfast, UK
| | - Jill Brooks
- Institute of Head and Neck Studies and Education, Institute of Cancer and Genomic Sciences, University of Birmingham, Birmingham, UK
| | - Paul Nankivell
- Institute of Head and Neck Studies and Education, Institute of Cancer and Genomic Sciences, University of Birmingham, Birmingham, UK
| | - Hisham Mehanna
- Institute of Head and Neck Studies and Education, Institute of Cancer and Genomic Sciences, University of Birmingham, Birmingham, UK
| | - Syed Ali Khurram
- School of Clinical Dentistry, University of Sheffield, Sheffield, UK
| | - Nasir M Rajpoot
- Tissue Image Analytics Centre, Department of Computer Science, University of Warwick, Coventry, UK.
| |
Collapse
|
39
|
Zeng P, Song R, Chen S, Li X, Li H, Chen Y, Gong Z, Cai G, Lin Y, Shi M, Huang K, Chen Z. Expert-guided StyleGAN2 image generation elevates AI diagnostic accuracy for maxillary sinus lesions. COMMUNICATIONS MEDICINE 2025; 5:185. [PMID: 40394291 PMCID: PMC12092618 DOI: 10.1038/s43856-025-00907-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2024] [Accepted: 05/12/2025] [Indexed: 05/22/2025] Open
Abstract
BACKGROUND The progress of artificial intelligence (AI) research in dental medicine is hindered by data acquisition challenges and imbalanced distributions. These problems are especially apparent when planning to develop AI-based diagnostic or analytic tools for various lesions, such as maxillary sinus lesions (MSL) including mucosal thickening and polypoid lesions. Traditional unsupervised generative models struggle to simultaneously control the image realism, diversity, and lesion-type specificity. This study establishes an expert-guided framework to overcome these limitations to elevate AI-based diagnostic accuracy. METHODS A StyleGAN2 framework was developed for generating clinically relevant MSL images (such as mucosal thickening and polypoid lesion) under expert control. The generated images were then integrated into training datasets to evaluate their effect on ResNet50's diagnostic performance. RESULTS Here we show: 1) Both lesion subtypes achieve satisfactory fidelity metrics, with structural similarity indices (SSIM > 0.996) and maximum mean discrepancy values (MMD < 0.032), and clinical validation scores close to those of real images; 2) Integrating baseline datasets with synthetic images significantly enhances diagnostic accuracy for both internal and external test sets, particularly improving area under the precision-recall curve (AUPRC) by approximately 8% and 14% for mucosal thickening and polypoid lesions in the internal test set, respectively. CONCLUSIONS The StyleGAN2-based image generation tool effectively addressed data scarcity and imbalance through high-quality MSL image synthesis, consequently boosting diagnostic model performance. This work not only facilitates AI-assisted preoperative assessment for maxillary sinus lift procedures but also establishes a methodological framework for overcoming data limitations in medical image analysis.
Collapse
Affiliation(s)
- Peisheng Zeng
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China
| | - Rihui Song
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong, China
| | - Shijie Chen
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China
| | - Xiaohang Li
- School of Mechanical and Automation Engineering, Wuyi University, Jiangmen, Guangdong, China
| | - Haopeng Li
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong, China
| | - Yue Chen
- School of Information Technology, Guangdong Industry Polytechnic University, Foshan, Guangdong, China
| | - Zhuohong Gong
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China
| | - Gengbin Cai
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China
| | - Yixiong Lin
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China
| | - Mengru Shi
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China.
| | - Kai Huang
- School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, Guangdong, China.
| | - Zetao Chen
- Hospital of Stomatology, Guanghua School of Stomatology, Sun Yat-sen University and Guangdong Provincial Key Laboratory of Stomatology, Guangzhou, Guangdong, China.
| |
Collapse
|
40
|
Yan L, Xu J, Ye X, Lin M, Gong Y, Fang Y, Chen S. Development and validation of ultrasound-based radiomics deep learning model to identify bone erosion in rheumatoid arthritis. Clin Rheumatol 2025:10.1007/s10067-025-07481-1. [PMID: 40389785 DOI: 10.1007/s10067-025-07481-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2025] [Revised: 04/05/2025] [Accepted: 05/04/2025] [Indexed: 05/21/2025]
Abstract
OBJECTIVE To develop and validate a deep learning radiomics fusion model (DLR) based on ultrasound (US) images to identify bone erosion in rheumatoid arthritis (RA) patients. METHODS A total of 432 patients with RA at two institutions were collected. Three hundred twelve patients from center 1 were randomly divided into a training set (N = 218) and an internal test set (N = 94) in a 7:3 ratio; meanwhile, 124 patients from center 2 were as an external test set. Radiomics (Rad) and deep learning (DL) features were extracted based on hand-crafted radiomics and deep transfer learning networks. The least absolute shrinkage and selection operator regression was employed to establish DLR fusion feature from the Rad and DL features. Subsequently, 10 machine learning algorithms were used to construct models and the final optimal model was selected. The performance of models was evaluated using receiver operating characteristic (ROC) and decision curve analysis (DCA). The diagnostic efficacy of sonographers was compared with and without the assistance of the optimal model. RESULTS LR was chosen as the optimal algorithm for model construction account for superior performance (Rad/DL/DLR: area under the curve [AUC] = 0.906/0.974/0.979) in the training set. In the internal test set, DLR_LR as the final model had the highest AUC (AUC = 0.966), which was also validated in the external test set (AUC = 0.932). With the aid of DLR_LR model, the overall performance of both junior and senior sonographers improved significantly (P < 0.05), and there was no significant difference between the junior sonographer with DLR_LR model assistance and the senior sonographer without assistance (P > 0.05). CONCLUSION DLR model based on US images is the best performer and is expected to become an important tool for identifying bone erosion in RA patients. Key Points • DLR model based on US images is the best performer in identifying BE in RA patients. • DLR model may assist the sonographers to improve the accuracy of BE evaluations.
Collapse
Affiliation(s)
- Lei Yan
- Department of Ultrasound, The First Affiliated Hospital of Fujian Medical University, 20# Chazhong Road, Fuzhou, 350005, Fujian, China
- Department of Ultrasound, National Regional Medical Center, First Affiliated Hospital of Fujian Medical University Binhai Campus, Fuzhou, China
| | - Jing Xu
- Department of Ultrasound, The First Affiliated Hospital of Fujian Medical University, 20# Chazhong Road, Fuzhou, 350005, Fujian, China
- Department of Ultrasound, National Regional Medical Center, First Affiliated Hospital of Fujian Medical University Binhai Campus, Fuzhou, China
| | - Xiaojian Ye
- Department of Ultrasound, The First Affiliated Hospital of Fujian Medical University, 20# Chazhong Road, Fuzhou, 350005, Fujian, China
- Department of Ultrasound, National Regional Medical Center, First Affiliated Hospital of Fujian Medical University Binhai Campus, Fuzhou, China
| | - Minghang Lin
- Department of Ultrasound, Fuzhou University Affiliated Provincial Hospital, Fuzhou, China
- Department of Ultrasound, Fuqing City Hospital, Fujian Medical University, Fuzhou, China
| | - Yiran Gong
- Department of Ultrasound, The First Affiliated Hospital of Fujian Medical University, 20# Chazhong Road, Fuzhou, 350005, Fujian, China
- Department of Ultrasound, National Regional Medical Center, First Affiliated Hospital of Fujian Medical University Binhai Campus, Fuzhou, China
| | - Yabin Fang
- Department of Ultrasound, The First Affiliated Hospital of Fujian Medical University, 20# Chazhong Road, Fuzhou, 350005, Fujian, China
- Department of Ultrasound, National Regional Medical Center, First Affiliated Hospital of Fujian Medical University Binhai Campus, Fuzhou, China
| | - Shuqiang Chen
- Department of Ultrasound, The First Affiliated Hospital of Fujian Medical University, 20# Chazhong Road, Fuzhou, 350005, Fujian, China.
- Department of Ultrasound, National Regional Medical Center, First Affiliated Hospital of Fujian Medical University Binhai Campus, Fuzhou, China.
- Department of Ultrasound, Fuzhou University Affiliated Provincial Hospital, Fuzhou, China.
| |
Collapse
|
41
|
Nambiar R, Bhat R, Achar H V B. Advancements in Hematologic Malignancy Detection: A Comprehensive Survey of Methodologies and Emerging Trends. ScientificWorldJournal 2025; 2025:1671766. [PMID: 40421320 PMCID: PMC12103971 DOI: 10.1155/tswj/1671766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2024] [Accepted: 04/24/2025] [Indexed: 05/28/2025] Open
Abstract
The investigation and diagnosis of hematologic malignancy using blood cell image analysis are major and emerging subjects that lie at the intersection of artificial intelligence and medical research. This survey systematically examines the state-of-the-art in blood cancer detection through image-based analysis, aimed at identifying the most effective computational strategies and highlighting emerging trends. This review focuses on three principal objectives, namely, to categorize and compare traditional machine learning (ML), deep learning (DL), and hybrid learning approaches; to evaluate performance metrics such as accuracy, precision, recall, and area under the ROC curve; and to identify methodological gaps and propose directions for future research. Methodologically, we organize the literature by categorizing the malignancy types-leukemia, lymphoma, and multiple myeloma-and particularizing the preprocessing steps, feature extraction techniques, network architectures, and ensemble strategies employed. For ML methods, we discuss classical classifiers including support vector machines and random forests; for DL, we analyze convolutional neural networks (e.g., AlexNet, VGG, and ResNet) and transformer-based models; and for hybrid systems, we examine combinations of CNNs with attention mechanisms or traditional classifiers. Our synthesis reveals that DL models consistently outperform ML baselines, achieving classification accuracies above 95% in benchmark datasets, with hybrid models pushing peak accuracy to 99.7%. However, challenges remain in data scarcity, class imbalance, and generalizability to clinical settings. We conclude by recommending the integration of multimodal data, semisupervised learning, and rigorous external validation to advance toward deployable diagnostic tools. This survey also provides a comprehensive roadmap for researchers and clinicians striving to harness AI for reliable hematologic cancer detection.
Collapse
Affiliation(s)
- Rajashree Nambiar
- Department of Robotics and AI Engineering, NMAM Institute of Technology, NITTE (Deemed to be University), Nitte, India
| | - Ranjith Bhat
- Department of Robotics and AI Engineering, NMAM Institute of Technology, NITTE (Deemed to be University), Nitte, India
| | - Balachandra Achar H V
- Department of Electronics and Communication Engineering, Manipal Institute of Technology Bengaluru, Manipal Academy of Higher Education, Manipal, India
| |
Collapse
|
42
|
Lin YT, Xiong X, Zheng YP, Zhou Q. Transfer Learning and Multi-Feature Fusion-Based Deep Learning Model for Idiopathic Macular Hole Diagnosis and Grading from Optical Coherence Tomography Images. Clin Ophthalmol 2025; 19:1593-1607. [PMID: 40396157 PMCID: PMC12091069 DOI: 10.2147/opth.s521558] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2025] [Accepted: 05/05/2025] [Indexed: 05/22/2025] Open
Abstract
Background Idiopathic macular hole is an ophthalmic disease that seriously affects vision, and its early diagnosis and treatment have important clinical significance to reduce the occurrence of blindness. At present, OCT is the gold standard for diagnosing this disease, but its application is limited due to the need for professional ophthalmologist to diagnose it. The introduction of artificial intelligence will break this situation and make its diagnosis efficient, and how to build an effective predictive model is the key to the problem, and more clinical trials are still needed to verify it. Objective This study aims to evaluate the role of deep learning systems in Idiopathic Macular Hole diagnosis, grading, and prediction. Methods A single-center, retrospective study used binocular OCT images from IMH patients at the First Affiliated Hospital of Nanchang University (November 2019 - January 2023). A deep learning algorithm, including traditional omics, Resnet101, and fusion models incorporating multi-feature fusion and transfer learning, was developed. Model performance was evaluated using accuracy and AUC. Logistic regression analyzed clinical factors, and a nomogram predicted surgical risk. Analysis was conducted with SPSS 22.0 and R 3.6.3. P < 0.05 was statistically significant. Results Among 229 OCT images, the traditional omics, Resnet101, and fusion models achieved accuracies of 93%, 94%, and 95%, respectively, in the training set. In the test set, the fusion model and Resnet101 correctly identified 39 images, while the traditional omics model identified 35. The nomogram had a C-index of 0.996, with macular hole diameter most strongly associated with surgical risk. Conclusion The deep learning system with transfer learning and multi-feature fusion effectively diagnoses and grades IMH from OCT images.
Collapse
Affiliation(s)
- Ye-Ting Lin
- Department of Ophthalmology, The First Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, Jiangxi, People’s Republic of China
| | - Xu Xiong
- Department of Ophthalmology, The First Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, Jiangxi, People’s Republic of China
| | - Ying-Ping Zheng
- Department of Product Design, Jiangxi Normal University, Nanchang, Jiangxi, People’s Republic of China
| | - Qiong Zhou
- Department of Ophthalmology, The First Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, Jiangxi, People’s Republic of China
| |
Collapse
|
43
|
Baker J, Elliott C, Boden A, Antypas A, Singh S, Aggarwal P, Jayasinghe N, Narasimhan P. What are the perceptions of AI in radiology among UK medical students and junior doctors? Acta Radiol 2025:2841851251339010. [PMID: 40375780 DOI: 10.1177/02841851251339010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/18/2025]
Abstract
BackgroundThe integration of artificial intelligence (AI) in radiology has the potential to improve diagnostic accuracy and efficiency. Medical students and junior doctors will likely use AI more frequently in the future, making their perceptions essential for identifying educational gaps.PurposeTo explore the perceptions of UK medical students and junior doctors regarding AI in radiology.Material and MethodsA cross-sectional survey was distributed across UK medical schools and foundation programs. A total of 250 responses were analyzed using descriptive statistics and non-parametric tests, focusing on career impact, clinical effectiveness, educational development, and ethical concerns.ResultsMost respondents (55.2%) were undeterred by career uncertainties related to AI, with 64% confident that AI would not replace radiologists. Up to 80.6% supported AI's clinical benefits, and 63.2% endorsed its educational integration. However, there were concerns about job displacement and insufficient AI training. Medical students were more worried about job security than junior doctors, while those committed to radiology were less apprehensive and viewed AI as complementary.ConclusionEducational programs and regulatory frameworks are essential to facilitate AI integration in radiology. Addressing concerns about job displacement and improving AI education will be key to preparing future radiologists for technological advancements.
Collapse
Affiliation(s)
- James Baker
- Radiology Department, Harrogate and District NHS Foundation Trust, Harrogate, UK
| | - Charlotte Elliott
- Radiology Department, Harrogate and District NHS Foundation Trust, Harrogate, UK
| | - Alexander Boden
- Norwich Medical School, University of East Anglia Norwich, Norwich, GBR
| | - Antony Antypas
- School of Biomedical Sciences, Faculty of Biological Sciences, University of Leeds, Leeds, UK
| | - Shwetabh Singh
- School of Biomedical Sciences, Faculty of Biological Sciences, University of Leeds, Leeds, UK
| | - Prashant Aggarwal
- Urology Department, Oxford University Hospitals NHS Foundation Trust, Thames Valley, UK
- Nuffield Department of Surgical Sciences, University of Oxford, Oxford, UK
| | | | - Padmanesan Narasimhan
- School of Population Health, Faculty of Medicine and Health Sciences, University of New South Wales, Sydney, NSW, Australia
| |
Collapse
|
44
|
Do W, van Nistelrooij N, Bergé S, Vinayahalingam S. Enhancing Craniomaxillofacial Surgeries with Artificial Intelligence Technologies. Oral Maxillofac Surg Clin North Am 2025:S1042-3699(25)00017-2. [PMID: 40382287 DOI: 10.1016/j.coms.2025.04.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/20/2025]
Abstract
Artificial intelligence (AI) can be applied in multiple subspecialties in craniomaxillofacial (CMF) surgeries. This article overviews AI fundamentals focusing on classification, object detection, and segmentation-core tasks used in CMF applications. The article then explores the development and integration of AI in dentoalveolar surgery, implantology, traumatology, oncology, craniofacial surgery, and orthognathic and feminization surgery. It highlights AI-driven advancements in diagnosis, pre-operative planning, intra-operative assistance, post-operative management, and outcome prediction. Finally, the challenges in AI adoption are discussed, including data limitations, algorithm validation, and clinical integration.
Collapse
Affiliation(s)
- Wesley Do
- Department of Oral & Maxillofacial Surgery, Radboud UMC, Philips van Leydenlaan 25, 6525 EX Nijmegen, The Netherlands
| | - Niels van Nistelrooij
- Department of Oral and Maxillofacial Surgery, Charité - Universitätsmedizin Berlin, corporate member of Freie Universität Berlin and Humboldt- Universität zu Berlin, Hindenburgdamm 30, 12203, Berlin, Germany
| | - Stefaan Bergé
- Department of Oral & Maxillofacial Surgery, Radboud UMC, Philips van Leydenlaan 25, 6525 EX Nijmegen, The Netherlands.
| | - Shankeeth Vinayahalingam
- Department of Oral & Maxillofacial Surgery, Radboud UMC, Philips van Leydenlaan 25, 6525 EX Nijmegen, The Netherlands
| |
Collapse
|
45
|
Galán J, González M, Moral P, García-Martín Á, Martínez JM. Transforming urban waste collection inventory: AI-Based container classification and Re-Identification. WASTE MANAGEMENT (NEW YORK, N.Y.) 2025; 199:25-35. [PMID: 40081303 DOI: 10.1016/j.wasman.2025.02.051] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Revised: 02/05/2025] [Accepted: 02/27/2025] [Indexed: 03/16/2025]
Abstract
This work lays the groundwork for creating an automated system for the inventory of urban waste elements. Our primary contribution is the development of, to the best of our knowledge, the first re-identification system for urban waste elements that uses Artificial Intelligence and Computer Vision, incorporating information from a classification module and geolocation context to enhance post-processing performance. This re-identification system helps to create and update inventories by determining if a new image matches an existing element in the inventory based on visual similarity or, if not, by adding it as a new identity (new class or new identity of an existing class). Such a system could be highly valuable to local authorities and waste management companies, offering improved facility maintenance, geolocation, and additional applications. This work also addresses the dynamic nature of urban environments and waste management elements by exploring Continual Learning strategies to adapt pretrained systems to new settings with different urban elements. Experimental results show that the proposed system operates effectively across various container types and city layouts. These findings were validated through testing in two different Spanish locations, a "City" and a "Campus", differing in size, illumination conditions, seasons, urban design and container appearance. For the final re-identification system, the baseline system achieves 53.18 mAP (mean Average Precision) in the simple scenario, compared to 21.54 mAP in the complex scenario, with additional challenging unseen variability. Incorporating the proposed post-processing techniques significantly improved results, reaching 74.14 mAP and 71.75 mAP in the simple and complex scenario respectively.
Collapse
Affiliation(s)
- Javier Galán
- Video Processing and Understanding Lab, Universidad Aut́onoma de Madrid 28049 Madrid, Spain
| | - Miguel González
- Video Processing and Understanding Lab, Universidad Aut́onoma de Madrid 28049 Madrid, Spain
| | - Paula Moral
- Video Processing and Understanding Lab, Universidad Aut́onoma de Madrid 28049 Madrid, Spain
| | - Álvaro García-Martín
- Video Processing and Understanding Lab, Universidad Aut́onoma de Madrid 28049 Madrid, Spain.
| | - José M Martínez
- Video Processing and Understanding Lab, Universidad Aut́onoma de Madrid 28049 Madrid, Spain
| |
Collapse
|
46
|
Batista S, Couceiro M, Filipe R, Rachinhas P, Isidoro J, Domingues I. Private Data Incrementalization: Data-Centric Model Development for Clinical Liver Segmentation. Bioengineering (Basel) 2025; 12:530. [PMID: 40428149 PMCID: PMC12108710 DOI: 10.3390/bioengineering12050530] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2025] [Revised: 05/09/2025] [Accepted: 05/13/2025] [Indexed: 05/29/2025] Open
Abstract
Machine Learning models, more specifically Artificial Neural Networks, are transforming medical imaging by enabling precise liver segmentation, a crucial task for diagnosing and treating liver diseases. However, these models often face challenges in adapting to diverse clinical data sources as differences in dataset volume, resolution, and origin impact generalization and performance. This study introduces a Private Data Incrementalization, a data-centric approach to enhance the adaptability of Artificial Neural Networks by progressively exposing them to varied clinical data. As the target of this study is not to propose a new image segmentation model, the existing medical imaging segmentation models-including U-Net, ResUNet++, Fully Convolutional Network, and a modified algorithm based on the Conditional Bernoulli Diffusion Model-are used. The study evaluates these four models using a curated private dataset of computed tomography scans from Coimbra University Hospital, supplemented by two public datasets, 3D-IRCADb01 and CHAOS. The Private Data Incrementalization method systematically increases the volume and diversity of training data, simulating real-world conditions where models must handle varied imaging contexts. Pre-processing and post-processing stages, incremental training, and performance evaluations reveal that structured exposure to diverse datasets improves segmentation performance, with ResUNet++ achieving the highest accuracy (0.9972) and Dice Similarity Coefficient (0.9449), and the best Average Symmetric Surface Distance (0.0053 mm), demonstrating the importance of dataset diversity and volume for segmentation models' robustness and generalization. Private Data Incrementalization thus offers a scalable strategy for building resilient segmentation models, ultimately benefiting clinical workflows, patient care, and healthcare resource management by addressing the variability inherent in clinical imaging data.
Collapse
Affiliation(s)
- Stephanie Batista
- Polytechnic University of Coimbra, Rua da Misericórdia, Lagar dos Cortiços, S. Martinho do Bispo, 3045-093 Coimbra, Portugal (M.C.)
| | - Miguel Couceiro
- Polytechnic University of Coimbra, Rua da Misericórdia, Lagar dos Cortiços, S. Martinho do Bispo, 3045-093 Coimbra, Portugal (M.C.)
- Institute of Applied Research (i2A), 3045-093 Coimbra, Portugal
- Laboratory for High Performance Computing (LaCED), 3030-199 Coimbra, Portugal
- Laboratory of Instrumentation and Experimental Particle Physics (LIP-Coimbra), Rua Larga da Universidade de Coimbra, 3004-516 Coimbra, Portugal
| | | | - Paulo Rachinhas
- Coimbra Hospital and University Center, 3004-561 Coimbra, Portugal; (P.R.); (J.I.)
| | - Jorge Isidoro
- Coimbra Hospital and University Center, 3004-561 Coimbra, Portugal; (P.R.); (J.I.)
| | - Inês Domingues
- Polytechnic University of Coimbra, Rua da Misericórdia, Lagar dos Cortiços, S. Martinho do Bispo, 3045-093 Coimbra, Portugal (M.C.)
- Medical Physics, Radiobiology and Radiological Protection Group, Research Centre of the Portuguese Institute of Oncology of Porto (CI-IPOP), 4200-072 Porto, Portugal
| |
Collapse
|
47
|
Zhang J, Huang YS, Wang YW, Xiang H, Lin X, Chang RF. Automated whole-breast ultrasound tumor diagnosis using attention-inception network. Phys Med 2025; 134:104989. [PMID: 40373703 DOI: 10.1016/j.ejmp.2025.104989] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/28/2024] [Revised: 04/11/2025] [Accepted: 04/29/2025] [Indexed: 05/17/2025] Open
Abstract
PURPOSE Automated Whole-Breast Ultrasound (ABUS) has been widely used as an important tool in breast cancer diagnosis due to the ability of this technique to provide complete three-dimensional (3D) images of breasts. To eliminate the risk of misdiagnosis, computer-aided diagnosis (CADx) systems have been proposed to assist radiologists. Convolutional neural networks (CNNs), renowned for the automatic feature extraction capabilities, have developed rapidly in medical image analysis, and this study proposes a CADx system based on 3D CNN for ABUS. MATERIALS AND METHODS This study used a private dataset collected at Sun Yat-Sen University Cancer Center (SYSUCC) from 396 breast tumor patients. First, the tumor volume of interest (VOI) was extracted and resized, and then the tumor was enhanced by histogram equalization. Second, a 3D U-Net++ was employed to segment the tumor mask. Finally, the VOI, the enhanced VOI, and the corresponding tumor mask were fed into a 3D Attention-Inception network to classify the tumor as benign or malignant. RESULTS The experiment results indicate an accuracy of 89.4%, a sensitivity of 91.2%, a specificity of 87.6%, and an area under the receiver operating characteristic curve (AUC) of 0.9262, which suggests that the proposed CADx system for ABUS images rivals the performance of experienced radiologists in tumor diagnosis tasks. CONCLUSION This study proposes a CADx system consisting of a 3D U-Net++ tumor segmentation model and a 3D attention inception neural network tumor classification model for diagnosis in ABUS images. The results indicate that the proposed CADx system is effective and efficient in tumor diagnosis tasks.
Collapse
Affiliation(s)
- Jun Zhang
- Graduate Institute of Biomedical Electronics and Bioinformatics, National Taiwan University, Taipei, Taiwan
| | - Yao-Sian Huang
- Department of Computer Science and Information Engineering, National Changhua University of Education, Changhua, Taiwan
| | - You-Wei Wang
- Department of Computer Science and Information Engineering, National Taiwan University, Taipei, Taiwan
| | - Huiling Xiang
- Department of Ultrasound, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Xi Lin
- Department of Ultrasound, Sun Yat-sen University Cancer Center, Guangzhou, China.
| | - Ruey-Feng Chang
- Graduate Institute of Biomedical Electronics and Bioinformatics, National Taiwan University, Taipei, Taiwan; Department of Computer Science and Information Engineering, National Taiwan University, Taipei, Taiwan; Graduate Institute of Network and Multimedia, National Taiwan University, Taipei, Taiwan.
| |
Collapse
|
48
|
Del Amor R, López-Pérez M, Meseguer P, Morales S, Terradez L, Aneiros-Fernandez J, Mateos J, Molina R, Naranjo V. A fusocelular skin dataset with whole slide images for deep learning models. Sci Data 2025; 12:788. [PMID: 40368949 PMCID: PMC12078617 DOI: 10.1038/s41597-025-05108-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2024] [Accepted: 04/28/2025] [Indexed: 05/16/2025] Open
Abstract
Cutaneous spindle cell (CSC) lesions encompass a spectrum from benign to malignant neoplasms, often posing significant diagnostic challenges. Computer-aided diagnosis systems offer a promising solution to make pathologists' decisions objective and faster. These systems usually require large-scale datasets with curated labels for effective training; however, manual annotation is time-consuming and expensive. To overcome this challenge, crowdsourcing has emerged as a popular and valuable strategy to scale up the labeling process by distributing the effort among different non-expert annotators. This work introduces AI4SkIN, the first public dataset Whole Slide Images (WSIs) for CSC neoplasms, annotated using an innovative crowdsourcing protocol. AI4SkIN dataset contains 641 Hematoxylin and Eosin stained WSIs with multiclass labels from both expert and trainee pathologists. The dataset improves CSC neoplasm diagnosis using advanced machine learning and crowdsourcing based on Gaussian Processes, showing that models trained on non-expert labels perform comparably to those using expert labels. In conclusion, we illustrate that AI4SkIN provides a good resource for developing and validating methods for multiclass CSC neoplasm classification.
Collapse
Affiliation(s)
- Rocío Del Amor
- Instituto Universitario de Investigación en Tecnología Centrada en el Ser Humano, HUMAN-tech Universitat Politècnica de València, Valencia, Spain.
- Artikode Intelligence S.L, Valencia, Spain.
| | - Miguel López-Pérez
- Instituto Universitario de Investigación en Tecnología Centrada en el Ser Humano, HUMAN-tech Universitat Politècnica de València, Valencia, Spain
| | - Pablo Meseguer
- Instituto Universitario de Investigación en Tecnología Centrada en el Ser Humano, HUMAN-tech Universitat Politècnica de València, Valencia, Spain
| | - Sandra Morales
- Instituto Universitario de Investigación en Tecnología Centrada en el Ser Humano, HUMAN-tech Universitat Politècnica de València, Valencia, Spain
| | - Liria Terradez
- Pathology Department. Hospital Clínico Universitario de Valencia, Universidad de Valencia, Valencia, Spain
| | | | - Javier Mateos
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada, Spain
| | - Rafael Molina
- Department of Computer Science and Artificial Intelligence, University of Granada, Granada, Spain
| | - Valery Naranjo
- Instituto Universitario de Investigación en Tecnología Centrada en el Ser Humano, HUMAN-tech Universitat Politècnica de València, Valencia, Spain.
- Artikode Intelligence S.L, Valencia, Spain.
| |
Collapse
|
49
|
Kahveci M, Uğur L. Prediction and Stage Classification of Pressure Ulcers in Intensive Care Patients by Machine Learning. Diagnostics (Basel) 2025; 15:1239. [PMID: 40428232 PMCID: PMC12109807 DOI: 10.3390/diagnostics15101239] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2025] [Revised: 04/23/2025] [Accepted: 05/07/2025] [Indexed: 05/29/2025] Open
Abstract
Background/Objective: Pressure ulcers are a serious clinical problem associated with high morbidity, mortality and healthcare costs, especially in intensive care unit (ICU) patients. Existing risk assessment tools, such as the Braden Score, are often inadequate in ICU patients and have poor discriminatory power between classes. This increases the need for more sensitive, predictive and integrative systems. The aim of this study was to classify pressure ulcer stages (Stages I-IV) with high accuracy using machine learning algorithms using demographic, clinical and laboratory data of ICU patients and to evaluate the model performance at a level that can be integrated into clinical decision support systems. Methods: A total of 200 patients hospitalized in the ICU were included in the study. Using demographic, clinical and laboratory data of the patients, six different machine learning algorithms (SVM, KNN, ANN, Decision Tree, Naive Bayes and Discriminant Analysis) were used for classification. The models were evaluated using confusion matrices, ROC-AUC analyses and metrics such as class-based sensitivity and error rate. Results: SVM, KNN and ANN models showed the highest success in classifying pressure ulcer stages, achieving 99% overall accuracy and excellent performance with AUC = 1.00. Variables such as Braden score, albumin and CRP levels contributed significantly to model performance. ROC curves showed that the models provided strong discrimination between classes. Key predictors of pressure ulcer severity included prolonged ICU stay (p < 0.001), low albumin (Stage I: 3.4 ± 0.5 g/dL vs. Stage IV: 2.4 ± 0.8 g/dL; p < 0.001) and high CRP (Stage I: 28 mg/L vs. Stage IV: 142 mg/L; p < 0.001). Conclusions: This study shows that machine learning algorithms offer high accuracy and generalization potential in pressure ulcer classification. In particular, the effectiveness of algorithms such as SVM, ANN and KNN in detecting early-stage ulcers is promising in terms of integration into clinical decision support systems. In future studies, the clinical validity of the model should be increased with multicenter datasets and visual-data-based hybrid models.
Collapse
Affiliation(s)
- Mürsel Kahveci
- Anesthesiology and Reanimation, Amasya Training and Reserch Hospital, Amasya University, Amasya 05100, Turkey;
| | - Levent Uğur
- Department of Mechanical Engineering, Faculty of Engineering, Amasya University, Amasya 05100, Turkey
| |
Collapse
|
50
|
Yinusa A, Faezipour M. A multi-layered defense against adversarial attacks in brain tumor classification using ensemble adversarial training and feature squeezing. Sci Rep 2025; 15:16804. [PMID: 40369011 PMCID: PMC12078522 DOI: 10.1038/s41598-025-00890-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2025] [Accepted: 05/02/2025] [Indexed: 05/16/2025] Open
Abstract
Deep learning, particularly convolutional neural networks (CNNs), has proven valuable for brain tumor classification, aiding diagnostic and therapeutic decisions in medical imaging. Despite their accuracy, these models are vulnerable to adversarial attacks, compromising their reliability in clinical settings. In this research, we utilized a VGG16-based CNN model to classify brain tumors, achieving 96% accuracy on clean magnetic resonance imaging (MRI) data. To assess robustness, we exposed the model to Fast Gradient Sign Method (FGSM) and Projected Gradient Descent (PGD) attacks, which reduced accuracy to 32% and 13%, respectively. We then applied a multi-layered defense strategy, including adversarial training with FGSM and PGD examples and feature squeezing techniques such as bit-depth reduction and Gaussian blurring. This approach improved model resilience, achieving 54% accuracy on FGSM and 47% on PGD adversarial examples. Our results highlight the importance of proactive defense strategies for maintaining the reliability of AI in medical imaging under adversarial conditions.
Collapse
Affiliation(s)
- Ahmeed Yinusa
- Computational and Data Science Program, Middle Tennessee State University, 1301 East Main Street, Murfreesboro, TN, 37132, USA
| | - Misa Faezipour
- Department of Engineering Technology, Middle Tennessee State University, 1301 East Main Street, Murfreesboro, TN, 37132, USA.
| |
Collapse
|