1
|
Wang S, Shen Y, Zeng F, Wang M, Li B, Shen D, Tang X, Wang B. Exploiting biochemical data to improve osteosarcoma diagnosis with deep learning. Health Inf Sci Syst 2024; 12:31. [PMID: 38645838 PMCID: PMC11026331 DOI: 10.1007/s13755-024-00288-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2023] [Accepted: 03/05/2024] [Indexed: 04/23/2024] Open
Abstract
Early and accurate diagnosis of osteosarcomas (OS) is of great clinical significance, and machine learning (ML) based methods are increasingly adopted. However, current ML-based methods for osteosarcoma diagnosis consider only X-ray images, usually fail to generalize to new cases, and lack explainability. In this paper, we seek to explore the capability of deep learning models in diagnosing primary OS, with higher accuracy, explainability, and generality. Concretely, we analyze the added value of integrating the biochemical data, i.e., alkaline phosphatase (ALP) and lactate dehydrogenase (LDH), and design a model that incorporates the numerical features of ALP and LDH and the visual features of X-ray imaging through a late fusion approach in the feature space. We evaluate this model on real-world clinic data with 848 patients aged from 4 to 81. The experimental results reveal the effectiveness of incorporating ALP and LDH simultaneously in a late fusion approach, with the accuracy of the considered 2608 cases increased to 97.17%, compared to 94.35% in the baseline. Grad-CAM visualizations consistent with orthopedic specialists further justified the model's explainability.
Collapse
Affiliation(s)
- Shidong Wang
- Musculoskeletal Tumor Center, Peking University People’s Hospital, Beijing, China
| | - Yangyang Shen
- School of Computer Science and Technology, Southeast University, Nanjing, China
| | - Fanwei Zeng
- Musculoskeletal Tumor Center, Peking University People’s Hospital, Beijing, China
| | - Meng Wang
- College of Design and Innovation, Tongji University, Shanghai, China
| | - Bohan Li
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China
- Ministry of Industry and Information Technology, Collaborative Innovation Center of Novel Software Technology and Industrialization, Nanjing, China
- National Engineering Laboratory for Integrated Aero-Space-Ground Ocean Big Data Application Technology, Xi’an, China
| | - Dian Shen
- School of Computer Science and Technology, Southeast University, Nanjing, China
| | - Xiaodong Tang
- Musculoskeletal Tumor Center, Peking University People’s Hospital, Beijing, China
| | - Beilun Wang
- School of Computer Science and Technology, Southeast University, Nanjing, China
| |
Collapse
|
2
|
Zhu J, Zou L, Xie X, Xu R, Tian Y, Zhang B. 2.5D deep learning based on multi-parameter MRI to differentiate primary lung cancer pathological subtypes in patients with brain metastases. Eur J Radiol 2024; 180:111712. [PMID: 39222565 DOI: 10.1016/j.ejrad.2024.111712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Revised: 08/17/2024] [Accepted: 08/27/2024] [Indexed: 09/04/2024]
Abstract
BACKGROUND Brain metastases (BMs) represents a severe neurological complication stemming from cancers originating from various sources. It is a highly challenging clinical task to accurately distinguish the pathological subtypes of brain metastatic tumors from lung cancer (LC).The utility of 2.5-dimensional (2.5D) deep learning (DL) in distinguishing pathological subtypes of LC with BMs is yet to be determined. METHODS A total of 250 patients were included in this retrospective study, divided in a 7:3 ratio into training set (N=175) and testing set (N=75). We devised a method to assemble a series of two-dimensional (2D) images by extracting adjacent slices from a central slice in both superior-inferior and anterior-posterior directions to form a 2.5D dataset. Multi-Instance learning (MIL) is a weakly supervised learning method that organizes training instances into "bags" and provides labels for entire bags, with the purpose of learning a classifier based on the labeled positive and negative bags to predict the corresponding class for an unknown bag. Therefore, we employed MIL to construct a comprehensive 2.5D feature set. Then we used the single-slice as input for constructing the 2D model. DL features were extracted from these slices using the pre-trained ResNet101. All feature sets were inputted into the support vector machine (SVM) for evaluation. The diagnostic performance of the classification models were evaluated using five-fold cross-validation, with accuracy and area under the curve (AUC) metrics calculated for analysis. RESULTS The optimal performance was obtained using the 2.5D DL model, which achieved the micro-AUC of 0.868 (95% confidence interval [CI], 0.817-0.919) and accuracy of 0.836 in the test cohort. The 2D model achieved the micro-AUC of 0.836 (95 % CI, 0.778-0.894) and accuracy of 0.827 in the test cohort. CONCLUSIONS The proposed 2.5D DL model is feasible and effective in identifying pathological subtypes of BMs from lung cancer.
Collapse
Affiliation(s)
- Jinling Zhu
- Department Of Radiology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Li Zou
- Department of Radiotherapy & Oncology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Xin Xie
- Department Of Radiology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Ruizhe Xu
- Department of Radiotherapy & Oncology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Ye Tian
- Department of Radiotherapy & Oncology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China.
| | - Bo Zhang
- Department Of Radiology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China.
| |
Collapse
|
3
|
Lv Q, Liang K, Tian C, Zhang Y, Li Y, Deng J, Yue W, Li W. Unveiling Thymoma Typing Through Hyperspectral Imaging and Deep Learning. JOURNAL OF BIOPHOTONICS 2024; 17:e202400325. [PMID: 39362657 DOI: 10.1002/jbio.202400325] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2024] [Revised: 07/30/2024] [Accepted: 08/15/2024] [Indexed: 10/05/2024]
Abstract
Thymoma, a rare tumor from thymic epithelial cells, presents diagnostic challenges because of the subjective nature of traditional methods, leading to high false-negative rates and long diagnosis times. This study introduces a thymoma classification technique that integrates hyperspectral imaging with deep learning. We initially capture pathological slice images of thymoma using a hyperspectral camera and delineate regions of interest to extract spectral data. This data undergoes reflectance calibration and noise reduction. Subsequently, we transform the spectral data into two-dimensional images via the Gramian Angular Field (GAF) method. A variant residual network is then utilized to extract features and classify these images. Our results demonstrate that this model significantly enhances classification accuracy and efficiency, achieving an average accuracy of 95%. The method proves highly effective in automated thymoma diagnosis, optimizing data utilization, and feature representation learning.
Collapse
Affiliation(s)
- Qize Lv
- Department of Biomedical Engineering Institute, School of Control Science and Engineering, Shandong University, Jinan, China
| | - Ke Liang
- Department of Pathology, Qilu Hospital of Shandong University, Jinan, China
| | - ChongXuan Tian
- Department of Biomedical Engineering Institute, School of Control Science and Engineering, Shandong University, Jinan, China
| | - YanHai Zhang
- Department of Biomedical Engineering Institute, School of Control Science and Engineering, Shandong University, Jinan, China
| | - YunZe Li
- Department of Biomedical Engineering Institute, School of Control Science and Engineering, Shandong University, Jinan, China
| | - JinLin Deng
- Department of Biomedical Engineering Institute, School of Control Science and Engineering, Shandong University, Jinan, China
| | - WeiMing Yue
- Department of Thoracic Surgery, Qilu Hospital, Shandong University, Jinan, China
| | - Wei Li
- Department of Biomedical Engineering Institute, School of Control Science and Engineering, Shandong University, Jinan, China
| |
Collapse
|
4
|
Panyarak W, Suttapak W, Mahasantipiya P, Charuakkra A, Boonsong N, Wantanajittikul K, Iamaroon A. CrossViT with ECAP: Enhanced deep learning for jaw lesion classification. Int J Med Inform 2024; 193:105666. [PMID: 39492085 DOI: 10.1016/j.ijmedinf.2024.105666] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2024] [Revised: 07/25/2024] [Accepted: 10/24/2024] [Indexed: 11/05/2024]
Abstract
BACKGROUND Radiolucent jaw lesions like ameloblastoma (AM), dentigerous cyst (DC), odontogenic keratocyst (OKC), and radicular cyst (RC) often share similar characteristics, making diagnosis challenging. In 2021, CrossViT, a novel deep learning approach using multi-scale vision transformers (ViT) with cross-attention, emerged for accurate image classification. Additionally, we introduced Extended Cropping and Padding (ECAP), a method to expand training data by iteratively cropping smaller images while preserving context. However, its application in dental radiographic classification remains unexplored. This study investigates the effectiveness of CrossViTs and ECAP against ResNets for classifying common radiolucent jaw lesions. METHODS We conducted a retrospective study involving 208 prevalent radiolucent jaw lesions (49 AMs, 59 DCs, 48 OKCs, and 54 RCs) observed in panoramic radiographs or orthopantomograms (OPGs) with confirmed histological diagnoses. Three experienced oral radiologists provided annotations with consensus. We implemented horizontal flip and ECAP technique with CrossViT-15, -18, ResNet-50, -101, and -152. A four-fold cross-validation approach was employed. The models' performance assessed through accuracy, specificity, precision, recall (sensitivity), F1-score, and area under the receiver operating characteristics (AUCs) metrics. RESULTS Models using the ECAP technique generally achieved better results, with ResNet-152 showing a statistically significant increase in F1-score. CrossViT models consistently achieved higher accuracy, precision, recall, and F1-score compared to ResNet models, regardless of ECAP usage. CrossViT-18 achieved the best overall performance. While all models showed positive ability to differentiate lesions, DC had the highest AUCs (0.89-0.90) and OKC the lowest (0.72-0.81). Only CrossViT-15 achieved AUCs above 0.80 for all four lesion types. CONCLUSION ECAP, a targeted padding data technique, improves deep learning model performance for radiolucent jaw lesion classification. This context-preserving approach is beneficial for tasks requiring an understanding of the lesion's surroundings. Combined with CrossViT models, ECAP shows promise for accurate classification, particularly for rare lesions with limited data.
Collapse
Affiliation(s)
- Wannakamon Panyarak
- Division of Oral and Maxillofacial Radiology, Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Suthep Road, Suthep Sub-district, Mueang Chiang Mai District, Chiang Mai 50200, Thailand.
| | - Wattanapong Suttapak
- Division of Computer Engineering, School of Information and Communication Technology, University of Phayao, Phahon Yothin Road, Mae Ka Sub-district, Mueang Phayao District, Phayao 56000, Thailand.
| | - Phattaranant Mahasantipiya
- Division of Oral and Maxillofacial Radiology, Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Suthep Road, Suthep Sub-district, Mueang Chiang Mai District, Chiang Mai 50200, Thailand.
| | - Arnon Charuakkra
- Division of Oral and Maxillofacial Radiology, Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Suthep Road, Suthep Sub-district, Mueang Chiang Mai District, Chiang Mai 50200, Thailand.
| | - Nattanit Boonsong
- Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Suthep Road, Suthep Sub-district, Mueang Chiang Mai District, Chiang Mai 50200, Thailand.
| | - Kittichai Wantanajittikul
- Department of Radiologic Technology, Faculty of Associated Medical Sciences, Chiang Mai University, Suthep Road, Suthep Sub-district, Mueang Chiang Mai District, Chiang Mai 50200, Thailand.
| | - Anak Iamaroon
- Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Suthep Road, Suthep Sub-district, Mueang Chiang Mai District, Chiang Mai 50200, Thailand.
| |
Collapse
|
5
|
Tao Y, Luo Y, Hu H, Wang W, Zhao Y, Wang S, Zheng Q, Zhang T, Zhang G, Li J, Ni M. Clinically applicable optimized periprosthetic joint infection diagnosis via AI based pathology. NPJ Digit Med 2024; 7:303. [PMID: 39462052 PMCID: PMC11513062 DOI: 10.1038/s41746-024-01301-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 10/16/2024] [Indexed: 10/28/2024] Open
Abstract
Periprosthetic joint infection (PJI) is a severe complication after joint replacement surgery that demands precise diagnosis for effective treatment. We enhanced PJI diagnostic accuracy through three steps: (1) developing a self-supervised PJI model with DINO v2 to create a large dataset; (2) comparing multiple intelligent models to identify the best one; and (3) using the optimal model for visual analysis to refine diagnostic practices. The self-supervised model generated 27,724 training samples and achieved a perfect AUC of 1, indicating flawless case differentiation. EfficientNet v2-S outperformed CAMEL2 at the image level, while CAMEL2 was superior at the patient level. By using the weakly supervised PJI model to adjust diagnostic criteria, we reduced the required high-power field diagnoses per slide from five to three. These findings demonstrate AI's potential to improve the accuracy and standardization of PJI pathology and have significant implications for infectious disease diagnostics.
Collapse
Affiliation(s)
- Ye Tao
- Orthopedics Department, Fourth Medical Center, Chinese PLA General Hospital, Beijing, China
| | - Yazhi Luo
- Department of computation, information and technology, Technical University of Munich, Munich, Germany
| | - Hanwen Hu
- Orthopedics Department, Fourth Medical Center, Chinese PLA General Hospital, Beijing, China
| | - Wei Wang
- Thorough Lab, Thorough Future, Beijing, China
| | - Ying Zhao
- Thorough Lab, Thorough Future, Beijing, China
| | - Shuhao Wang
- Thorough Lab, Thorough Future, Beijing, China
| | - Qingyuan Zheng
- Orthopedics Department, Fourth Medical Center, Chinese PLA General Hospital, Beijing, China
| | - Tianwei Zhang
- Orthopedics Department, Fourth Medical Center, Chinese PLA General Hospital, Beijing, China
| | - Guoqiang Zhang
- Orthopedics Department, Fourth Medical Center, Chinese PLA General Hospital, Beijing, China
| | - Jie Li
- Department of Pathology, First Medical Center, Chinese PLA General Hospital, Beijing, China.
| | - Ming Ni
- Orthopedics Department, Fourth Medical Center, Chinese PLA General Hospital, Beijing, China.
| |
Collapse
|
6
|
Pan J, Lin PC, Gong SC, Wang Z, Cao R, Lv Y, Zhang K, Wang L. Feasibility study of opportunistic osteoporosis screening on chest CT using a multi-feature fusion DCNN model. Arch Osteoporos 2024; 19:98. [PMID: 39414670 PMCID: PMC11485148 DOI: 10.1007/s11657-024-01455-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/28/2023] [Accepted: 10/01/2024] [Indexed: 10/18/2024]
Abstract
A multi-feature fusion DCNN model for automated evaluation of lumbar vertebrae L1 on chest combined with clinical information and radiomics permits estimation of volumetric bone mineral density for evaluation of osteoporosis. PURPOSE To develop a multi-feature deep learning model based on chest CT, combined with clinical information and radiomics to explore the feasibility in screening for osteoporosis based on estimation of volumetric bone mineral density. METHODS The chest CT images of 1048 health check subjects were retrospectively collected as the master dataset, and the images of 637 subjects obtained from a different CT scanner were used for the external validation cohort. The subjects were divided into three categories according to the quantitative CT (QCT) examination, namely, normal group, osteopenia group, and osteoporosis group. Firstly, a deep learning-based segmentation model was constructed. Then, classification models were established and selected, and then, an optimal model to build bone density value prediction regression model was chosen. RESULTS The DSC value was 0.951 ± 0.030 in the testing dataset and 0.947 ± 0.060 in the external validation cohort. The multi-feature fusion model based on the lumbar 1 vertebra had the best performance in the diagnosis. The area under the curve (AUC) of diagnosing normal, osteopenia, and osteoporosis was 0.992, 0.973, and 0.989. The mean absolute errors (MAEs) of the bone density prediction regression model in the test set and external testing dataset are 8.20 mg/cm3 and 9.23 mg/cm3, respectively, and the root mean square errors (RMSEs) are 10.25 mg/cm3 and 11.91 mg/cm3, respectively. The R-squared values are 0.942 and 0.923, respectively. The Pearson correlation coefficients are 0.972 and 0.965. CONCLUSION The multi-feature fusion DCNN model based on only the lumbar 1 vertebrae and clinical variables can perform bone density three-classification diagnosis and estimate volumetric bone mineral density. If confirmed in independent populations, this automated opportunistic chest CT evaluation can help clinical screening of large-sample populations to identify subjects at high risk of osteoporotic fracture.
Collapse
Affiliation(s)
- Jing Pan
- Department of Radiology, The Second Affiliated Hospital of Nantong University, Nantong, 226001, Jiangsu, China
- Department of Radiology, Nanjing Hospital of Chinese Medicine Affiliated to Nanjing University of Chinese Medicine, Nanjing, 210000, Jiangsu, China
| | - Peng-Cheng Lin
- School of Electrical Engineering, Nantong University, Nantong, 226001, Jiangsu, China
| | - Shen-Chu Gong
- Department of Radiology, The Second Affiliated Hospital of Nantong University, Nantong, 226001, Jiangsu, China
| | - Ze Wang
- Department of Radiology, The Second Affiliated Hospital of Nantong University, Nantong, 226001, Jiangsu, China
| | - Rui Cao
- Department of Radiology, The Second Affiliated Hospital of Nantong University, Nantong, 226001, Jiangsu, China
| | - Yuan Lv
- Department of Radiology, The Second Affiliated Hospital of Nantong University, Nantong, 226001, Jiangsu, China
| | - Kun Zhang
- School of Electrical Engineering, Nantong University, Nantong, 226001, Jiangsu, China.
| | - Lin Wang
- Department of Radiology, The Second Affiliated Hospital of Nantong University, Nantong, 226001, Jiangsu, China.
| |
Collapse
|
7
|
Zhu Y, Wang J, Xue C, Zhai X, Xiao C, Lu T. Deep Learning and Habitat Radiomics for the Prediction of Glioma Pathology Using Multiparametric MRI: A Multicenter Study. Acad Radiol 2024:S1076-6332(24)00671-8. [PMID: 39322536 DOI: 10.1016/j.acra.2024.09.021] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Revised: 09/05/2024] [Accepted: 09/06/2024] [Indexed: 09/27/2024]
Abstract
RATIONALE AND OBJECTIVES Recent radiomics studies on predicting pathological outcomes of glioma have shown immense potential. However, the predictive ability remains suboptimal due to the tumor intrinsic heterogeneity. We aimed to achieve better pathological prediction outcomes by combining habitat analysis with deep learning. MATERIALS AND METHODS 387 cases of primary glioma from three hospitals were collected, along with their T1 contrast-enhanced and T2-weighted MR sequences, pathological reports and clinical histories. The training set consisted of 264 patients, 82 patients composed the test set, and 41 patients were used as the validation set for hyperparameter tuning and optimal model selection. All groups were sourced from different centers. Through radiomics, deep learning, habitat analysis and combined analysis, we extracted imaging features separately and jointly modeled them with clinical features. We identified the optimal models for predicting glioma grades, Ki67 expression levels, P53 mutation and IDH1 mutation. RESULTS Using a LightGBM model with DenseNet161 features based on habitat subregions, the best tumor grade prediction model was achieved. A LightGBM model with ResNet50 features based on habitat subregions yielded the best Ki67 expression level prediction model. An SVM model with Radiomics and Inception_v3 features provided the best prediction of P53 mutation. The best model for predicting IDH1 mutation was achieved by an MLP model with Radiomics features based on habitat subregions. Clinical features might be potentially helpful for the prediction with relatively weak evidence. CONCLUSION Habitat+Deep Learning feature extraction methods were optimal for predicting grades and Ki67 levels. Deep Learning is optimal for predicting P53 mutation, while the combination of Habitat+ Radiomics models yielded the best prediction for IDH1 mutation.
Collapse
Affiliation(s)
- Yunyang Zhu
- Department of Neurosurgery, The First Affiliated Hospital of Soochow University, Suzhou, China (Y.Z., J.W., T.L.)
| | - Jing Wang
- Department of Neurosurgery, The First Affiliated Hospital of Soochow University, Suzhou, China (Y.Z., J.W., T.L.)
| | - Chen Xue
- Department of Radiology, the Affiliated Brain Hospital of Nanjing Medical University, Nanjing, China (C.X., C.X.)
| | - Xiaoyang Zhai
- The First Affiliated Hospital of Xinxiang University, Xinxiang, China (X.Z.)
| | - Chaoyong Xiao
- Department of Radiology, the Affiliated Brain Hospital of Nanjing Medical University, Nanjing, China (C.X., C.X.)
| | - Ting Lu
- Department of Neurosurgery, The First Affiliated Hospital of Soochow University, Suzhou, China (Y.Z., J.W., T.L.).
| |
Collapse
|
8
|
Pan Y, Gou F, Xiao C, Liu J, Zhou J. Semi-supervised recognition for artificial intelligence assisted pathology image diagnosis. Sci Rep 2024; 14:21984. [PMID: 39304708 DOI: 10.1038/s41598-024-70750-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2024] [Accepted: 08/20/2024] [Indexed: 09/22/2024] Open
Abstract
The analysis and interpretation of cytopathological images are crucial in modern medical diagnostics. However, manually locating and identifying relevant cells from the vast amount of image data can be a daunting task. This challenge is particularly pronounced in developing countries where there may be a shortage of medical expertise to handle such tasks. The challenge of acquiring large amounts of high-quality labelled data remains, many researchers have begun to use semi-supervised learning methods to learn from unlabeled data. Although current semi-supervised learning models partially solve the issue of limited labelled data, they are inefficient in exploiting unlabeled samples. To address this, we introduce a new AI-assisted semi-supervised scheme, the Reliable-Unlabeled Semi-Supervised Segmentation (RU3S) model. This model integrates the ResUNet-SE-ASPP-Attention (RSAA) model, which includes the Squeeze-and-Excitation (SE) network, Atrous Spatial Pyramid Pooling (ASPP) structure, Attention module, and ResUNet architecture. Our model leverages unlabeled data effectively, improving accuracy significantly. A novel confidence filtering strategy is introduced to make better use of unlabeled samples, addressing the scarcity of labelled data. Experimental results show a 2.0% improvement in mIoU accuracy over the current state-of-the-art semi-supervised segmentation model ST, demonstrating our approach's effectiveness in solving this medical problem.
Collapse
Affiliation(s)
- Yao Pan
- School of Computer Science, Jiangxi University of Traditional Chinese Medicine, Nanchang, 330004, China
| | - Fangfang Gou
- State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang, 550025, China.
| | - Chunwen Xiao
- The Second People's Hospital of Huaihua, Huaihua, 418000, China
| | - Jun Liu
- The Second People's Hospital of Huaihua, Huaihua, 418000, China.
| | - Jing Zhou
- Hunan University of Medicine General Hospital, Huaihua, 418000, China.
| |
Collapse
|
9
|
Li S, Hu Y, Yang L, Lv B, Kong X, Qiang G. DSEception: a noval neural networks architecture for enhancing pneumonia and tuberculosis diagnosis. Front Bioeng Biotechnol 2024; 12:1454652. [PMID: 39291256 PMCID: PMC11405223 DOI: 10.3389/fbioe.2024.1454652] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2024] [Accepted: 08/23/2024] [Indexed: 09/19/2024] Open
Abstract
Background Pneumonia and tuberculosis are prevalent pulmonary diseases globally, each demanding specific care measures. However, distinguishing between these two conditions imposes challenges due to the high skill requirements for doctors, the impact of imaging positions and respiratory intensity of patients, and the associated high healthcare costs, emphasizing the imperative need for intelligent and efficient diagnostic methods. Method This study aims to develop a highly accurate automatic diagnosis and classification method for various lung diseases (Normal, Pneumonia, and Tuberculosis). We propose a hybrid model, which is based on the InceptionV3 architecture, enhanced by introducing Deepwise Separable Convolution after the Inception modules and incorporating the Squeeze-and-Excitation mechanism. This architecture successfully enables the model to extract richer features without significantly increasing the parameter count and computational workload, thereby markedly improving the performance in predicting and classifying lung diseases. To objectively assess the proposed model, external testing and five-fold cross-validation were conducted. Additionally, widely used baseline models in the scholarly community were constructed for comparison. Result In the external testing phase, the our model achieved an average accuracy (ACC) of 90.48% and an F1-score (F1) of 91.44%, which is an approximate 4% improvement over the best-performing baseline model, ResNet. In the five-fold cross-validation, our model's average ACC and F1 reached 88.27% ± 2.76% and 89.29% ± 2.69%, respectively, demonstrating exceptional predictive performance and stability. The results indicate that our model holds promise for deployment in clinical settings to assist in the diagnosis of lung diseases, potentially reducing misdiagnosis rates and patient losses. Conclusion Utilizing deep learning for automatic assistance in the diagnosis of pneumonia and tuberculosis holds clinical significance by enhancing diagnostic accuracy, reducing healthcare costs, enabling rapid screening and large-scale detection, and facilitating personalized treatment approaches, thereby contributing to widespread accessibility and improved healthcare services in the future.
Collapse
Affiliation(s)
- Shengyi Li
- Internet of Things Engineering, Beijing-Dublin international College, Beijing University of Technology, Beijing, China
| | - Yue Hu
- China Academy of Chinese Medical Sciences, Guang'anmen Hospital, Beijing, China
| | - Lexin Yang
- Internet of Things Engineering, Beijing-Dublin international College, Beijing University of Technology, Beijing, China
| | - Baohua Lv
- Department of Radiology, Taian City Central Hospital, Qingdao University, Qingdao, Shandong, China
| | - Xue Kong
- Department of Radiology, Taian City Central Hospital, Qingdao University, Qingdao, Shandong, China
| | - Guangliang Qiang
- Department of Thoracic Surgery, Peking University Third Hospital, Beijing, China
| |
Collapse
|
10
|
Nguyen CV, Duong HM, Do CD. MELEP: A Novel Predictive Measure of Transferability in Multi-label ECG Diagnosis. JOURNAL OF HEALTHCARE INFORMATICS RESEARCH 2024; 8:506-522. [PMID: 39131101 PMCID: PMC11310184 DOI: 10.1007/s41666-024-00168-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2023] [Revised: 05/04/2024] [Accepted: 06/04/2024] [Indexed: 08/13/2024]
Abstract
In practical electrocardiography (ECG) interpretation, the scarcity of well-annotated data is a common challenge. Transfer learning techniques are valuable in such situations, yet the assessment of transferability has received limited attention. To tackle this issue, we introduce MELEP, which stands for Muti-label Expected Log of Empirical Predictions, a measure designed to estimate the effectiveness of knowledge transfer from a pre-trained model to a downstream multi-label ECG diagnosis task. MELEP is generic, working with new target data with different label sets, and computationally efficient, requiring only a single forward pass through the pre-trained model. To the best of our knowledge, MELEP is the first transferability metric specifically designed for multi-label ECG classification problems. Our experiments show that MELEP can predict the performance of pre-trained convolutional and recurrent deep neural networks, on small and imbalanced ECG data. Specifically, we observed strong correlation coefficients (with absolute values exceeding 0.6 in most cases) between MELEP and the actual average F1 scores of the fine-tuned models. Our work highlights the potential of MELEP to expedite the selection of suitable pre-trained models for ECG diagnosis tasks, saving time and effort that would otherwise be spent on fine-tuning these models.
Collapse
Affiliation(s)
- Cuong V. Nguyen
- College of Engineering and Computer Science, VinUniversity, Hanoi, Vietnam
| | - Hieu Minh Duong
- College of Engineering and Computer Science, VinUniversity, Hanoi, Vietnam
| | - Cuong D. Do
- College of Engineering and Computer Science, VinUniversity, Hanoi, Vietnam
- VinUni-Illinois Smart Health Center, VinUniversity, Hanoi, Vietnam
| |
Collapse
|
11
|
Ling R, Wang M, Lu J, Wu S, Wu P, Ge J, Wang L, Liu Y, Jiang J, Shi K, Yan Z, Zuo C, Jiang J. Radiomics-Guided Deep Learning Networks Classify Differential Diagnosis of Parkinsonism. Brain Sci 2024; 14:680. [PMID: 39061420 PMCID: PMC11274493 DOI: 10.3390/brainsci14070680] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2024] [Revised: 06/17/2024] [Accepted: 06/26/2024] [Indexed: 07/28/2024] Open
Abstract
The differential diagnosis between atypical Parkinsonian syndromes may be challenging and critical. We aimed to proposed a radiomics-guided deep learning (DL) model to discover interpretable DL features and further verify the proposed model through the differential diagnosis of Parkinsonian syndromes. We recruited 1495 subjects for 18F-fluorodeoxyglucose positron emission tomography (18F-FDG PET) scanning, including 220 healthy controls and 1275 patients diagnosed with idiopathic Parkinson's disease (IPD), multiple system atrophy (MSA), or progressive supranuclear palsy (PSP). Baseline radiomics and two DL models were developed and tested for the Parkinsonian diagnosis. The DL latent features were extracted from the last layer and subsequently guided by radiomics. The radiomics-guided DL model outperformed the baseline radiomics approach, suggesting the effectiveness of the DL approach. DenseNet showed the best diagnosis ability (sensitivity: 95.7%, 90.1%, and 91.2% for IPD, MSA, and PSP, respectively) using retained DL features in the test dataset. The retained DL latent features were significantly associated with radiomics features and could be interpreted through biological explanations of handcrafted radiomics features. The radiomics-guided DL model offers interpretable high-level abstract information for differential diagnosis of Parkinsonian disorders and holds considerable promise for personalized disease monitoring.
Collapse
Affiliation(s)
- Ronghua Ling
- School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China;
- School of Medical Imaging, Shanghai University of Medicine & Health Science, Shanghai 201318, China;
| | - Min Wang
- School of Life Sciences, Shanghai University, Shanghai 200444, China (J.J.)
| | - Jiaying Lu
- Department of Nuclear Medicine & PET Center, National Clinical Research Center for Aging and Medicine, & National Center for Neurological Disorders, Huashan Hospital, Fudan University, Shanghai 200437, China
| | - Shaoyou Wu
- School of Life Sciences, Shanghai University, Shanghai 200444, China (J.J.)
| | - Ping Wu
- Department of Nuclear Medicine & PET Center, National Clinical Research Center for Aging and Medicine, & National Center for Neurological Disorders, Huashan Hospital, Fudan University, Shanghai 200437, China
| | - Jingjie Ge
- Department of Nuclear Medicine & PET Center, National Clinical Research Center for Aging and Medicine, & National Center for Neurological Disorders, Huashan Hospital, Fudan University, Shanghai 200437, China
| | - Luyao Wang
- School of Life Sciences, Shanghai University, Shanghai 200444, China (J.J.)
| | - Yingqian Liu
- School of Electrical Engineering, Shandong University of Aeronautics, Binzhou 256601, China
| | - Juanjuan Jiang
- School of Medical Imaging, Shanghai University of Medicine & Health Science, Shanghai 201318, China;
| | - Kuangyu Shi
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, 3010 Bern, Switzerland
- Computer Aided Medical Procedures, School of Computation, Information and Technology, Technical University of Munich, 85748 Munich, Germany
| | - Zhuangzhi Yan
- School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China;
- School of Life Sciences, Shanghai University, Shanghai 200444, China (J.J.)
| | - Chuantao Zuo
- Department of Nuclear Medicine & PET Center, National Clinical Research Center for Aging and Medicine, & National Center for Neurological Disorders, Huashan Hospital, Fudan University, Shanghai 200437, China
| | - Jiehui Jiang
- School of Life Sciences, Shanghai University, Shanghai 200444, China (J.J.)
| |
Collapse
|
12
|
Safai A, Froines C, Slater R, Linderman RE, Bogost J, Pacheco C, Voland R, Pak J, Tiwari P, Channa R, Domalpally A. Quantifying Geographic Atrophy in Age-Related Macular Degeneration: A Comparative Analysis Across 12 Deep Learning Models. Invest Ophthalmol Vis Sci 2024; 65:42. [PMID: 39046755 PMCID: PMC11271806 DOI: 10.1167/iovs.65.8.42] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Accepted: 07/02/2024] [Indexed: 07/25/2024] Open
Abstract
Purpose AI algorithms have shown impressive performance in segmenting geographic atrophy (GA) from fundus autofluorescence (FAF) images. However, selection of artificial intelligence (AI) architecture is an important variable in model development. Here, we explore 12 distinct AI architecture combinations to determine the most effective approach for GA segmentation. Methods We investigated various AI architectures, each with distinct combinations of encoders and decoders. The architectures included three decoders-FPN (Feature Pyramid Network), UNet, and PSPNet (Pyramid Scene Parsing Network)-and serve as the foundation framework for segmentation task. Encoders including EfficientNet, ResNet (Residual Networks), VGG (Visual Geometry Group) and Mix Vision Transformer (mViT) have a role in extracting optimum latent features for accurate GA segmentation. Performance was measured through comparison of GA areas between human and AI predictions and Dice Coefficient (DC). Results The training dataset included 601 FAF images from AREDS2 study and validation included 156 FAF images from the GlaxoSmithKline study. The mean absolute difference between grader measured and AI predicted areas ranged from -0.08 (95% CI = -1.35, 1.19) to 0.73 mm2 (95% CI = -5.75,4.29) and DC between 0.884-0.993. The best-performing models were UNet and FPN frameworks with mViT, and the least-performing models were PSPNet framework. Conclusions The choice of AI architecture impacts GA segmentation performance. Vision transformers with FPN and UNet architectures demonstrate stronger suitability for this task compared to Convolutional Neural Network- and PSPNet-based models. Selecting an AI architecture must be tailored to the specific goals of the project, and developers should consider which architecture is ideal for their project.
Collapse
Affiliation(s)
- Apoorva Safai
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
- Depts of Radiology and Biomedical Engineering, University of Wisconsin, Madison, Wisconsin, United States
| | - Colin Froines
- Wisconsin Reading Center, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Robert Slater
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Rachel E. Linderman
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
- Wisconsin Reading Center, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Jacob Bogost
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Caleb Pacheco
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Rickie Voland
- Wisconsin Reading Center, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Jeong Pak
- Wisconsin Reading Center, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Pallavi Tiwari
- Depts of Radiology and Biomedical Engineering, University of Wisconsin, Madison, Wisconsin, United States
| | - Roomasa Channa
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| | - Amitha Domalpally
- A-EYE Research Unit, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
- Wisconsin Reading Center, Dept of Ophthalmology and Visual Sciences, University of Wisconsin, Madison, Wisconsin, United States
| |
Collapse
|
13
|
Kondejkar T, Al-Heejawi SMA, Breggia A, Ahmad B, Christman R, Ryan ST, Amal S. Multi-Scale Digital Pathology Patch-Level Prostate Cancer Grading Using Deep Learning: Use Case Evaluation of DiagSet Dataset. Bioengineering (Basel) 2024; 11:624. [PMID: 38927860 PMCID: PMC11200755 DOI: 10.3390/bioengineering11060624] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/06/2024] [Revised: 06/03/2024] [Accepted: 06/12/2024] [Indexed: 06/28/2024] Open
Abstract
Prostate cancer remains a prevalent health concern, emphasizing the critical need for early diagnosis and precise treatment strategies to mitigate mortality rates. The accurate prediction of cancer grade is paramount for timely interventions. This paper introduces an approach to prostate cancer grading, framing it as a classification problem. Leveraging ResNet models on multi-scale patch-level digital pathology and the Diagset dataset, the proposed method demonstrates notable success, achieving an accuracy of 0.999 in identifying clinically significant prostate cancer. The study contributes to the evolving landscape of cancer diagnostics, offering a promising avenue for improved grading accuracy and, consequently, more effective treatment planning. By integrating innovative deep learning techniques with comprehensive datasets, our approach represents a step forward in the pursuit of personalized and targeted cancer care.
Collapse
Affiliation(s)
- Tanaya Kondejkar
- College of Engineering, Northeastern University, Boston, MA 02115, USA; (T.K.); (S.M.A.A.-H.)
| | | | - Anne Breggia
- MaineHealth Institute for Research, Scarborough, ME 04074, USA;
| | - Bilal Ahmad
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.); (S.T.R.)
| | - Robert Christman
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.); (S.T.R.)
| | - Stephen T. Ryan
- Maine Medical Center, Portland, ME 04102, USA; (B.A.); (R.C.); (S.T.R.)
| | - Saeed Amal
- The Roux Institute, Department of Bioengineering, College of Engineering, Northeastern University, Boston, MA 02115, USA
| |
Collapse
|
14
|
Yeh WC, Kuo CY, Chen JM, Ku TH, Yao DJ, Ho YC, Lin RY. Pioneering Data Processing for Convolutional Neural Networks to Enhance the Diagnostic Accuracy of Traditional Chinese Medicine Pulse Diagnosis for Diabetes. Bioengineering (Basel) 2024; 11:561. [PMID: 38927797 PMCID: PMC11201186 DOI: 10.3390/bioengineering11060561] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2024] [Revised: 05/18/2024] [Accepted: 05/28/2024] [Indexed: 06/28/2024] Open
Abstract
Traditional Chinese medicine (TCM) has relied on pulse diagnosis as a cornerstone of healthcare assessment for thousands of years. Despite its long history and widespread use, TCM pulse diagnosis has faced challenges in terms of diagnostic accuracy and consistency due to its dependence on subjective interpretation and theoretical analysis. This study introduces an approach to enhance the accuracy of TCM pulse diagnosis for diabetes by leveraging the power of deep learning algorithms, specifically LeNet and ResNet models, for pulse waveform analysis. LeNet and ResNet models were applied to analyze TCM pulse waveforms using a diverse dataset comprising both healthy individuals and patients with diabetes. The integration of these advanced algorithms with modern TCM pulse measurement instruments shows great promise in reducing practitioner-dependent variability and improving the reliability of diagnoses. This research bridges the gap between ancient wisdom and cutting-edge technology in healthcare. LeNet-F, incorporating special feature extraction of a pulse based on TMC, showed improved training and test accuracies (73% and 67%, respectively, compared with LeNet's 70% and 65%). Moreover, ResNet models consistently outperformed LeNet, with ResNet18-F achieving the highest accuracy (82%) in training and 74% in testing. The advanced preprocessing techniques and additional features contribute significantly to ResNet18-F's superior performance, indicating the importance of feature engineering strategies. Furthermore, the study identifies potential avenues for future research, including optimizing preprocessing techniques to handle pulse waveform variations and noise levels, integrating additional time-frequency domain features, developing domain-specific feature selection algorithms, and expanding the scope to other diseases. These advancements aim to refine traditional Chinese medicine pulse diagnosis, enhancing its accuracy and reliability while integrating it into modern technology for more effective healthcare approaches.
Collapse
Affiliation(s)
- Wei-Chang Yeh
- Department of Industrial Engineering and Engineering Management, National Tsing Hua University, Hsinchu 30013, Taiwan; (W.-C.Y.); (R.-Y.L.)
| | - Chen-Yi Kuo
- Department of Industrial Engineering and Engineering Management, National Tsing Hua University, Hsinchu 30013, Taiwan; (W.-C.Y.); (R.-Y.L.)
| | | | | | - Da-Jeng Yao
- Department of Power Mechanical Engineering, National Tsing Hua University, Hsinchu 30013, Taiwan; (D.-J.Y.)
| | - Ya-Chi Ho
- Department of Power Mechanical Engineering, National Tsing Hua University, Hsinchu 30013, Taiwan; (D.-J.Y.)
| | - Ruei-Yu Lin
- Department of Industrial Engineering and Engineering Management, National Tsing Hua University, Hsinchu 30013, Taiwan; (W.-C.Y.); (R.-Y.L.)
| |
Collapse
|
15
|
Peng J, Ouyang C, Peng H, Hu W, Wang Y, Jiang P. MultiFuseYOLO: Redefining Wine Grape Variety Recognition through Multisource Information Fusion. SENSORS (BASEL, SWITZERLAND) 2024; 24:2953. [PMID: 38733058 PMCID: PMC11086123 DOI: 10.3390/s24092953] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Revised: 05/03/2024] [Accepted: 05/04/2024] [Indexed: 05/13/2024]
Abstract
Based on the current research on the wine grape variety recognition task, it has been found that traditional deep learning models relying only on a single feature (e.g., fruit or leaf) for classification can face great challenges, especially when there is a high degree of similarity between varieties. In order to effectively distinguish these similar varieties, this study proposes a multisource information fusion method, which is centered on the SynthDiscrim algorithm, aiming to achieve a more comprehensive and accurate wine grape variety recognition. First, this study optimizes and improves the YOLOV7 model and proposes a novel target detection and recognition model called WineYOLO-RAFusion, which significantly improves the fruit localization precision and recognition compared with YOLOV5, YOLOX, and YOLOV7, which are traditional deep learning models. Secondly, building upon the WineYOLO-RAFusion model, this study incorporated the method of multisource information fusion into the model, ultimately forming the MultiFuseYOLO model. Experiments demonstrated that MultiFuseYOLO significantly outperformed other commonly used models in terms of precision, recall, and F1 score, reaching 0.854, 0.815, and 0.833, respectively. Moreover, the method improved the precision of the hard to distinguish Chardonnay and Sauvignon Blanc varieties, which increased the precision from 0.512 to 0.813 for Chardonnay and from 0.533 to 0.775 for Sauvignon Blanc. In conclusion, the MultiFuseYOLO model offers a reliable and comprehensive solution to the task of wine grape variety identification, especially in terms of distinguishing visually similar varieties and realizing high-precision identifications.
Collapse
Affiliation(s)
- Jialiang Peng
- College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China; (J.P.); (C.O.); (H.P.)
| | - Cheng Ouyang
- College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China; (J.P.); (C.O.); (H.P.)
| | - Hao Peng
- College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China; (J.P.); (C.O.); (H.P.)
| | - Wenwu Hu
- College of Mechanical and Electrical Engineering, Hunan Agricultural University, Changsha 410128, China;
| | - Yi Wang
- College of Information and Intelligence, Hunan Agricultural University, Changsha 410128, China; (J.P.); (C.O.); (H.P.)
| | - Ping Jiang
- College of Mechanical and Electrical Engineering, Hunan Agricultural University, Changsha 410128, China;
| |
Collapse
|
16
|
Wu DJ, Kollitz M, Ward M, Dharnipragada RS, Gupta R, Sabal LT, Singla A, Tummala R, Dusenbery K, Watanabe Y. Prediction of Obliteration After the Gamma Knife Radiosurgery of Arteriovenous Malformations Using Hand-Crafted Radiomics and Deep-Learning Methods. Cureus 2024; 16:e58835. [PMID: 38784357 PMCID: PMC11114484 DOI: 10.7759/cureus.58835] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 04/23/2024] [Indexed: 05/25/2024] Open
Abstract
INTRODUCTION Brain arteriovenous malformations (bAVMs) are vascular abnormalities that can be treated with embolization or radiotherapy to prevent the risk of future rupture. In this study, we use hand-crafted radiomics and deep learning techniques to predict favorable vs. unfavorable outcomes following Gamma Knife radiosurgery (GKRS) of bAVMs and compare their prediction performances. METHODS One hundred twenty-six patients seen at one academic medical center for GKRS obliteration of bAVMs over 15 years were retrospectively reviewed. Forty-two patients met the inclusion criteria. Favorable outcomes were defined as complete nidus obliteration demonstrated on cerebral angiogram and asymptomatic recovery. Unfavorable outcomes were defined as incomplete obliteration or complications relating to the AVM that developed after GKRS. Outcome predictions were made using a random forest model with hand-crafted radiomic features and a fine-tuned ResNet-34 convolutional neural network (CNN) model. The performance was evaluated by using a ten-fold cross-validation technique. RESULTS The average accuracy and area-under-curve (AUC) values of the Random Forest Classifier (RFC) with radiomics features were 68.5 ±9.80% and 0.705 ±0.086, whereas those of the ResNet-34 model were 60.0 ±11.9% and 0.694 ±0.124. Four radiomics features used with RFC discriminated unfavorable response cases from favorable response cases with statistical significance. When cropped images were used with ResNet-34, the accuracy and AUC decreased to 59.3 ± 14.2% and 55.4 ±10.4%, respectively. CONCLUSIONS A hand-crafted radiomics model and a pre-trained CNN model can be fine-tuned on pre-treatment MRI scans to predict clinical outcomes of AVM patients undergoing GKRS with equivalent prediction performance. The outcome predictions are promising but require further external validation on more patients.
Collapse
Affiliation(s)
- David J Wu
- Medicine, University of Minnesota School of Medicine, Minneapolis, USA
| | - Megan Kollitz
- Radiology, University of Minnesota School of Medicine, Minneapolis, USA
| | - Mitchell Ward
- Neurosurgery, University of Minnesota School of Medicine, Minneapolis, USA
| | | | - Ribhav Gupta
- Medicine, University of Minnesota School of Medicine, Minneapolis, USA
| | - Luke T Sabal
- Neurosurgery, University of Minnesota School of Medicine, Minneapolis, USA
| | - Ayush Singla
- Computer Science, Stanford University, Stanford, USA
| | | | | | - Yoichi Watanabe
- Radiation Oncology, University of Minnesota, Minneapolis, USA
| |
Collapse
|
17
|
Ponomarchuk E, Thomas G, Song M, Krokhmal A, Kvashennikova A, Wang YN, Khokhlova V, Khokhlova T. Histology-based quantification of boiling histotripsy outcomes via ResNet-18 network: Towards mechanical dose metrics. ULTRASONICS 2024; 138:107225. [PMID: 38141356 DOI: 10.1016/j.ultras.2023.107225] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/07/2023] [Revised: 11/21/2023] [Accepted: 12/13/2023] [Indexed: 12/25/2023]
Abstract
This work was focused on the newly developed ultrasonic approach for non-invasive surgery - boiling histotripsy (BH) - recently proposed for mechanical ablation of tissues using pulsed high intensity focused ultrasound (HIFU). The BH lesion is known to depend in size and shape on exposure parameters and mechanical properties, structure and composition of tissue being treated. The aim of this work was to advance the concept of BH dose by investigating quantitative relationships between the parameters of the lesion, pulsing protocols, and targeted tissue properties. A HIFU focus of a 1.5 MHz 256-element array driven by power-enhanced Verasonics system was electronically steered along the grid within 12 × 4 × 12 mm volume to produce volumetric lesions in porcine liver (soft, with abundant collagenous structures) and bovine myocardium (stiff, homogenous cellular) ex vivo tissues with various pulsing protocols (1-10 ms pulses, 1-15 pulses per point). Quantification of the lesion size and completeness was performed through serial histological sectioning, and a computer vision approach using a combination of manual and automated detection of fully fractionated and residual tissue based on neural network ResNet-18 was developed. Histological sample fixation led to underestimation of BH ablation rate compared to the ultrasound-based estimations, and provided similar qualitative feedback as did gross inspection. This suggests that gross observation may be sufficient for qualitatively evaluating the BH treatment completeness. BH efficiency in liver tissue was shown to be insensitive to the changes in pulsing protocol within the tested parameter range, whereas in bovine myocardium the efficiency increased with either increasing pulse length or number of pulses per point or both. The results imply that one universal mechanical dose metric applicable to an arbitrary tissue type is unlikely to be established. The dose metric as a product of the BH pulse duration and the number of pulses per sonication point (BHD1) was shown to be more relevant for initial planning of fractionation of collagenous tissues. The dose metric as a number of pulses per point (BHD2) is more suitable for the treatment planning of softer targets primarily containing cellular tissue, allowing for significant acceleration of treatment using shorter pulses.
Collapse
Affiliation(s)
| | - Gilles Thomas
- Center for Industrial and Medical Ultrasound, University of Washington, Seattle, USA
| | - Minho Song
- Department of Gastroenterology, University of Washington, Seattle, USA
| | - Alisa Krokhmal
- Physics Faculty, Lomonosov Moscow State University, Moscow, Russian Federation
| | | | - Yak-Nam Wang
- Center for Industrial and Medical Ultrasound, University of Washington, Seattle, USA
| | - Vera Khokhlova
- Physics Faculty, Lomonosov Moscow State University, Moscow, Russian Federation; Center for Industrial and Medical Ultrasound, University of Washington, Seattle, USA
| | - Tatiana Khokhlova
- Department of Gastroenterology, University of Washington, Seattle, USA
| |
Collapse
|
18
|
Yousefpour Shahrivar R, Karami F, Karami E. Enhancing Fetal Anomaly Detection in Ultrasonography Images: A Review of Machine Learning-Based Approaches. Biomimetics (Basel) 2023; 8:519. [PMID: 37999160 PMCID: PMC10669151 DOI: 10.3390/biomimetics8070519] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 10/05/2023] [Accepted: 10/26/2023] [Indexed: 11/25/2023] Open
Abstract
Fetal development is a critical phase in prenatal care, demanding the timely identification of anomalies in ultrasound images to safeguard the well-being of both the unborn child and the mother. Medical imaging has played a pivotal role in detecting fetal abnormalities and malformations. However, despite significant advances in ultrasound technology, the accurate identification of irregularities in prenatal images continues to pose considerable challenges, often necessitating substantial time and expertise from medical professionals. In this review, we go through recent developments in machine learning (ML) methods applied to fetal ultrasound images. Specifically, we focus on a range of ML algorithms employed in the context of fetal ultrasound, encompassing tasks such as image classification, object recognition, and segmentation. We highlight how these innovative approaches can enhance ultrasound-based fetal anomaly detection and provide insights for future research and clinical implementations. Furthermore, we emphasize the need for further research in this domain where future investigations can contribute to more effective ultrasound-based fetal anomaly detection.
Collapse
Affiliation(s)
- Ramin Yousefpour Shahrivar
- Department of Biology, College of Convergent Sciences and Technologies, Science and Research Branch, Islamic Azad University, Tehran, 14515-775, Iran
| | - Fatemeh Karami
- Department of Medical Genetics, Applied Biophotonics Research Center, Science and Research Branch, Islamic Azad University, Tehran, 14515-775, Iran
| | - Ebrahim Karami
- Department of Engineering and Applied Sciences, Memorial University of Newfoundland, St. John’s, NL A1B 3X5, Canada
| |
Collapse
|
19
|
Zhou T, Liu F, Ye X, Wang H, Lu H. CCGL-YOLOV5:A cross-modal cross-scale global-local attention YOLOV5 lung tumor detection model. Comput Biol Med 2023; 165:107387. [PMID: 37659112 DOI: 10.1016/j.compbiomed.2023.107387] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2023] [Revised: 07/29/2023] [Accepted: 08/25/2023] [Indexed: 09/04/2023]
Abstract
BACKGROUND Multimodal medical image detection is a key technology in medical image analysis, which plays an important role in tumor diagnosis. There are different sizes lesions and different shapes lesions in multimodal lung tumor images, which makes it difficult to effectively extract key features of lung tumor lesions. METHODS A Cross-modal Cross-scale Clobal-Local Attention YOLOV5 Lung Tumor Detection Model (CCGL-YOLOV5) is proposed in this paper. The main works are as follows: Firstly, the Cross-Modal Fusion Transformer Module (CMFTM) is designed to improve the multimodal key lesion feature extraction ability and fusion ability through the interactive assisted fusion of multimodal features; Secondly, the Global-Local Feature Interaction Module (GLFIM) is proposed to enhance the interaction ability between multimodal global features and multimodal local features through bidirectional interactive branches. Thirdly, the Cross-Scale Attention Fusion Module (CSAFM) is designed to obtain rich multi-scale features through grouping multi-scale attention for feature fusion. RESULTS The comparison experiments with advanced networks are done. The Acc, Rec, mAP, F1 score and FPS of CCGL-YOLOV5 model on multimodal lung tumor PET/CT dataset are 97.83%, 97.39%, 96.67%, 97.61% and 98.59, respectively; The experimental results show that the performance of CCGL-YOLOV5 model in this paper are better than other typical models. CONCLUSION The CCGL-YOLOV5 model can effectively use the multimodal feature information. There are important implications for multimodal medical image research and clinical disease diagnosis in CCGL-YOLOV5 model.
Collapse
Affiliation(s)
- Tao Zhou
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China
| | - Fengzhen Liu
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China.
| | - Xinyu Ye
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China
| | - Hongwei Wang
- School of Computer Science and Engineering, North Minzu University, Yinchuan, 750021, China; Key Laboratory of Image and Graphics Intelligent Processing of State Ethnic Affairs Commission, North Minzu University, Yinchuan, 750021, China
| | - Huiling Lu
- School of Medical Information and Engineering, Ningxia Medical University, Yinchuan, 750004, China.
| |
Collapse
|
20
|
Namatevs I, Nikulins A, Edelmers E, Neimane L, Slaidina A, Radzins O, Sudars K. Modular Neural Networks for Osteoporosis Detection in Mandibular Cone-Beam Computed Tomography Scans. Tomography 2023; 9:1772-1786. [PMID: 37888733 PMCID: PMC10611366 DOI: 10.3390/tomography9050141] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Revised: 09/12/2023] [Accepted: 09/19/2023] [Indexed: 10/28/2023] Open
Abstract
In this technical note, we examine the capabilities of deep convolutional neural networks (DCNNs) for diagnosing osteoporosis through cone-beam computed tomography (CBCT) scans of the mandible. The evaluation was conducted using 188 patients' mandibular CBCT images utilizing DCNN models built on the ResNet-101 framework. We adopted a segmented three-phase method to assess osteoporosis. Stage 1 focused on mandibular bone slice identification, Stage 2 pinpointed the coordinates for mandibular bone cross-sectional views, and Stage 3 computed the mandibular bone's thickness, highlighting osteoporotic variances. The procedure, built using ResNet-101 networks, showcased efficacy in osteoporosis detection using CBCT scans: Stage 1 achieved a remarkable 98.85% training accuracy, Stage 2 minimized L1 loss to a mere 1.02 pixels, and the last stage's bone thickness computation algorithm reported a mean squared error of 0.8377. These findings underline the significant potential of AI in osteoporosis identification and its promise for enhanced medical care. The compartmentalized method endorses a sturdier DCNN training and heightened model transparency. Moreover, the outcomes illustrate the efficacy of a modular transfer learning method for osteoporosis detection, even when relying on limited mandibular CBCT datasets. The methodology given is accompanied by the source code available on GitLab.
Collapse
Affiliation(s)
- Ivars Namatevs
- Institute of Electronics and Computer Science, LV-1006 Riga, Latvia; (A.N.); (E.E.); (K.S.)
| | - Arturs Nikulins
- Institute of Electronics and Computer Science, LV-1006 Riga, Latvia; (A.N.); (E.E.); (K.S.)
| | - Edgars Edelmers
- Institute of Electronics and Computer Science, LV-1006 Riga, Latvia; (A.N.); (E.E.); (K.S.)
- Department of Morphology, Institute of Anatomy and Anthropology, Rīga Stradiņš University, LV-1010 Riga, Latvia
| | - Laura Neimane
- Department of Conservative Dentistry and Oral Health, Institute of Stomatology, Rīga Stradiņš University, LV-1007 Riga, Latvia;
| | - Anda Slaidina
- Department of Prosthetic Dentistry, Institute of Stomatology, Rīga Stradiņš University, LV-1007 Riga, Latvia;
| | - Oskars Radzins
- Department of Orthodontics, Institute of Stomatology, Rīga Stradiņš University, LV-1007 Riga, Latvia;
| | - Kaspars Sudars
- Institute of Electronics and Computer Science, LV-1006 Riga, Latvia; (A.N.); (E.E.); (K.S.)
| |
Collapse
|