1
|
Zhu J, Zou L, Xie X, Xu R, Tian Y, Zhang B. 2.5D deep learning based on multi-parameter MRI to differentiate primary lung cancer pathological subtypes in patients with brain metastases. Eur J Radiol 2024; 180:111712. [PMID: 39222565 DOI: 10.1016/j.ejrad.2024.111712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Revised: 08/17/2024] [Accepted: 08/27/2024] [Indexed: 09/04/2024]
Abstract
BACKGROUND Brain metastases (BMs) represents a severe neurological complication stemming from cancers originating from various sources. It is a highly challenging clinical task to accurately distinguish the pathological subtypes of brain metastatic tumors from lung cancer (LC).The utility of 2.5-dimensional (2.5D) deep learning (DL) in distinguishing pathological subtypes of LC with BMs is yet to be determined. METHODS A total of 250 patients were included in this retrospective study, divided in a 7:3 ratio into training set (N=175) and testing set (N=75). We devised a method to assemble a series of two-dimensional (2D) images by extracting adjacent slices from a central slice in both superior-inferior and anterior-posterior directions to form a 2.5D dataset. Multi-Instance learning (MIL) is a weakly supervised learning method that organizes training instances into "bags" and provides labels for entire bags, with the purpose of learning a classifier based on the labeled positive and negative bags to predict the corresponding class for an unknown bag. Therefore, we employed MIL to construct a comprehensive 2.5D feature set. Then we used the single-slice as input for constructing the 2D model. DL features were extracted from these slices using the pre-trained ResNet101. All feature sets were inputted into the support vector machine (SVM) for evaluation. The diagnostic performance of the classification models were evaluated using five-fold cross-validation, with accuracy and area under the curve (AUC) metrics calculated for analysis. RESULTS The optimal performance was obtained using the 2.5D DL model, which achieved the micro-AUC of 0.868 (95% confidence interval [CI], 0.817-0.919) and accuracy of 0.836 in the test cohort. The 2D model achieved the micro-AUC of 0.836 (95 % CI, 0.778-0.894) and accuracy of 0.827 in the test cohort. CONCLUSIONS The proposed 2.5D DL model is feasible and effective in identifying pathological subtypes of BMs from lung cancer.
Collapse
Affiliation(s)
- Jinling Zhu
- Department Of Radiology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Li Zou
- Department of Radiotherapy & Oncology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Xin Xie
- Department Of Radiology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Ruizhe Xu
- Department of Radiotherapy & Oncology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China
| | - Ye Tian
- Department of Radiotherapy & Oncology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China.
| | - Bo Zhang
- Department Of Radiology, The Second Affiliated Hospital of Soochow University, Suzhou, 215004, China.
| |
Collapse
|
2
|
Guo B, Chen Y, Lin J, Huang B, Bai X, Guo C, Gao B, Gong Q, Bai X. Self-supervised learning for accurately modelling hierarchical evolutionary patterns of cerebrovasculature. Nat Commun 2024; 15:9235. [PMID: 39455566 PMCID: PMC11511858 DOI: 10.1038/s41467-024-53550-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2023] [Accepted: 10/16/2024] [Indexed: 10/28/2024] Open
Abstract
Cerebrovascular abnormalities are critical indicators of stroke and neurodegenerative diseases like Alzheimer's disease (AD). Understanding the normal evolution of brain vessels is essential for detecting early deviations and enabling timely interventions. Here, for the first time, we proposed a pipeline exploring the joint evolution of cortical volumes (CVs) and arterial volumes (AVs) in a large cohort of 2841 individuals. Using advanced deep learning for vessel segmentation, we built normative models of CVs and AVs across spatially hierarchical brain regions. We found that while AVs generally decline with age, distinct trends appear in regions like the circle of Willis. Comparing healthy individuals with those affected by AD or stroke, we identified significant reductions in both CVs and AVs, wherein patients with AD showing the most severe impact. Our findings reveal gender-specific effects and provide critical insights into how these conditions alter brain structure, potentially guiding future clinical assessments and interventions.
Collapse
Affiliation(s)
- Bin Guo
- Xiamen Key Laboratory of Psychoradiology and Neuromodulation, Department of Radiology, West China Xiamen Hospital of Sichuan University, Xiamen, China
- Image Processing Center, Beihang University, Beijing, China
| | - Ying Chen
- Image Processing Center, Beihang University, Beijing, China
- Institute for Stroke and Dementia Research, Klinikum der Universität München, Ludwig-Maximilians University Munich, Munich, Germany
| | - Jinping Lin
- Xiamen Key Laboratory of Psychoradiology and Neuromodulation, Department of Radiology, West China Xiamen Hospital of Sichuan University, Xiamen, China
| | - Bin Huang
- Department of Radiology, Affiliated Hospital of Guizhou Medical University, Guizhou, China
| | - Xiangzhuo Bai
- Zhongxiang Hospital of Traditional Chinese Medicine, Hubei, China
| | | | - Bo Gao
- Department of Radiology, Affiliated Hospital of Guizhou Medical University, Guizhou, China
| | - Qiyong Gong
- Xiamen Key Laboratory of Psychoradiology and Neuromodulation, Department of Radiology, West China Xiamen Hospital of Sichuan University, Xiamen, China.
- Functional and Molecular Imaging Key Laboratory of Sichuan Province, West China Hospital of Sichuan University, Chengdu, China.
- Huaxi MR Research Center (HMRRC), Department of Radiology, West China Hospital of Sichuan University, Chengdu, China.
- Research Unit of Psychoradiology, Chinese Academy of Medical Sciences, Chengdu, China.
| | - Xiangzhi Bai
- Image Processing Center, Beihang University, Beijing, China.
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing, China.
- Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, China.
| |
Collapse
|
3
|
Zhang R, Chung ACS. EfficientQ: An efficient and accurate post-training neural network quantization method for medical image segmentation. Med Image Anal 2024; 97:103277. [PMID: 39094461 DOI: 10.1016/j.media.2024.103277] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2023] [Revised: 05/09/2024] [Accepted: 07/15/2024] [Indexed: 08/04/2024]
Abstract
Model quantization is a promising technique that can simultaneously compress and accelerate a deep neural network by limiting its computation bit-width, which plays a crucial role in the fast-growing AI industry. Despite model quantization's success in producing well-performing low-bit models, the quantization process itself can still be expensive, which may involve a long fine-tuning stage on a large, well-annotated training set. To make the quantization process more efficient in terms of both time and data requirements, this paper proposes a fast and accurate post-training quantization method, namely EfficientQ. We develop this new method with a layer-wise optimization strategy and leverage the powerful alternating direction method of multipliers (ADMM) algorithm to ensure fast convergence. Furthermore, a weight regularization scheme is incorporated to provide more guidance for the optimization of the discrete weights, and a self-adaptive attention mechanism is proposed to combat the class imbalance problem. Extensive comparison and ablation experiments are conducted on two publicly available medical image segmentation datasets, i.e., LiTS and BraTS2020, and the results demonstrate the superiority of the proposed method over various existing post-training quantization methods in terms of both accuracy and optimization speed. Remarkably, with EfficientQ, the quantization of a practical 3D UNet only requires less than 5 min on a single GPU and one data sample. The source code is available at https://github.com/rongzhao-zhang/EfficientQ.
Collapse
Affiliation(s)
- Rongzhao Zhang
- Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong, China.
| | - Albert C S Chung
- Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong, China.
| |
Collapse
|
4
|
Wen F, Chen Z, Wang X, Dou M, Yang J, Yao Y, Shen Y. Deep learning based clinical target volumes contouring for prostate cancer: Easy and efficient application. J Appl Clin Med Phys 2024; 25:e14482. [PMID: 39120487 PMCID: PMC11466469 DOI: 10.1002/acm2.14482] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2023] [Revised: 05/30/2024] [Accepted: 07/09/2024] [Indexed: 08/10/2024] Open
Abstract
BACKGROUND Radiotherapy has been crucial in prostate cancer treatment. However, manual segmentation is labor intensive and highly variable among radiation oncologists. In this study, a deep learning based automated contouring model is constructed for clinical target volumes (CTVs) of intact and postoperative prostate cancer. METHODS Computed tomography (CT) data sets of 197 prostate cancer patients were collected. Two auto-delineation models were built for radical radiotherapy and postoperative radiotherapy of prostate cancer respectively, and each model included CTVn for pelvic lymph nodes and CTVp for prostate tumors or prostate tumor beds. RESULTS In the radical radiotherapy model, the volumetric dice (VD) coefficient of CTVn calculated by AI, was higher than that of the one delineated by the junior physicians (0.85 vs. 0.82, p = 0.018); In the postoperative radiotherapy model, the quantitative parameter of CTVn and CTVp, counted by AI, was better than that of the junior physicians. The median delineation time for AI was 0.23 min in the postoperative model and 0.26 min in the radical model, which were significantly shorter than those of the physicians (50.40 and 45.43 min, respectively, p < 0.001). The correction time of the senior physician for AI was much shorter compared with that for the junior physicians in both models (p < 0.001). CONCLUSION Using deep learning and attention mechanism, a highly consistent and time-saving contouring model was built for CTVs of pelvic lymph nodes and prostate tumors or prostate tumor beds for prostate cancer, which also might be a good approach to train junior radiation oncologists.
Collapse
Affiliation(s)
- Feng Wen
- Department of Radiation OncologyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
- Abdominal Oncology Ward, Cancer CenterWest China Hospital, Sichuan UniversityChengduChina
| | - Zhebin Chen
- Chengdu Institute of Computer ApplicationChinese Academy of Sciences, SichuanChengduChina
- University of Chinese Academy of SciencesBeijingChina
| | - Xin Wang
- Department of Radiation OncologyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
- Abdominal Oncology Ward, Cancer CenterWest China Hospital, Sichuan UniversityChengduChina
| | - Meng Dou
- Chengdu Institute of Computer ApplicationChinese Academy of Sciences, SichuanChengduChina
- University of Chinese Academy of SciencesBeijingChina
| | - Jialuo Yang
- Department of Medicine OncologyShifang people's HospitalShifangChina
| | - Yu Yao
- Chengdu Institute of Computer ApplicationChinese Academy of Sciences, SichuanChengduChina
- University of Chinese Academy of SciencesBeijingChina
| | - Yali Shen
- Department of Radiation OncologyCancer CenterWest China Hospital, Sichuan UniversityChengduChina
- Abdominal Oncology Ward, Cancer CenterWest China Hospital, Sichuan UniversityChengduChina
| |
Collapse
|
5
|
Chen W, Liao M, Bao S, An S, Li W, Liu X, Huang G, Gong H, Luo Q, Xiao C, Li A. A hierarchically annotated dataset drives tangled filament recognition in digital neuron reconstruction. PATTERNS (NEW YORK, N.Y.) 2024; 5:101007. [PMID: 39233689 PMCID: PMC11368685 DOI: 10.1016/j.patter.2024.101007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/29/2024] [Revised: 04/25/2024] [Accepted: 05/23/2024] [Indexed: 09/06/2024]
Abstract
Reconstructing neuronal morphology is vital for classifying neurons and mapping brain connectivity. However, it remains a significant challenge due to its complex structure, dense distribution, and low image contrast. In particular, AI-assisted methods often yield numerous errors that require extensive manual intervention. Therefore, reconstructing hundreds of neurons is already a daunting task for general research projects. A key issue is the lack of specialized training for challenging regions due to inadequate data and training methods. This study extracted 2,800 challenging neuronal blocks and categorized them into multiple density levels. Furthermore, we enhanced images using an axial continuity-based network that improved three-dimensional voxel resolution while reducing the difficulty of neuron recognition. Comparing the pre- and post-enhancement results in automatic algorithms using fluorescence micro-optical sectioning tomography (fMOST) data, we observed a significant increase in the recall rate. Our study not only enhances the throughput of reconstruction but also provides a fundamental dataset for tangled neuron reconstruction.
Collapse
Affiliation(s)
- Wu Chen
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Mingwei Liao
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Shengda Bao
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Sile An
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Wenwei Li
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Xin Liu
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Ganghua Huang
- Key Laboratory of Biomedical Engineering of Hainan Province, School of Biomedical Engineering, Hainan University, Haikou 570228, China
| | - Hui Gong
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
- HUST-Suzhou Institute for Brainsmatics, JITRI, Suzhou 215123, China
| | - Qingming Luo
- Key Laboratory of Biomedical Engineering of Hainan Province, School of Biomedical Engineering, Hainan University, Haikou 570228, China
| | - Chi Xiao
- Key Laboratory of Biomedical Engineering of Hainan Province, School of Biomedical Engineering, Hainan University, Haikou 570228, China
| | - Anan Li
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, Huazhong University of Science and Technology, Wuhan 430074, China
- HUST-Suzhou Institute for Brainsmatics, JITRI, Suzhou 215123, China
- Key Laboratory of Biomedical Engineering of Hainan Province, School of Biomedical Engineering, Hainan University, Haikou 570228, China
| |
Collapse
|
6
|
Ye RZ, Lipatov K, Diedrich D, Bhattacharyya A, Erickson BJ, Pickering BW, Herasevich V. Automatic ARDS surveillance with chest X-ray recognition using convolutional neural networks. J Crit Care 2024; 82:154794. [PMID: 38552452 DOI: 10.1016/j.jcrc.2024.154794] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Revised: 11/20/2023] [Accepted: 12/01/2023] [Indexed: 06/01/2024]
Abstract
OBJECTIVE This study aims to design, validate and assess the accuracy a deep learning model capable of differentiation Chest X-Rays between pneumonia, acute respiratory distress syndrome (ARDS) and normal lungs. MATERIALS AND METHODS A diagnostic performance study was conducted using Chest X-Ray images from adult patients admitted to a medical intensive care unit between January 2003 and November 2014. X-ray images from 15,899 patients were assigned one of three prespecified categories: "ARDS", "Pneumonia", or "Normal". RESULTS A two-step convolutional neural network (CNN) pipeline was developed and tested to distinguish between the three patterns with sensitivity ranging from 91.8% to 97.8% and specificity ranging from 96.6% to 98.8%. The CNN model was validated with a sensitivity of 96.3% and specificity of 96.6% using a previous dataset of patients with Acute Lung Injury (ALI)/ARDS. DISCUSSION The results suggest that a deep learning model based on chest x-ray pattern recognition can be a useful tool in distinguishing patients with ARDS from patients with normal lungs, providing faster results than digital surveillance tools based on text reports. CONCLUSION A CNN-based deep learning model showed clinically significant performance, providing potential for faster ARDS identification. Future research should prospectively evaluate these tools in a clinical setting.
Collapse
Affiliation(s)
- Run Zhou Ye
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA.; Division of Endocrinology, Department of Medicine, Centre de Recherche du CHUS, Sherbrooke QC J1H 5N4, Canada
| | - Kirill Lipatov
- Critical Care Medicine, Mayo Clinic, Eau Claire, WI, United States
| | - Daniel Diedrich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | | | - Bradley J Erickson
- Department of Diagnostic Radiology, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | - Brian W Pickering
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA..
| |
Collapse
|
7
|
Konar D, Bhattacharyya S, Gandhi TK, Panigrahi BK, Jiang R. 3-D Quantum-Inspired Self-Supervised Tensor Network for Volumetric Segmentation of Medical Images. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10312-10325. [PMID: 37022399 DOI: 10.1109/tnnls.2023.3240238] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
This article introduces a novel shallow 3-D self-supervised tensor neural network in quantum formalism for volumetric segmentation of medical images with merits of obviating training and supervision. The proposed network is referred to as the 3-D quantum-inspired self-supervised tensor neural network (3-D-QNet). The underlying architecture of 3-D-QNet is composed of a trinity of volumetric layers, viz., input, intermediate, and output layers interconnected using an S -connected third-order neighborhood-based topology for voxelwise processing of 3-D medical image data, suitable for semantic segmentation. Each of the volumetric layers contains quantum neurons designated by qubits or quantum bits. The incorporation of tensor decomposition in quantum formalism leads to faster convergence of network operations to preclude the inherent slow convergence problems faced by the classical supervised and self-supervised networks. The segmented volumes are obtained once the network converges. The suggested 3-D-QNet is tailored and tested on the BRATS 2019 Brain MR image dataset and the Liver Tumor Segmentation Challenge (LiTS17) dataset extensively in our experiments. The 3-D-QNet has achieved promising dice similarity (DS) as compared with the time-intensive supervised convolutional neural network (CNN)-based models, such as 3-D-UNet, voxelwise residual network (VoxResNet), Dense-Res-Inception Net (DRINet), and 3-D-ESPNet, thereby showing a potential advantage of our self-supervised shallow network on facilitating semantic segmentation.
Collapse
|
8
|
Li Y, Wynne J, Wang J, Roper J, Chang CW, Patel AB, Shelton J, Liu T, Mao H, Yang X. MRI-based prostate cancer classification using 3D efficient capsule network. Med Phys 2024; 51:4748-4758. [PMID: 38346111 DOI: 10.1002/mp.16975] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2023] [Revised: 12/13/2023] [Accepted: 01/21/2024] [Indexed: 07/10/2024] Open
Abstract
BACKGROUND Prostate cancer (PCa) is the most common cancer in men and the second leading cause of male cancer-related death. Gleason score (GS) is the primary driver of PCa risk-stratification and medical decision-making, but can only be assessed at present via biopsy under anesthesia. Magnetic resonance imaging (MRI) is a promising non-invasive method to further characterize PCa, providing additional anatomical and functional information. Meanwhile, the diagnostic power of MRI is limited by qualitative or, at best, semi-quantitative interpretation criteria, leading to inter-reader variability. PURPOSES Computer-aided diagnosis employing quantitative MRI analysis has yielded promising results in non-invasive prediction of GS. However, convolutional neural networks (CNNs) do not implicitly impose a frame of reference to the objects. Thus, CNNs do not encode the positional information properly, limiting method robustness against simple image variations such as flipping, scaling, or rotation. Capsule network (CapsNet) has been proposed to address this limitation and achieves promising results in this domain. In this study, we develop a 3D Efficient CapsNet to stratify GS-derived PCa risk using T2-weighted (T2W) MRI images. METHODS In our method, we used 3D CNN modules to extract spatial features and primary capsule layers to encode vector features. We then propose to integrate fully-connected capsule layers (FC Caps) to create a deeper hierarchy for PCa grading prediction. FC Caps comprises a secondary capsule layer which routes active primary capsules and a final capsule layer which outputs PCa risk. To account for data imbalance, we propose a novel dynamic weighted margin loss. We evaluate our method on a public PCa T2W MRI dataset from the Cancer Imaging Archive containing data from 976 patients. RESULTS Two groups of experiments were performed: (1) we first identified high-risk disease by classifying low + medium risk versus high risk; (2) we then stratified disease in one-versus-one fashion: low versus high risk, medium versus high risk, and low versus medium risk. Five-fold cross validation was performed. Our model achieved an area under receiver operating characteristic curve (AUC) of 0.83 and 0.64 F1-score for low versus high grade, 0.79 AUC and 0.75 F1-score for low + medium versus high grade, 0.75 AUC and 0.69 F1-score for medium versus high grade and 0.59 AUC and 0.57 F1-score for low versus medium grade. Our method outperformed state-of-the-art radiomics-based classification and deep learning methods with the highest metrics for each experiment. Our divide-and-conquer strategy achieved weighted Cohen's Kappa score of 0.41, suggesting moderate agreement with ground truth PCa risks. CONCLUSIONS In this study, we proposed a novel 3D Efficient CapsNet for PCa risk stratification and demonstrated its feasibility. This developed tool provided a non-invasive approach to assess PCa risk from T2W MR images, which might have potential to personalize the treatment of PCa and reduce the number of unnecessary biopsies.
Collapse
Affiliation(s)
- Yuheng Li
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
- The Wallace H. Coulter Department of Biomedical Engineering, Emory University and Georgia Institute of Technology, Atlanta, Georgia, USA
| | - Jacob Wynne
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Jing Wang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Justin Roper
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Chih-Wei Chang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Ashish B Patel
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Joseph Shelton
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Tian Liu
- Department of Radiation Oncology, Icahn School of Medicine at Mount Sinai, New York, New York, USA
| | - Hui Mao
- The Wallace H. Coulter Department of Biomedical Engineering, Emory University and Georgia Institute of Technology, Atlanta, Georgia, USA
- Department of Radiology and Imaging Science and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
- The Wallace H. Coulter Department of Biomedical Engineering, Emory University and Georgia Institute of Technology, Atlanta, Georgia, USA
| |
Collapse
|
9
|
Zhang L, Ning G, Liang H, Han B, Liao H. One-shot neuroanatomy segmentation through online data augmentation and confidence aware pseudo label. Med Image Anal 2024; 95:103182. [PMID: 38688039 DOI: 10.1016/j.media.2024.103182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Revised: 11/26/2023] [Accepted: 04/18/2024] [Indexed: 05/02/2024]
Abstract
Recently, deep learning-based brain segmentation methods have achieved great success. However, most approaches focus on supervised segmentation, which requires many high-quality labeled images. In this paper, we pay attention to one-shot segmentation, aiming to learn from one labeled image and a few unlabeled images. We propose an end-to-end unified network that joints deformation modeling and segmentation tasks. Our network consists of a shared encoder, a deformation modeling head, and a segmentation head. In the training phase, the atlas and unlabeled images are input to the encoder to get multi-scale features. The features are then fed to the multi-scale deformation modeling module to estimate the atlas-to-image deformation field. The deformation modeling module implements the estimation at the feature level in a coarse-to-fine manner. Then, we employ the field to generate the augmented image pair through online data augmentation. We do not apply any appearance transformations cause the shared encoder could capture appearance variations. Finally, we adopt supervised segmentation loss for the augmented image. Considering that the unlabeled images still contain rich information, we introduce confidence aware pseudo label for them to further boost the segmentation performance. We validate our network on three benchmark datasets. Experimental results demonstrate that our network significantly outperforms other deep single-atlas-based and traditional multi-atlas-based segmentation methods. Notably, the second dataset is collected from multi-center, and our network still achieves promising segmentation performance on both the seen and unseen test sets, revealing its robustness. The source code will be available at https://github.com/zhangliutong/brainseg.
Collapse
Affiliation(s)
- Liutong Zhang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Guochen Ning
- School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Hanying Liang
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Boxuan Han
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Hongen Liao
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China; School of Biomedical Engineering, Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China.
| |
Collapse
|
10
|
Jain S, Dhir R, Sikka G. View adaptive unified self-supervised technique for abdominal organ segmentation. Comput Biol Med 2024; 177:108659. [PMID: 38823366 DOI: 10.1016/j.compbiomed.2024.108659] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2023] [Revised: 03/05/2024] [Accepted: 05/24/2024] [Indexed: 06/03/2024]
Abstract
Automatic abdominal organ segmentation is an essential prerequisite for accurate volumetric analysis, disease diagnosis, and tracking by medical practitioners. However, the deformable shapes, variable locations, overlapping with nearby organs, and similar contrast make the segmentation challenging. Moreover, the requirement of a large manually labeled dataset makes it harder. Hence, a semi-supervised contrastive learning approach is utilized to perform the automatic abdominal organ segmentation. Existing 3D deep learning models based on contrastive learning are not able to capture the 3D context of medical volumetric data along three planes/views: axial, sagittal, and coronal views. In this work, a semi-supervised view-adaptive unified model (VAU-model) is proposed to make the 3D deep learning model as view-adaptive to learn 3D context along each view in a unified manner. This method utilizes the novel optimization function that assists the 3D model to learn the 3D context of volumetric medical data along each view in a single model. The effectiveness of the proposed approach is validated on the three types of datasets: BTCV, NIH, and MSD quantitatively and qualitatively. The results demonstrate that the VAU model achieves an average Dice score of 81.61% which is a 3.89% improvement compared to the previous best results for pancreas segmentation in multi-organ dataset BTCV. It also achieves an average Dice score of 77.76% and 76.76% for the pancreas under the single organ non-pathological NIH dataset, and pathological MSD dataset.
Collapse
Affiliation(s)
- Suchi Jain
- Computer Science and Engineering, Dr. B.R. Ambedkar National Institute of Technology, Jalandhar, Punjab, 144008, India.
| | - Renu Dhir
- Computer Science and Engineering, Dr. B.R. Ambedkar National Institute of Technology, Jalandhar, Punjab, 144008, India
| | - Geeta Sikka
- Computer Science and Engineering, National Institute of Technology, Delhi, 110036, India
| |
Collapse
|
11
|
Teramoto S, Uga Y. Convolutional neural networks combined with conventional filtering to semantically segment plant roots in rapidly scanned X-ray computed tomography volumes with high noise levels. PLANT METHODS 2024; 20:73. [PMID: 38773503 PMCID: PMC11106967 DOI: 10.1186/s13007-024-01208-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/09/2023] [Accepted: 05/15/2024] [Indexed: 05/23/2024]
Abstract
BACKGROUND X-ray computed tomography (CT) is a powerful tool for measuring plant root growth in soil. However, a rapid scan with larger pots, which is required for throughput-prioritized crop breeding, results in high noise levels, low resolution, and blurred root segments in the CT volumes. Moreover, while plant root segmentation is essential for root quantification, detailed conditional studies on segmenting noisy root segments are scarce. The present study aimed to investigate the effects of scanning time and deep learning-based restoration of image quality on semantic segmentation of blurry rice (Oryza sativa) root segments in CT volumes. RESULTS VoxResNet, a convolutional neural network-based voxel-wise residual network, was used as the segmentation model. The training efficiency of the model was compared using CT volumes obtained at scan times of 33, 66, 150, 300, and 600 s. The learning efficiencies of the samples were similar, except for scan times of 33 and 66 s. In addition, The noise levels of predicted volumes differd among scanning conditions, indicating that the noise level of a scan time ≥ 150 s does not affect the model training efficiency. Conventional filtering methods, such as median filtering and edge detection, increased the training efficiency by approximately 10% under any conditions. However, the training efficiency of 33 and 66 s-scanned samples remained relatively low. We concluded that scan time must be at least 150 s to not affect segmentation. Finally, we constructed a semantic segmentation model for 150 s-scanned CT volumes, for which the Dice loss reached 0.093. This model could not predict the lateral roots, which were not included in the training data. This limitation will be addressed by preparing appropriate training data. CONCLUSIONS A semantic segmentation model can be constructed even with rapidly scanned CT volumes with high noise levels. Given that scanning times ≥ 150 s did not affect the segmentation results, this technique holds promise for rapid and low-dose scanning. This study offers insights into images other than CT volumes with high noise levels that are challenging to determine when annotating.
Collapse
Affiliation(s)
- Shota Teramoto
- Institute of Crop Sciences, National Agriculture & Food Research Organization, Tsukuba, Ibaraki, 305-8602, Japan.
| | - Yusaku Uga
- Institute of Crop Sciences, National Agriculture & Food Research Organization, Tsukuba, Ibaraki, 305-8602, Japan
| |
Collapse
|
12
|
Khandelwal P, Duong MT, Sadaghiani S, Lim S, Denning AE, Chung E, Ravikumar S, Arezoumandan S, Peterson C, Bedard M, Capp N, Ittyerah R, Migdal E, Choi G, Kopp E, Loja B, Hasan E, Li J, Bahena A, Prabhakaran K, Mizsei G, Gabrielyan M, Schuck T, Trotman W, Robinson J, Ohm DT, Lee EB, Trojanowski JQ, McMillan C, Grossman M, Irwin DJ, Detre JA, Tisdall MD, Das SR, Wisse LEM, Wolk DA, Yushkevich PA. Automated deep learning segmentation of high-resolution 7 Tesla postmortem MRI for quantitative analysis of structure-pathology correlations in neurodegenerative diseases. IMAGING NEUROSCIENCE (CAMBRIDGE, MASS.) 2024; 2:1-30. [PMID: 39301426 PMCID: PMC11409836 DOI: 10.1162/imag_a_00171] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/13/2023] [Revised: 04/01/2024] [Accepted: 04/15/2024] [Indexed: 09/22/2024]
Abstract
Postmortem MRI allows brain anatomy to be examined at high resolution and to link pathology measures with morphometric measurements. However, automated segmentation methods for brain mapping in postmortem MRI are not well developed, primarily due to limited availability of labeled datasets, and heterogeneity in scanner hardware and acquisition protocols. In this work, we present a high-resolution dataset of 135 postmortem human brain tissue specimens imaged at 0.3 mm3 isotropic using a T2w sequence on a 7T whole-body MRI scanner. We developed a deep learning pipeline to segment the cortical mantle by benchmarking the performance of nine deep neural architectures, followed by post-hoc topological correction. We evaluate the reliability of this pipeline via overlap metrics with manual segmentation in 6 specimens, and intra-class correlation between cortical thickness measures extracted from the automatic segmentation and expert-generated reference measures in 36 specimens. We also segment four subcortical structures (caudate, putamen, globus pallidus, and thalamus), white matter hyperintensities, and the normal appearing white matter, providing a limited evaluation of accuracy. We show generalizing capabilities across whole-brain hemispheres in different specimens, and also on unseen images acquired at 0.28 mm3 and 0.16 mm3 isotropic T2*w fast low angle shot (FLASH) sequence at 7T. We report associations between localized cortical thickness and volumetric measurements across key regions, and semi-quantitative neuropathological ratings in a subset of 82 individuals with Alzheimer's disease (AD) continuum diagnoses. Our code, Jupyter notebooks, and the containerized executables are publicly available at the project webpage (https://pulkit-khandelwal.github.io/exvivo-brain-upenn/).
Collapse
Affiliation(s)
- Pulkit Khandelwal
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, United States
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
| | - Michael Tran Duong
- Department of Bioengineering, University of Pennsylvania, Philadelphia, PA, United States
| | - Shokufeh Sadaghiani
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Sydney Lim
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Amanda E. Denning
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Eunice Chung
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Sadhana Ravikumar
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Sanaz Arezoumandan
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Claire Peterson
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Madigan Bedard
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Noah Capp
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Ranjit Ittyerah
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Elyse Migdal
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Grace Choi
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Emily Kopp
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Bridget Loja
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Eusha Hasan
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Jiacheng Li
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Alejandra Bahena
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Karthik Prabhakaran
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Gabor Mizsei
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Marianna Gabrielyan
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Theresa Schuck
- Department of Pathology and Laboratory Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - Winifred Trotman
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - John Robinson
- Department of Pathology and Laboratory Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - Daniel T. Ohm
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Edward B. Lee
- Department of Pathology and Laboratory Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - John Q. Trojanowski
- Department of Pathology and Laboratory Medicine, University of Pennsylvania, Philadelphia, PA, United States
| | - Corey McMillan
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Murray Grossman
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - David J. Irwin
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - John A. Detre
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - M. Dylan Tisdall
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| | - Sandhitsu R. Das
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | | | - David A. Wolk
- Department of Neurology, University of Pennsylvania, Philadelphia, PA, United States
| | - Paul A. Yushkevich
- Penn Image Computing and Science Laboratory, University of Pennsylvania, Philadelphia, PA, United States
- Department of Radiology, University of Pennsylvania, Philadelphia, PA, United States
| |
Collapse
|
13
|
Chen Q, Zhang J, Meng R, Zhou L, Li Z, Feng Q, Shen D. Modality-Specific Information Disentanglement From Multi-Parametric MRI for Breast Tumor Segmentation and Computer-Aided Diagnosis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1958-1971. [PMID: 38206779 DOI: 10.1109/tmi.2024.3352648] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/13/2024]
Abstract
Breast cancer is becoming a significant global health challenge, with millions of fatalities annually. Magnetic Resonance Imaging (MRI) can provide various sequences for characterizing tumor morphology and internal patterns, and becomes an effective tool for detection and diagnosis of breast tumors. However, previous deep-learning based tumor segmentation methods from multi-parametric MRI still have limitations in exploring inter-modality information and focusing task-informative modality/modalities. To address these shortcomings, we propose a Modality-Specific Information Disentanglement (MoSID) framework to extract both inter- and intra-modality attention maps as prior knowledge for guiding tumor segmentation. Specifically, by disentangling modality-specific information, the MoSID framework provides complementary clues for the segmentation task, by generating modality-specific attention maps to guide modality selection and inter-modality evaluation. Our experiments on two 3D breast datasets and one 2D prostate dataset demonstrate that the MoSID framework outperforms other state-of-the-art multi-modality segmentation methods, even in the cases of missing modalities. Based on the segmented lesions, we further train a classifier to predict the patients' response to radiotherapy. The prediction accuracy is comparable to the case of using manually-segmented tumors for treatment outcome prediction, indicating the robustness and effectiveness of the proposed segmentation method. The code is available at https://github.com/Qianqian-Chen/MoSID.
Collapse
|
14
|
Zhao J, Jiang T, Lin Y, Chan LC, Chan PK, Wen C, Chen H. Adaptive Fusion of Deep Learning With Statistical Anatomical Knowledge for Robust Patella Segmentation From CT Images. IEEE J Biomed Health Inform 2024; 28:2842-2853. [PMID: 38446653 DOI: 10.1109/jbhi.2024.3372576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/08/2024]
Abstract
Kneeosteoarthritis (KOA), as a leading joint disease, can be decided by examining the shapes of patella to spot potential abnormal variations. To assist doctors in the diagnosis of KOA, a robust automatic patella segmentation method is highly demanded in clinical practice. Deep learning methods, especially convolutional neural networks (CNNs) have been widely applied to medical image segmentation in recent years. Nevertheless, poor image quality and limited data still impose challenges to segmentation via CNNs. On the other hand, statistical shape models (SSMs) can generate shape priors which give anatomically reliable segmentation to varying instances. Thus, in this work, we propose an adaptive fusion framework, explicitly combining deep neural networks and anatomical knowledge from SSM for robust patella segmentation. Our adaptive fusion framework will accordingly adjust the weight of segmentation candidates in fusion based on their segmentation performance. We also propose a voxel-wise refinement strategy to make the segmentation of CNNs more anatomically correct. Extensive experiments and thorough assessment have been conducted on various mainstream CNN backbones for patella segmentation in low-data regimes, which demonstrate that our framework can be flexibly attached to a CNN model, significantly improving its performance when labeled training data are limited and input image data are of poor quality.
Collapse
|
15
|
Wu H, Peng L, Du D, Xu H, Lin G, Zhou Z, Lu L, Lv W. BAF-Net: bidirectional attention-aware fluid pyramid feature integrated multimodal fusion network for diagnosis and prognosis. Phys Med Biol 2024; 69:105007. [PMID: 38593831 DOI: 10.1088/1361-6560/ad3cb2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Accepted: 04/09/2024] [Indexed: 04/11/2024]
Abstract
Objective. To go beyond the deficiencies of the three conventional multimodal fusion strategies (i.e. input-, feature- and output-level fusion), we propose a bidirectional attention-aware fluid pyramid feature integrated fusion network (BAF-Net) with cross-modal interactions for multimodal medical image diagnosis and prognosis.Approach. BAF-Net is composed of two identical branches to preserve the unimodal features and one bidirectional attention-aware distillation stream to progressively assimilate cross-modal complements and to learn supplementary features in both bottom-up and top-down processes. Fluid pyramid connections were adopted to integrate the hierarchical features at different levels of the network, and channel-wise attention modules were exploited to mitigate cross-modal cross-level incompatibility. Furthermore, depth-wise separable convolution was introduced to fuse the cross-modal cross-level features to alleviate the increase in parameters to a great extent. The generalization abilities of BAF-Net were evaluated in terms of two clinical tasks: (1) an in-house PET-CT dataset with 174 patients for differentiation between lung cancer and pulmonary tuberculosis. (2) A public multicenter PET-CT head and neck cancer dataset with 800 patients from nine centers for overall survival prediction.Main results. On the LC-PTB dataset, improved performance was found in BAF-Net (AUC = 0.7342) compared with input-level fusion model (AUC = 0.6825;p< 0.05), feature-level fusion model (AUC = 0.6968;p= 0.0547), output-level fusion model (AUC = 0.7011;p< 0.05). On the H&N cancer dataset, BAF-Net (C-index = 0.7241) outperformed the input-, feature-, and output-level fusion model, with 2.95%, 3.77%, and 1.52% increments of C-index (p= 0.3336, 0.0479 and 0.2911, respectively). The ablation experiments demonstrated the effectiveness of all the designed modules regarding all the evaluated metrics in both datasets.Significance. Extensive experiments on two datasets demonstrated better performance and robustness of BAF-Net than three conventional fusion strategies and PET or CT unimodal network in terms of diagnosis and prognosis.
Collapse
Affiliation(s)
- Huiqin Wu
- Department of Medical Imaging, Guangdong Second Provincial General Hospital, Guangzhou, Guangdong, 518037, People's Republic of China
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
| | - Lihong Peng
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
| | - Dongyang Du
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
| | - Hui Xu
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
| | - Guoyu Lin
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
| | - Zidong Zhou
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
| | - Lijun Lu
- School of Biomedical Engineering and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, Guangdong, 510515, People's Republic of China
- Pazhou Lab, Guangzhou, Guangdong, 510330, People's Republic of China
| | - Wenbing Lv
- School of Information and Yunnan Key Laboratory of Intelligent Systems and Computing, Yunnan University, Kunming, Yunnan, 650504, People's Republic of China
| |
Collapse
|
16
|
Kim DH, Seo J, Lee JH, Jeon ET, Jeong D, Chae HD, Lee E, Kang JH, Choi YH, Kim HJ, Chai JW. Automated Detection and Segmentation of Bone Metastases on Spine MRI Using U-Net: A Multicenter Study. Korean J Radiol 2024; 25:363-373. [PMID: 38528694 PMCID: PMC10973735 DOI: 10.3348/kjr.2023.0671] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2022] [Revised: 12/11/2023] [Accepted: 01/13/2024] [Indexed: 03/27/2024] Open
Abstract
OBJECTIVE To develop and evaluate a deep learning model for automated segmentation and detection of bone metastasis on spinal MRI. MATERIALS AND METHODS We included whole spine MRI scans of adult patients with bone metastasis: 662 MRI series from 302 patients (63.5 ± 11.5 years; male:female, 151:151) from three study centers obtained between January 2015 and August 2021 for training and internal testing (random split into 536 and 126 series, respectively) and 49 MRI series from 20 patients (65.9 ± 11.5 years; male:female, 11:9) from another center obtained between January 2018 and August 2020 for external testing. Three sagittal MRI sequences, including non-contrast T1-weighted image (T1), contrast-enhanced T1-weighted Dixon fat-only image (FO), and contrast-enhanced fat-suppressed T1-weighted image (CE), were used. Seven models trained using the 2D and 3D U-Nets were developed with different combinations (T1, FO, CE, T1 + FO, T1 + CE, FO + CE, and T1 + FO + CE). The segmentation performance was evaluated using Dice coefficient, pixel-wise recall, and pixel-wise precision. The detection performance was analyzed using per-lesion sensitivity and a free-response receiver operating characteristic curve. The performance of the model was compared with that of five radiologists using the external test set. RESULTS The 2D U-Net T1 + CE model exhibited superior segmentation performance in the external test compared to the other models, with a Dice coefficient of 0.699 and pixel-wise recall of 0.653. The T1 + CE model achieved per-lesion sensitivities of 0.828 (497/600) and 0.857 (150/175) for metastases in the internal and external tests, respectively. The radiologists demonstrated a mean per-lesion sensitivity of 0.746 and a mean per-lesion positive predictive value of 0.701 in the external test. CONCLUSION The deep learning models proposed for automated segmentation and detection of bone metastases on spinal MRI demonstrated high diagnostic performance.
Collapse
Affiliation(s)
- Dong Hyun Kim
- Department of Radiology, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
- College of Medicine, Seoul National University, Seoul, Republic of Korea
| | - Jiwoon Seo
- Department of Radiology, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
- College of Medicine, Seoul National University, Seoul, Republic of Korea.
| | - Ji Hyun Lee
- Department of Radiology, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
| | - Eun-Tae Jeon
- Department of Radiology, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
| | | | - Hee Dong Chae
- College of Medicine, Seoul National University, Seoul, Republic of Korea
- Department of Radiology, Seoul National University Hospital, Seoul, Republic of Korea
| | - Eugene Lee
- College of Medicine, Seoul National University, Seoul, Republic of Korea
- Department of Radiology, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
| | - Ji Hee Kang
- Department of Radiology, Konkuk University Medical Center, Seoul, Republic of Korea
| | - Yoon-Hee Choi
- Department of Physical Medicine and Rehabilitation, Soonchunhyang University Seoul Hospital, Seoul, Republic of Korea
| | - Hyo Jin Kim
- Department of Radiology, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
- College of Medicine, Seoul National University, Seoul, Republic of Korea
| | - Jee Won Chai
- Department of Radiology, SMG-SNU Boramae Medical Center, Seoul, Republic of Korea
- College of Medicine, Seoul National University, Seoul, Republic of Korea
| |
Collapse
|
17
|
Ilesanmi AE, Ilesanmi TO, Ajayi BO. Reviewing 3D convolutional neural network approaches for medical image segmentation. Heliyon 2024; 10:e27398. [PMID: 38496891 PMCID: PMC10944240 DOI: 10.1016/j.heliyon.2024.e27398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2023] [Revised: 02/27/2024] [Accepted: 02/28/2024] [Indexed: 03/19/2024] Open
Abstract
Background Convolutional neural networks (CNNs) assume pivotal roles in aiding clinicians in diagnosis and treatment decisions. The rapid evolution of imaging technology has established three-dimensional (3D) CNNs as a formidable framework for delineating organs and anomalies in medical images. The prominence of 3D CNN frameworks is steadily growing within medical image segmentation and classification. Thus, our proposition entails a comprehensive review, encapsulating diverse 3D CNN algorithms for the segmentation of medical image anomalies and organs. Methods This study systematically presents an exhaustive review of recent 3D CNN methodologies. Rigorous screening of abstracts and titles were carried out to establish their relevance. Research papers disseminated across academic repositories were meticulously chosen, analyzed, and appraised against specific criteria. Insights into the realm of anomalies and organ segmentation were derived, encompassing details such as network architecture and achieved accuracies. Results This paper offers an all-encompassing analysis, unveiling the prevailing trends in 3D CNN segmentation. In-depth elucidations encompass essential insights, constraints, observations, and avenues for future exploration. A discerning examination indicates the preponderance of the encoder-decoder network in segmentation tasks. The encoder-decoder framework affords a coherent methodology for the segmentation of medical images. Conclusion The findings of this study are poised to find application in clinical diagnosis and therapeutic interventions. Despite inherent limitations, CNN algorithms showcase commendable accuracy levels, solidifying their potential in medical image segmentation and classification endeavors.
Collapse
Affiliation(s)
- Ademola E. Ilesanmi
- University of Pennsylvania, 3710 Hamilton Walk, 6th Floor, Philadelphia, PA, 19104, United States
| | | | - Babatunde O. Ajayi
- National Astronomical Research Institute of Thailand, Chiang Mai 50180, Thailand
| |
Collapse
|
18
|
Zheng H, Hung ALY, Miao Q, Song W, Scalzo F, Raman SS, Zhao K, Sung K. AtPCa-Net: anatomical-aware prostate cancer detection network on multi-parametric MRI. Sci Rep 2024; 14:5740. [PMID: 38459100 PMCID: PMC10923873 DOI: 10.1038/s41598-024-56405-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2023] [Accepted: 03/06/2024] [Indexed: 03/10/2024] Open
Abstract
Multi-parametric MRI (mpMRI) is widely used for prostate cancer (PCa) diagnosis. Deep learning models show good performance in detecting PCa on mpMRI, but domain-specific PCa-related anatomical information is sometimes overlooked and not fully explored even by state-of-the-art deep learning models, causing potential suboptimal performances in PCa detection. Symmetric-related anatomical information is commonly used when distinguishing PCa lesions from other visually similar but benign prostate tissue. In addition, different combinations of mpMRI findings are used for evaluating the aggressiveness of PCa for abnormal findings allocated in different prostate zones. In this study, we investigate these domain-specific anatomical properties in PCa diagnosis and how we can adopt them into the deep learning framework to improve the model's detection performance. We propose an anatomical-aware PCa detection Network (AtPCa-Net) for PCa detection on mpMRI. Experiments show that the AtPCa-Net can better utilize the anatomical-related information, and the proposed anatomical-aware designs help improve the overall model performance on both PCa detection and patient-level classification.
Collapse
Affiliation(s)
- Haoxin Zheng
- Radiological Sciences, University of California, Los Angeles, Los Angeles, 90095, USA.
- Computer Science, University of California, Los Angeles, Los Angeles, 90095, USA.
| | - Alex Ling Yu Hung
- Radiological Sciences, University of California, Los Angeles, Los Angeles, 90095, USA
- Computer Science, University of California, Los Angeles, Los Angeles, 90095, USA
| | - Qi Miao
- Radiological Sciences, University of California, Los Angeles, Los Angeles, 90095, USA
| | - Weinan Song
- Electrical and Computer Engineering, University of California, Los Angeles, Los Angeles, 90095, USA
| | - Fabien Scalzo
- Computer Science, University of California, Los Angeles, Los Angeles, 90095, USA
- The Seaver College, Pepperdine University, Los Angeles, 90363, USA
| | - Steven S Raman
- Radiological Sciences, University of California, Los Angeles, Los Angeles, 90095, USA
| | - Kai Zhao
- Radiological Sciences, University of California, Los Angeles, Los Angeles, 90095, USA
| | - Kyunghyun Sung
- Radiological Sciences, University of California, Los Angeles, Los Angeles, 90095, USA
| |
Collapse
|
19
|
Wang R, Chow SSL, Serafin RB, Xie W, Han Q, Baraznenok E, Lan L, Bishop KW, Liu JTC. Direct three-dimensional segmentation of prostate glands with nnU-Net. JOURNAL OF BIOMEDICAL OPTICS 2024; 29:036001. [PMID: 38434772 PMCID: PMC10905031 DOI: 10.1117/1.jbo.29.3.036001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Revised: 02/06/2024] [Accepted: 02/09/2024] [Indexed: 03/05/2024]
Abstract
Significance In recent years, we and others have developed non-destructive methods to obtain three-dimensional (3D) pathology datasets of clinical biopsies and surgical specimens. For prostate cancer risk stratification (prognostication), standard-of-care Gleason grading is based on examining the morphology of prostate glands in thin 2D sections. This motivates us to perform 3D segmentation of prostate glands in our 3D pathology datasets for the purposes of computational analysis of 3D glandular features that could offer improved prognostic performance. Aim To facilitate prostate cancer risk assessment, we developed a computationally efficient and accurate deep learning model for 3D gland segmentation based on open-top light-sheet microscopy datasets of human prostate biopsies stained with a fluorescent analog of hematoxylin and eosin (H&E). Approach For 3D gland segmentation based on our H&E-analog 3D pathology datasets, we previously developed a hybrid deep learning and computer vision-based pipeline, called image translation-assisted segmentation in 3D (ITAS3D), which required a complex two-stage procedure and tedious manual optimization of parameters. To simplify this procedure, we use the 3D gland-segmentation masks previously generated by ITAS3D as training datasets for a direct end-to-end deep learning-based segmentation model, nnU-Net. The inputs to this model are 3D pathology datasets of prostate biopsies rapidly stained with an inexpensive fluorescent analog of H&E and the outputs are 3D semantic segmentation masks of the gland epithelium, gland lumen, and surrounding stromal compartments within the tissue. Results nnU-Net demonstrates remarkable accuracy in 3D gland segmentations even with limited training data. Moreover, compared with the previous ITAS3D pipeline, nnU-Net operation is simpler and faster, and it can maintain good accuracy even with lower-resolution inputs. Conclusions Our trained DL-based 3D segmentation model will facilitate future studies to demonstrate the value of computational 3D pathology for guiding critical treatment decisions for patients with prostate cancer.
Collapse
Affiliation(s)
- Rui Wang
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| | - Sarah S. L. Chow
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| | - Robert B. Serafin
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| | - Weisi Xie
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| | - Qinghua Han
- University of Washington, Department of Bioengineering, Seattle, Washington, United States
| | - Elena Baraznenok
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
- University of Washington, Department of Bioengineering, Seattle, Washington, United States
| | - Lydia Lan
- University of Washington, Department of Bioengineering, Seattle, Washington, United States
- University of Washington, Department of Biology, Seattle, Washington, United States
| | - Kevin W. Bishop
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
- University of Washington, Department of Bioengineering, Seattle, Washington, United States
| | - Jonathan T. C. Liu
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
- University of Washington, Department of Bioengineering, Seattle, Washington, United States
- University of Washington, Department of Laboratory Medicine and Pathology, Seattle, Washington, United States
| |
Collapse
|
20
|
Zhao Q, Chang CW, Yang X, Zhao L. Robust explanation supervision for false positive reduction in pulmonary nodule detection. Med Phys 2024; 51:1687-1701. [PMID: 38224306 PMCID: PMC10939846 DOI: 10.1002/mp.16937] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 11/08/2023] [Accepted: 12/12/2023] [Indexed: 01/16/2024] Open
Abstract
BACKGROUND Lung cancer is the deadliest and second most common cancer in the United States due to the lack of symptoms for early diagnosis. Pulmonary nodules are small abnormal regions that can be potentially correlated to the occurrence of lung cancer. Early detection of these nodules is critical because it can significantly improve the patient's survival rates. Thoracic thin-sliced computed tomography (CT) scanning has emerged as a widely used method for diagnosing and prognosis lung abnormalities. PURPOSE The standard clinical workflow of detecting pulmonary nodules relies on radiologists to analyze CT images to assess the risk factors of cancerous nodules. However, this approach can be error-prone due to the various nodule formation causes, such as pollutants and infections. Deep learning (DL) algorithms have recently demonstrated remarkable success in medical image classification and segmentation. As an ever more important assistant to radiologists in nodule detection, it is imperative ensure the DL algorithm and radiologist to better understand the decisions from each other. This study aims to develop a framework integrating explainable AI methods to achieve accurate pulmonary nodule detection. METHODS A robust and explainable detection (RXD) framework is proposed, focusing on reducing false positives in pulmonary nodule detection. Its implementation is based on an explanation supervision method, which uses nodule contours of radiologists as supervision signals to force the model to learn nodule morphologies, enabling improved learning ability on small dataset, and enable small dataset learning ability. In addition, two imputation methods are applied to the nodule region annotations to reduce the noise within human annotations and allow the model to have robust attributions that meet human expectations. The 480, 265, and 265 CT image sets from the public Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) dataset are used for training, validation, and testing. RESULTS Using only 10, 30, 50, and 100 training samples sequentially, our method constantly improves the classification performance and explanation quality of baseline in terms of Area Under the Curve (AUC) and Intersection over Union (IoU). In particular, our framework with a learnable imputation kernel improves IoU from baseline by 24.0% to 80.0%. A pre-defined Gaussian imputation kernel achieves an even greater improvement, from 38.4% to 118.8% from baseline. Compared to the baseline trained on 100 samples, our method shows less drop in AUC when trained on fewer samples. A comprehensive comparison of interpretability shows that our method aligns better with expert opinions. CONCLUSIONS A pulmonary nodule detection framework was demonstrated using public thoracic CT image datasets. The framework integrates the robust explanation supervision (RES) technique to ensure the performance of nodule classification and morphology. The method can reduce the workload of radiologists and enable them to focus on the diagnosis and prognosis of the potential cancerous pulmonary nodules at the early stage to improve the outcomes for lung cancer patients.
Collapse
Affiliation(s)
- Qilong Zhao
- Department of Computer Science, Emory University, Atlanta, GA 30308
| | - Chih-Wei Chang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308
| | - Liang Zhao
- Department of Computer Science, Emory University, Atlanta, GA 30308
| |
Collapse
|
21
|
Fortunati V, Su J, Wolff L, van Doormaal PJ, Hofmeijer J, Martens J, Bokkers RPH, van Zwam WH, van der Lugt A, van Walsum T. Siamese model for collateral score prediction from computed tomography angiography images in acute ischemic stroke. FRONTIERS IN NEUROIMAGING 2024; 2:1239703. [PMID: 38274412 PMCID: PMC10809990 DOI: 10.3389/fnimg.2023.1239703] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Accepted: 12/26/2023] [Indexed: 01/27/2024]
Abstract
Introduction Imaging biomarkers, such as the collateral score as determined from Computed Tomography Angiography (CTA) images, play a role in treatment decision making for acute stroke patients. In this manuscript, we present an end-to-end learning approach for automatic determination of a collateral score from a CTA image. Our aim was to investigate whether such end-to-end learning approaches can be used for this classification task, and whether the resulting classification can be used in existing outcome prediction models. Methods The method consists of a preprocessing step, where the CTA image is aligned to an atlas and divided in the two hemispheres: the affected side and the healthy side. Subsequently, a VoxResNet based convolutional neural network is used to extract features at various resolutions from the input images. This is done by using a Siamese model, such that the classification is driven by the comparison between the affected and healthy using a unique set of features for both hemispheres. After masking the resulting features for both sides with the vascular region and global average pooling (per hemisphere) and concatenation of the resulting features, a fully connected layer is used to determine the categorized collateral score. Experiments Several experiments have been performed to optimize the model hyperparameters and training procedure, and to validate the final model performance. The hyperparameter optimization and subsequent model training was done using CTA images from the MR CLEAN Registry, a Dutch multi-center multi-vendor registry of acute stroke patients that underwent endovascular treatment. A separate set of images, from the MR CLEAN Trial, served as an external validation set, where collateral scoring was assessed and compared with both human observers and a recent more traditional model. In addition, the automated collateral scores have been used in an existing functional outcome prediction model that uses both imaging and non-imaging clinical parameters. Conclusion The results show that end-to-end learning of collateral scoring in CTA images is feasible, and does perform similar to more traditional methods, and the performance also is within the inter-observer variation. Furthermore, the results demonstrate that the end-to-end classification results also can be used in an existing functional outcome prediction model.
Collapse
Affiliation(s)
| | - Jiahang Su
- Biomedical Imaging Group Rotterdam, Department of Radiology & Nuclear Medicine, Erasmus MC, University Medical Center Rotterdam, Rotterdam, Netherlands
| | - Lennard Wolff
- Department of Radiology & Nuclear Medicine, Erasmus MC, University Medical Center Rotterdam, Rotterdam, Netherlands
| | - Pieter-Jan van Doormaal
- Department of Radiology & Nuclear Medicine, Erasmus MC, University Medical Center Rotterdam, Rotterdam, Netherlands
| | - Jeanette Hofmeijer
- Clinical Neurophysiology, MIRA Institute for Biomedical Technology and Technical Medicine, University of Twente, Enschede, Netherlands
- Department of Neurology, Rijnstate Hospital, Arnhem, Netherlands
| | - Jasper Martens
- Department of Radiology and Nuclear Medicine, Rijnstate Hospital, Arnhem, Netherlands
| | | | - Wim H. van Zwam
- Department of Radiology & Nuclear Medicine, Maastricht UMC, Cardiovascular Research Institute Maastricht, Maastricht, Netherlands
| | - Aad van der Lugt
- Department of Radiology & Nuclear Medicine, Erasmus MC, University Medical Center Rotterdam, Rotterdam, Netherlands
| | - Theo van Walsum
- Biomedical Imaging Group Rotterdam, Department of Radiology & Nuclear Medicine, Erasmus MC, University Medical Center Rotterdam, Rotterdam, Netherlands
| |
Collapse
|
22
|
Luan S, Wu K, Wu Y, Zhu B, Wei W, Xue X. Accurate and robust auto-segmentation of head and neck organ-at-risks based on a novel CNN fine-tuning workflow. J Appl Clin Med Phys 2024; 25:e14248. [PMID: 38128058 PMCID: PMC10795444 DOI: 10.1002/acm2.14248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2023] [Revised: 12/08/2023] [Accepted: 12/11/2023] [Indexed: 12/23/2023] Open
Abstract
PURPOSE Obvious inconsistencies in auto-segmentations exist among various AI software. In this study, we have developed a novel convolutional neural network (CNN) fine-tuning workflow to achieve precise and robust localized segmentation. METHODS The datasets include Hubei Cancer Hospital dataset, Cetuximab Head and Neck Public Dataset, and Québec Public Dataset. Seven organs-at-risks (OARs), including brain stem, left parotid gland, esophagus, left optic nerve, optic chiasm, mandible, and pharyngeal constrictor, were selected. The auto-segmentation results from four commercial AI software were first compared with the manual delineations. Then a new multi-scale lightweight residual CNN model with an attention module (named as HN-Net) was trained and tested on 40 samples and 10 samples from Hubei Cancer Hospital, respectively. To enhance the network's accuracy and generalization ability, the fine-tuning workflow utilized an uncertainty estimation method for automatic selection of candidate samples of worthiness from Cetuximab Head and Neck Public Dataset for further training. The segmentation performances were evaluated on the Hubei Cancer Hospital dataset and/or the entire Québec Public Dataset. RESULTS A maximum difference of 0.13 and 0.7 mm in average Dice value and Hausdorff distance value for the seven OARs were observed by four AI software. The proposed HN-Net achieved an average Dice value of 0.14 higher than that of the AI software, and it also outperformed other popular CNN models (HN-Net: 0.79, U-Net: 0.78, U-Net++: 0.78, U-Net-Multi-scale: 0.77, AI software: 0.65). Additionally, the HN-Net fine-tuning workflow by using the local datasets and external public datasets further improved the automatic segmentation with the average Dice value by 0.02. CONCLUSION The delineations of commercial AI software need to be carefully reviewed, and localized further training is necessary for clinical practice. The proposed fine-tuning workflow could be feasibly adopted to implement an accurate and robust auto-segmentation model by using local datasets and external public datasets.
Collapse
Affiliation(s)
- Shunyao Luan
- Department of Radiation OncologyHubei Cancer Hospital, Tongji Medical CollegeHuazhong University of Science and TechnologyWuhanChina
- School of Integrated CircuitsLaboratory for OptoelectronicsHuazhong University of Science and TechnologyWuhanChina
| | - Kun Wu
- Department of Radiation OncologyHubei Cancer Hospital, Tongji Medical CollegeHuazhong University of Science and TechnologyWuhanChina
| | - Yuan Wu
- Department of Radiation OncologyHubei Cancer Hospital, Tongji Medical CollegeHuazhong University of Science and TechnologyWuhanChina
| | - Benpeng Zhu
- School of Integrated CircuitsLaboratory for OptoelectronicsHuazhong University of Science and TechnologyWuhanChina
| | - Wei Wei
- Department of Radiation OncologyHubei Cancer Hospital, Tongji Medical CollegeHuazhong University of Science and TechnologyWuhanChina
| | - Xudong Xue
- Department of Radiation OncologyHubei Cancer Hospital, Tongji Medical CollegeHuazhong University of Science and TechnologyWuhanChina
| |
Collapse
|
23
|
Xu B, Yang J, Hong P, Fan X, Sun Y, Zhang L, Yang B, Xu L, Avolio A. Coronary artery segmentation in CCTA images based on multi-scale feature learning. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:973-991. [PMID: 38943423 DOI: 10.3233/xst-240093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/01/2024]
Abstract
BACKGROUND Coronary artery segmentation is a prerequisite in computer-aided diagnosis of Coronary Artery Disease (CAD). However, segmentation of coronary arteries in Coronary Computed Tomography Angiography (CCTA) images faces several challenges. The current segmentation approaches are unable to effectively address these challenges and existing problems such as the need for manual interaction or low segmentation accuracy. OBJECTIVE A Multi-scale Feature Learning and Rectification (MFLR) network is proposed to tackle the challenges and achieve automatic and accurate segmentation of coronary arteries. METHODS The MFLR network introduces a multi-scale feature extraction module in the encoder to effectively capture contextual information under different receptive fields. In the decoder, a feature correction and fusion module is proposed, which employs high-level features containing multi-scale information to correct and guide low-level features, achieving fusion between the two-level features to further improve segmentation performance. RESULTS The MFLR network achieved the best performance on the dice similarity coefficient, Jaccard index, Recall, F1-score, and 95% Hausdorff distance, for both in-house and public datasets. CONCLUSION Experimental results demonstrate the superiority and good generalization ability of the MFLR approach. This study contributes to the accurate diagnosis and treatment of CAD, and it also informs other segmentation applications in medicine.
Collapse
Affiliation(s)
- Bu Xu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Jinzhong Yang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Peng Hong
- Software College, Northeastern University, Shenyang, China
| | - Xiaoxue Fan
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yu Sun
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of North Theater Command, Shenyang, China
| | - Libo Zhang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of North Theater Command, Shenyang, China
| | - Benqiang Yang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of North Theater Command, Shenyang, China
| | - Lisheng Xu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Medical Image Computing, Ministry of Education, Shenyang, China
- Engineering Research Center of Medical Imaging and Intelligent Analysis, Ministry of Education, Shenyang, China
| | - Alberto Avolio
- Macquarie Medical School, Faculty of Medicine, Health and Human Sciences, Macquarie University, Sydney, Australia
| |
Collapse
|
24
|
Yu Hung AL, Zheng H, Zhao K, Du X, Pang K, Miao Q, Raman SS, Terzopoulos D, Sung K. CSAM: A 2.5D Cross-Slice Attention Module for Anisotropic Volumetric Medical Image Segmentation. IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION. IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION 2024; 2024:5911-5920. [PMID: 39193208 PMCID: PMC11349312 DOI: 10.1109/wacv57701.2024.00582] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/29/2024]
Abstract
A large portion of volumetric medical data, especially magnetic resonance imaging (MRI) data, is anisotropic, as the through-plane resolution is typically much lower than the in-plane resolution. Both 3D and purely 2D deep learning-based segmentation methods are deficient in dealing with such volumetric data since the performance of 3D methods suffers when confronting anisotropic data, and 2D methods disregard crucial volumetric information. Insufficient work has been done on 2.5D methods, in which 2D convolution is mainly used in concert with volumetric information. These models focus on learning the relationship across slices, but typically have many parameters to train. We offer a Cross-Slice Attention Module (CSAM) with minimal trainable parameters, which captures information across all the slices in the volume by applying semantic, positional, and slice attention on deep feature maps at different scales. Our extensive experiments using different network architectures and tasks demonstrate the usefulness and generalizability of CSAM. Associated code is available at https://github.com/aL3x-O-o-Hung/CSAM.
Collapse
Affiliation(s)
| | | | - Kai Zhao
- University of California, Los Angeles
| | - Xiaoxi Du
- University of California, Los Angeles
| | | | - Qi Miao
- University of California, Los Angeles
| | | | | | | |
Collapse
|
25
|
Mhlanga ST, Viriri S. Deep learning techniques for isointense infant brain tissue segmentation: a systematic literature review. Front Med (Lausanne) 2023; 10:1240360. [PMID: 38193036 PMCID: PMC10773803 DOI: 10.3389/fmed.2023.1240360] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Accepted: 11/01/2023] [Indexed: 01/10/2024] Open
Abstract
Introduction To improve comprehension of initial brain growth in wellness along with sickness, it is essential to precisely segment child brain magnetic resonance imaging (MRI) into white matter (WM) and gray matter (GM), along with cerebrospinal fluid (CSF). Nonetheless, in the isointense phase (6-8 months of age), the inborn myelination and development activities, WM along with GM display alike stages of intensity in both T1-weighted and T2-weighted MRI, making tissue segmentation extremely difficult. Methods The comprehensive review of studies related to isointense brain MRI segmentation approaches is highlighted in this publication. The main aim and contribution of this study is to aid researchers by providing a thorough review to make their search for isointense brain MRI segmentation easier. The systematic literature review is performed from four points of reference: (1) review of studies concerning isointense brain MRI segmentation; (2) research contribution and future works and limitations; (3) frequently applied evaluation metrics and datasets; (4) findings of this studies. Results and discussion The systemic review is performed on studies that were published in the period of 2012 to 2022. A total of 19 primary studies of isointense brain MRI segmentation were selected to report the research question stated in this review.
Collapse
Affiliation(s)
| | - Serestina Viriri
- School of Mathematics, Statistics and Computer Science, University of KwaZulu-Natal, Durban, South Africa
| |
Collapse
|
26
|
Saha PK, Nadeem SA, Comellas AP. A Survey on Artificial Intelligence in Pulmonary Imaging. WILEY INTERDISCIPLINARY REVIEWS. DATA MINING AND KNOWLEDGE DISCOVERY 2023; 13:e1510. [PMID: 38249785 PMCID: PMC10796150 DOI: 10.1002/widm.1510] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/18/2022] [Accepted: 06/21/2023] [Indexed: 01/23/2024]
Abstract
Over the last decade, deep learning (DL) has contributed a paradigm shift in computer vision and image recognition creating widespread opportunities of using artificial intelligence in research as well as industrial applications. DL has been extensively studied in medical imaging applications, including those related to pulmonary diseases. Chronic obstructive pulmonary disease, asthma, lung cancer, pneumonia, and, more recently, COVID-19 are common lung diseases affecting nearly 7.4% of world population. Pulmonary imaging has been widely investigated toward improving our understanding of disease etiologies and early diagnosis and assessment of disease progression and clinical outcomes. DL has been broadly applied to solve various pulmonary image processing challenges including classification, recognition, registration, and segmentation. This paper presents a survey of pulmonary diseases, roles of imaging in translational and clinical pulmonary research, and applications of different DL architectures and methods in pulmonary imaging with emphasis on DL-based segmentation of major pulmonary anatomies such as lung volumes, lung lobes, pulmonary vessels, and airways as well as thoracic musculoskeletal anatomies related to pulmonary diseases.
Collapse
Affiliation(s)
- Punam K Saha
- Departments of Radiology and Electrical and Computer Engineering, University of Iowa, Iowa City, IA, 52242
| | | | | |
Collapse
|
27
|
Sun H, Yang S, Chen L, Liao P, Liu X, Liu Y, Wang N. Brain tumor image segmentation based on improved FPN. BMC Med Imaging 2023; 23:172. [PMID: 37904116 PMCID: PMC10617057 DOI: 10.1186/s12880-023-01131-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2022] [Accepted: 10/19/2023] [Indexed: 11/01/2023] Open
Abstract
PURPOSE Automatic segmentation of brain tumors by deep learning algorithm is one of the research hotspots in the field of medical image segmentation. An improved FPN network for brain tumor segmentation is proposed to improve the segmentation effect of brain tumor. MATERIALS AND METHODS Aiming at the problem that the traditional full convolutional neural network (FCN) has weak processing ability, which leads to the loss of details in tumor segmentation, this paper proposes a brain tumor image segmentation method based on the improved feature pyramid networks (FPN) convolutional neural network. In order to improve the segmentation effect of brain tumors, we improved the model, introduced the FPN structure into the U-Net structure, captured the context multi-scale information by using the different scale information in the U-Net model and the multi receptive field high-level features in the FPN convolutional neural network, and improved the adaptability of the model to different scale features. RESULTS Performance evaluation indicators show that the proposed improved FPN model has 99.1% accuracy, 92% DICE rating and 86% Jaccard index. The performance of the proposed method outperforms other segmentation models in each metric. In addition, the schematic diagram of the segmentation results shows that the segmentation results of our algorithm are closer to the ground truth, showing more brain tumour details, while the segmentation results of other algorithms are smoother. CONCLUSIONS The experimental results show that this method can effectively segment brain tumor regions and has certain generalization, and the segmentation effect is better than other networks. It has positive significance for clinical diagnosis of brain tumors.
Collapse
Affiliation(s)
- Haitao Sun
- Department of Radiotherapy Room, Zhongshan Hospital of Traditional Chinese Medicine, ZhongShanGuangdong Province, 528400, China
| | - Shuai Yang
- Department of Radiotherapy and Minimally Invasive Surgery, The Cancer Center of The Fifth Affiliated Hospital of Sun Yat-Sen University, Zhuhai, 519020, China
| | - Lijuan Chen
- Department of Radiotherapy Room, Zhongshan Hospital of Traditional Chinese Medicine, ZhongShanGuangdong Province, 528400, China
| | - Pingyan Liao
- Department of Radiotherapy Room, Zhongshan Hospital of Traditional Chinese Medicine, ZhongShanGuangdong Province, 528400, China
| | - Xiangping Liu
- Department of Radiotherapy Room, Zhongshan Hospital of Traditional Chinese Medicine, ZhongShanGuangdong Province, 528400, China
| | - Ying Liu
- Department of the Radiotherapy, The Fifth Affiliated Hospital of Guangzhou Medical University, Guangzhou, 510060, China
| | - Ning Wang
- Department of Radiotherapy Room, Zhongshan Hospital of Traditional Chinese Medicine, ZhongShanGuangdong Province, 528400, China.
| |
Collapse
|
28
|
Yang J, Jiang H, Tassew T, Sun P, Ma J, Xia Y, Yap PT, Chen G. Towards Accurate Microstructure Estimation via 3D Hybrid Graph Transformer. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2023; 14227:25-34. [PMID: 39219989 PMCID: PMC11361334 DOI: 10.1007/978-3-031-43993-3_3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
Deep learning has drawn increasing attention in microstructure estimation with undersampled diffusion MRI (dMRI) data. A representative method is the hybrid graph transformer (HGT), which achieves promising performance by integrating q -space graph learning and x -space transformer learning into a unified framework. However, this method overlooks the 3D spatial information as it relies on training with 2D slices. To address this limitation, we propose 3D hybrid graph transformer (3D-HGT), an advanced microstructure estimation model capable of making full use of 3D spatial information and angular information. To tackle the large computation burden associated with 3D x -space learning, we propose an efficient q -space learning model based on simplified graph neural networks. Furthermore, we propose a 3D x -space learning module based on the transformer. Extensive experiments on data from the human connectome project show that our 3D-HGT outperforms state-of-the-art methods, including HGT, in both quantitative and qualitative evaluations.
Collapse
Affiliation(s)
- Junqing Yang
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an, China
| | - Haotian Jiang
- School of Computer Science and Technology, Heilongjiang University, Harbin, China
| | - Tewodros Tassew
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an, China
| | - Peng Sun
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an, China
| | - Jiquan Ma
- School of Computer Science and Technology, Heilongjiang University, Harbin, China
| | - Yong Xia
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an, China
| | - Pew-Thian Yap
- Department of Radiology, University of North Carolina, Chapel Hill, NC, USA
- Biomedical Research Imaging Center, University of North Carolina, Chapel Hill, NC, USA
| | - Geng Chen
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an, China
| |
Collapse
|
29
|
Zeng A, Wu C, Lin G, Xie W, Hong J, Huang M, Zhuang J, Bi S, Pan D, Ullah N, Khan KN, Wang T, Shi Y, Li X, Xu X. ImageCAS: A large-scale dataset and benchmark for coronary artery segmentation based on computed tomography angiography images. Comput Med Imaging Graph 2023; 109:102287. [PMID: 37634975 DOI: 10.1016/j.compmedimag.2023.102287] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2023] [Revised: 05/03/2023] [Accepted: 08/03/2023] [Indexed: 08/29/2023]
Abstract
Cardiovascular disease (CVD) accounts for about half of non-communicable diseases. Vessel stenosis in the coronary artery is considered to be the major risk of CVD. Computed tomography angiography (CTA) is one of the widely used noninvasive imaging modalities in coronary artery diagnosis due to its superior image resolution. Clinically, segmentation of coronary arteries is essential for the diagnosis and quantification of coronary artery disease. Recently, a variety of works have been proposed to address this problem. However, on one hand, most works rely on in-house datasets, and only a few works published their datasets to the public which only contain tens of images. On the other hand, their source code have not been published, and most follow-up works have not made comparison with existing works, which makes it difficult to judge the effectiveness of the methods and hinders the further exploration of this challenging yet critical problem in the community. In this paper, we propose a large-scale dataset for coronary artery segmentation on CTA images. In addition, we have implemented a benchmark in which we have tried our best to implement several typical existing methods. Furthermore, we propose a strong baseline method which combines multi-scale patch fusion and two-stage processing to extract the details of vessels. Comprehensive experiments show that the proposed method achieves better performance than existing works on the proposed large-scale dataset. The benchmark and the dataset are published at https://github.com/XiaoweiXu/ImageCAS-A-Large-Scale-Dataset-and-Benchmark-for-Coronary-Artery-Segmentation-based-on-CT.
Collapse
Affiliation(s)
- An Zeng
- School of Computer Science, Guangdong University of Technology, Guangzhou, China
| | - Chunbiao Wu
- School of Computer Science, Guangdong University of Technology, Guangzhou, China
| | - Guisen Lin
- Department of Radiology, Shenzhen Children's Hospital, Shenzhen, China
| | - Wen Xie
- Guangdong Provincial Key Laboratory of South China Structural Heart Disease, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, 510080, China
| | - Jin Hong
- Guangdong Provincial Key Laboratory of South China Structural Heart Disease, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, 510080, China
| | - Meiping Huang
- Guangdong Provincial Key Laboratory of South China Structural Heart Disease, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, 510080, China
| | - Jian Zhuang
- Guangdong Provincial Key Laboratory of South China Structural Heart Disease, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, 510080, China
| | - Shanshan Bi
- Department of Computer Science and Engineering, Missouri University of Science and Technology, Rolla, MO, United States
| | - Dan Pan
- Department of Computer Science, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Najeeb Ullah
- Department of Computer Science, University of Engineering and Technology, Mardan, KP, Pakistan
| | - Kaleem Nawaz Khan
- Department of Computer Science, University of Engineering and Technology, Mardan, KP, Pakistan
| | - Tianchen Wang
- Department of Computer Science and Engineering, University of Notre Dame, Indiana, United States
| | - Yiyu Shi
- Department of Computer Science and Engineering, University of Notre Dame, Indiana, United States
| | - Xiaomeng Li
- Department of Electronic and Computer Engineering, The Hong Kong University of Science and Technology, Hong Kong Special Administrative Region, China
| | - Xiaowei Xu
- Guangdong Provincial Key Laboratory of South China Structural Heart Disease, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, 510080, China.
| |
Collapse
|
30
|
Dou M, Chen Z, Tang Y, Sheng L, Zhou J, Wang X, Yao Y. Segmentation of rectal tumor from multi-parametric MRI images using an attention-based fusion network. Med Biol Eng Comput 2023; 61:2379-2389. [PMID: 37084029 DOI: 10.1007/s11517-023-02828-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Accepted: 03/08/2023] [Indexed: 04/22/2023]
Abstract
Accurate segmentation of rectal tumors is the most crucial task in determining the stage of rectal cancer and developing suitable therapies. However, complex image backgrounds, irregular edge, and poor contrast hinder the related research. This study presents an attention-based multi-modal fusion module to effectively integrate complementary information from different MRI images and suppress redundancy. In addition, a deep learning-based segmentation model (AF-UNet) is designed to achieve accurate segmentation of rectal tumors. This model takes multi-parametric MRI images as input and effectively integrates the features from different multi-parametric MRI images by embedding the attention fusion module. Finally, three types of MRI images (T2, ADC, DWI) of 250 patients with rectal cancer were collected, with the tumor regions delineated by two oncologists. The experimental results show that the proposed method is superior to the most advanced image segmentation method with a Dice coefficient of [Formula: see text], which is also better than other multi-modal fusion methods. Framework of the AF-UNet. This model takes multi-modal MRI images as input, and integrates complementary information using attention mechanism and suppresses redundancy.
Collapse
Affiliation(s)
- Meng Dou
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Zhebin Chen
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Yuanling Tang
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Leiming Sheng
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Jitao Zhou
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Xin Wang
- Department of Abdominal Oncology, West China Hospital, Sichuan University, Chengdu, China
| | - Yu Yao
- Chengdu Institute of Computer Application, Chinese Academy of Sciences, Chengdu, China.
- University of Chinese Academy of Sciences, Beijing, China.
| |
Collapse
|
31
|
Liu L, Chang J, Liu Z, Zhang P, Xu X, Shang H. Hybrid Contextual Semantic Network for Accurate Segmentation and Detection of Small-Size Stroke Lesions From MRI. IEEE J Biomed Health Inform 2023; 27:4062-4073. [PMID: 37155390 DOI: 10.1109/jbhi.2023.3273771] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Stroke is a cerebrovascular disease with high mortality and disability rates. The occurrence of the stroke typically produces lesions of different sizes, with the accurate segmentation and detection of small-size stroke lesions being closely related to the prognosis of patients. However, the large lesions are usually correctly identified, the small-size lesions are usually ignored. This article provides a hybrid contextual semantic network (HCSNet) that can accurately and simultaneously segment and detect small-size stroke lesions from magnetic resonance images. HCSNet inherits the advantages of the encoder-decoder architecture and applies a novel hybrid contextual semantic module that generates high-quality contextual semantic features from the spatial and channel contextual semantic features through the skip connection layer. Moreover, a mixing-loss function is proposed to optimize HCSNet for unbalanced small-size lesions. HCSNet is trained and evaluated on 2D magnetic resonance images produced from the Anatomical Tracings of Lesions After Stroke challenge (ATLAS R2.0). Extensive experiments demonstrate that HCSNet outperforms several other state-of-the-art methods in its ability to segment and detect small-size stroke lesions. Visualization and ablation experiments reveal that the hybrid semantic module improves the segmentation and detection performance of HCSNet.
Collapse
|
32
|
Tian F, Tian Z, Chen Z, Zhang D, Du S. Surface-GCN: Learning interaction experience for organ segmentation in 3D medical images. Med Phys 2023; 50:5030-5044. [PMID: 36738103 DOI: 10.1002/mp.16280] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2022] [Revised: 12/26/2022] [Accepted: 01/13/2023] [Indexed: 02/05/2023] Open
Abstract
BACKGROUND Accurate segmentation of organs has a great significance for clinical diagnosis, but it is still hard work due to the obscure imaging boundaries caused by tissue adhesion on medical images. Based on the image continuity in medical image volumes, segmentation on these slices could be inferred from adjacent slices with a clear organ boundary. Radiologists can delineate a clear organ boundary by observing adjacent slices. PURPOSE Inspired by the radiologists' delineating procedure, we design an organ segmentation model based on boundary information of adjacent slices and a human-machine interactive learning strategy to introduce clinical experience. METHODS We propose an interactive organ segmentation method for medical image volume based on Graph Convolution Network (GCN) called Surface-GCN. First, we propose a Surface Feature Extraction Network (SFE-Net) to capture surface features of a target organ, and supervise it by a Mini-batch Adaptive Surface Matching (MBASM) module. Then, to predict organ boundaries precisely, we design an automatic segmentation module based on a Surface Convolution Unit (SCU), which propagates information on organ surfaces to refine the generated boundaries. In addition, an interactive segmentation module is proposed to learn radiologists' experience of interactive corrections on organ surfaces to reduce interaction clicks. RESULTS We evaluate the proposed method on one prostate MR image dataset and two abdominal multi-organ CT datasets. The experimental results show that our method outperforms other state-of-the-art methods. For prostate segmentation, the proposed method conducts a DSC score of 94.49% on PROMISE12 test dataset. For abdominal multi-organ segmentation, the proposed method achieves DSC scores of 95, 91, 95, and 88% for the left kidney, gallbladder, spleen, and esophagus, respectively. For interactive segmentation, the proposed method reduces 5-10 interaction clicks to reach the same accuracy. CONCLUSIONS To overcome the medical organ segmentation challenge, we propose a Graph Convolutional Network called Surface-GCN by imitating radiologist interactions and learning clinical experience. On single and multiple organ segmentation tasks, the proposed method could obtain more accurate segmentation boundaries compared with other state-of-the-art methods.
Collapse
Affiliation(s)
- Fengrui Tian
- School of Software Engineering, Xi'an Jiaotong University, Xi'an, China
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Zhiqiang Tian
- School of Software Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Zhang Chen
- School of Software Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Dong Zhang
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
- School of Automation Science and Engineering, Xi'an Jiaotong University, Xi'an, China
| | - Shaoyi Du
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| |
Collapse
|
33
|
Lim CC, Ling AHW, Chong YF, Mashor MY, Alshantti K, Aziz ME. Comparative Analysis of Image Processing Techniques for Enhanced MRI Image Quality: 3D Reconstruction and Segmentation Using 3D U-Net Architecture. Diagnostics (Basel) 2023; 13:2377. [PMID: 37510120 PMCID: PMC10377862 DOI: 10.3390/diagnostics13142377] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 06/29/2023] [Accepted: 07/05/2023] [Indexed: 07/30/2023] Open
Abstract
Osteosarcoma is a common type of bone tumor, particularly prevalent in children and adolescents between the ages of 5 and 25 who are experiencing growth spurts during puberty. Manual delineation of tumor regions in MRI images can be laborious and time-consuming, and results may be subjective and difficult to replicate. Therefore, a convolutional neural network (CNN) was developed to automatically segment osteosarcoma cancerous cells in three types of MRI images. The study consisted of five main stages. First, 3692 DICOM format MRI images were acquired from 46 patients, including T1-weighted, T2-weighted, and T1-weighted with injection of Gadolinium (T1W + Gd) images. Contrast stretching and median filter were applied to enhance image intensity and remove noise, and the pre-processed images were reconstructed into NIfTI format files for deep learning. The MRI images were then transformed to fit the CNN's requirements. A 3D U-Net architecture was proposed with optimized parameters to build an automatic segmentation model capable of segmenting osteosarcoma from the MRI images. The 3D U-Net segmentation model achieved excellent results, with mean dice similarity coefficients (DSC) of 83.75%, 85.45%, and 87.62% for T1W, T2W, and T1W + Gd images, respectively. However, the study found that the proposed method had some limitations, including poorly defined borders, missing lesion portions, and other confounding factors. In summary, an automatic segmentation method based on a CNN has been developed to address the challenge of manually segmenting osteosarcoma cancerous cells in MRI images. While the proposed method showed promise, the study revealed limitations that need to be addressed to improve its efficacy.
Collapse
Affiliation(s)
- Chee Chin Lim
- Faculty of Electronic Engineering & Technology, Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
- Sport Engineering Research Centre (SERC), Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
| | - Apple Ho Wei Ling
- Faculty of Electronic Engineering & Technology, Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
| | - Yen Fook Chong
- Sport Engineering Research Centre (SERC), Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
| | - Mohd Yusoff Mashor
- Faculty of Electronic Engineering & Technology, Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
- Sport Engineering Research Centre (SERC), Universiti Malaysia Perlis, Arau 02600, Perlis, Malaysia
| | | | - Mohd Ezane Aziz
- Department of Radiology, Universiti Sains Malaysia, Kubang Kerian 16150, Kelantan, Malaysia
| |
Collapse
|
34
|
Disagreement attention: Let us agree to disagree on computed tomography segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104769] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/19/2023]
|
35
|
Oh JH, Kim HG, Lee KM. Developing and Evaluating Deep Learning Algorithms for Object Detection: Key Points for Achieving Superior Model Performance. Korean J Radiol 2023; 24:698-714. [PMID: 37404112 DOI: 10.3348/kjr.2022.0765] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2022] [Revised: 04/29/2023] [Accepted: 05/16/2023] [Indexed: 07/06/2023] Open
Abstract
In recent years, artificial intelligence, especially object detection-based deep learning in computer vision, has made significant advancements, driven by the development of computing power and the widespread use of graphic processor units. Object detection-based deep learning techniques have been applied in various fields, including the medical imaging domain, where remarkable achievements have been reported in disease detection. However, the application of deep learning does not always guarantee satisfactory performance, and researchers have been employing trial-and-error to identify the factors contributing to performance degradation and enhance their models. Moreover, due to the black-box problem, the intermediate processes of a deep learning network cannot be comprehended by humans; as a result, identifying problems in a deep learning model that exhibits poor performance can be challenging. This article highlights potential issues that may cause performance degradation at each deep learning step in the medical imaging domain and discusses factors that must be considered to improve the performance of deep learning models. Researchers who wish to begin deep learning research can reduce the required amount of trial-and-error by understanding the issues discussed in this study.
Collapse
Affiliation(s)
- Jang-Hoon Oh
- Department of Radiology, Kyung Hee University Hospital, Kyung Hee University College of Medicine, Seoul, Korea
| | - Hyug-Gi Kim
- Department of Radiology, Kyung Hee University Hospital, Kyung Hee University College of Medicine, Seoul, Korea
| | - Kyung Mi Lee
- Department of Radiology, Kyung Hee University Hospital, Kyung Hee University College of Medicine, Seoul, Korea.
| |
Collapse
|
36
|
Zeng X, Guo Y, Zaman A, Hassan H, Lu J, Xu J, Yang H, Miao X, Cao A, Yang Y, Chen R, Kang Y. Tubular Structure Segmentation via Multi-Scale Reverse Attention Sparse Convolution. Diagnostics (Basel) 2023; 13:2161. [PMID: 37443556 DOI: 10.3390/diagnostics13132161] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2023] [Revised: 06/17/2023] [Accepted: 06/20/2023] [Indexed: 07/15/2023] Open
Abstract
Cerebrovascular and airway structures are tubular structures used for transporting blood and gases, respectively, providing essential support for the normal activities of the human body. Accurately segmenting these tubular structures is the basis of morphology research and pathological detection. Nevertheless, accurately segmenting these structures from images presents great challenges due to their complex morphological and topological characteristics. To address this challenge, this paper proposes a framework UARAI based on the U-Net multi-scale reverse attention network and sparse convolution network. The framework utilizes a multi-scale structure to effectively extract the global and deep detail features of vessels and airways. Further, it enhances the extraction ability of fine-edged features by a joint reverse attention module. In addition, the sparse convolution structure is introduced to improve the features' expression ability without increasing the model's complexity. Finally, the proposed training sample cropping strategy reduces the influence of block boundaries on the accuracy of tubular structure segmentation. The experimental findings demonstrate that the UARAI-based metrics, namely Dice and IoU, achieve impressive scores of 90.31% and 82.33% for cerebrovascular segmentation and 93.34% and 87.51% for airway segmentation, respectively. Compared to commonly employed segmentation techniques, the proposed method exhibits remarkable accuracy and robustness in delineating tubular structures such as cerebrovascular and airway structures. These results hold significant promise in facilitating medical image analysis and clinical diagnosis, offering invaluable support to healthcare professionals.
Collapse
Affiliation(s)
- Xueqiang Zeng
- School of Applied Technology, Shenzhen University, Shenzhen 518060, China
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
| | - Yingwei Guo
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China
| | - Asim Zaman
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
- School of Biomedical Engineering, Medical School, Shenzhen University, Shenzhen 518060, China
| | - Haseeb Hassan
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
| | - Jiaxi Lu
- School of Applied Technology, Shenzhen University, Shenzhen 518060, China
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
| | - Jiaxuan Xu
- State Key Laboratory of Respiratory Disease, National Center for Respiratory Medicine, National Clinical Research Center for Respiratory Disease, First Affiliated Hospital, Guangzhou Medical University, Guangzhou 510120, China
| | - Huihui Yang
- School of Applied Technology, Shenzhen University, Shenzhen 518060, China
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
| | - Xiaoqiang Miao
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China
| | - Anbo Cao
- School of Applied Technology, Shenzhen University, Shenzhen 518060, China
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
| | - Yingjian Yang
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China
| | - Rongchang Chen
- Shenzhen Institute of Respiratory Diseases, Shenzhen People's Hospital, Shenzhen 518001, China
- The Second Clinical Medical College, Jinan University, Guangzhou 518001, China
- The First Affiliated Hospital, Southern University of Science and Technology, Shenzhen 518001, China
| | - Yan Kang
- School of Applied Technology, Shenzhen University, Shenzhen 518060, China
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen 518118, China
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China
- Engineering Research Centre of Medical Imaging and Intelligent Analysis, Ministry of Education, Shenyang 110169, China
| |
Collapse
|
37
|
Vahedifard F, Adepoju JO, Supanich M, Ai HA, Liu X, Kocak M, Marathu KK, Byrd SE. Review of deep learning and artificial intelligence models in fetal brain magnetic resonance imaging. World J Clin Cases 2023; 11:3725-3735. [PMID: 37383127 PMCID: PMC10294149 DOI: 10.12998/wjcc.v11.i16.3725] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/04/2023] [Revised: 01/30/2023] [Accepted: 05/06/2023] [Indexed: 06/02/2023] Open
Abstract
Central nervous system abnormalities in fetuses are fairly common, happening in 0.1% to 0.2% of live births and in 3% to 6% of stillbirths. So initial detection and categorization of fetal Brain abnormalities are critical. Manually detecting and segmenting fetal brain magnetic resonance imaging (MRI) could be time-consuming, and susceptible to interpreter experience. Artificial intelligence (AI) algorithms and machine learning approaches have a high potential for assisting in the early detection of these problems, improving the diagnosis process and follow-up procedures. The use of AI and machine learning techniques in fetal brain MRI was the subject of this narrative review paper. Using AI, anatomic fetal brain MRI processing has investigated models to predict specific landmarks and segmentation automatically. All gestation age weeks (17-38 wk) and different AI models (mainly Convolutional Neural Network and U-Net) have been used. Some models' accuracy achieved 95% and more. AI could help preprocess and post-process fetal images and reconstruct images. Also, AI can be used for gestational age prediction (with one-week accuracy), fetal brain extraction, fetal brain segmentation, and placenta detection. Some fetal brain linear measurements, such as Cerebral and Bone Biparietal Diameter, have been suggested. Classification of brain pathology was studied using diagonal quadratic discriminates analysis, K-nearest neighbor, random forest, naive Bayes, and radial basis function neural network classifiers. Deep learning methods will become more powerful as more large-scale, labeled datasets become available. Having shared fetal brain MRI datasets is crucial because there aren not many fetal brain pictures available. Also, physicians should be aware of AI's function in fetal brain MRI, particularly neuroradiologists, general radiologists, and perinatologists.
Collapse
Affiliation(s)
- Farzan Vahedifard
- Department of Diagnostic Radiology and Nuclear Medicine, Rush Medical College, Chicago, IL 606012, United States
| | - Jubril O Adepoju
- Department of Diagnostic Radiology and Nuclear Medicine, Rush Medical College, Chicago, IL 606012, United States
| | - Mark Supanich
- Division for Diagnostic Medical Physics, Department of Radiology and Nuclear Medicine, Rush University Medical Center, Chicago, IL 606012, United States
| | - Hua Asher Ai
- Division for Diagnostic Medical Physics, Department of Radiology and Nuclear Medicine, Rush University Medical Center, Chicago, IL 606012, United States
| | - Xuchu Liu
- Department of Diagnostic Radiology and Nuclear Medicine, Rush Medical College, Chicago, IL 606012, United States
| | - Mehmet Kocak
- Department of Diagnostic Radiology and Nuclear Medicine, Rush Medical College, Chicago, IL 606012, United States
| | - Kranthi K Marathu
- Department of Diagnostic Radiology and Nuclear Medicine, Rush Medical College, Chicago, IL 606012, United States
| | - Sharon E Byrd
- Department of Diagnostic Radiology and Nuclear Medicine, Rush Medical College, Chicago, IL 606012, United States
| |
Collapse
|
38
|
Okyere FG, Cudjoe D, Sadeghi-Tehran P, Virlet N, Riche AB, Castle M, Greche L, Mohareb F, Simms D, Mhada M, Hawkesford MJ. Machine Learning Methods for Automatic Segmentation of Images of Field- and Glasshouse-Based Plants for High-Throughput Phenotyping. PLANTS (BASEL, SWITZERLAND) 2023; 12:2035. [PMID: 37653952 PMCID: PMC10224253 DOI: 10.3390/plants12102035] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Revised: 05/03/2023] [Accepted: 05/10/2023] [Indexed: 07/15/2023]
Abstract
Image segmentation is a fundamental but critical step for achieving automated high- throughput phenotyping. While conventional segmentation methods perform well in homogenous environments, the performance decreases when used in more complex environments. This study aimed to develop a fast and robust neural-network-based segmentation tool to phenotype plants in both field and glasshouse environments in a high-throughput manner. Digital images of cowpea (from glasshouse) and wheat (from field) with different nutrient supplies across their full growth cycle were acquired. Image patches from 20 randomly selected images from the acquired dataset were transformed from their original RGB format to multiple color spaces. The pixels in the patches were annotated as foreground and background with a pixel having a feature vector of 24 color properties. A feature selection technique was applied to choose the sensitive features, which were used to train a multilayer perceptron network (MLP) and two other traditional machine learning models: support vector machines (SVMs) and random forest (RF). The performance of these models, together with two standard color-index segmentation techniques (excess green (ExG) and excess green-red (ExGR)), was compared. The proposed method outperformed the other methods in producing quality segmented images with over 98%-pixel classification accuracy. Regression models developed from the different segmentation methods to predict Soil Plant Analysis Development (SPAD) values of cowpea and wheat showed that images from the proposed MLP method produced models with high predictive power and accuracy comparably. This method will be an essential tool for the development of a data analysis pipeline for high-throughput plant phenotyping. The proposed technique is capable of learning from different environmental conditions, with a high level of robustness.
Collapse
Affiliation(s)
- Frank Gyan Okyere
- Sustainable Soils and Crops, Rothamsted Research, Harpenden AL5 2JQ, UK
- School of Water, Energy and Environment, Soil, Agrifood and Biosciences, Cranfield University, Bedford MK43 0AL, UK
| | - Daniel Cudjoe
- Sustainable Soils and Crops, Rothamsted Research, Harpenden AL5 2JQ, UK
- School of Water, Energy and Environment, Soil, Agrifood and Biosciences, Cranfield University, Bedford MK43 0AL, UK
| | | | - Nicolas Virlet
- Sustainable Soils and Crops, Rothamsted Research, Harpenden AL5 2JQ, UK
| | - Andrew B. Riche
- Sustainable Soils and Crops, Rothamsted Research, Harpenden AL5 2JQ, UK
| | - March Castle
- Sustainable Soils and Crops, Rothamsted Research, Harpenden AL5 2JQ, UK
| | - Latifa Greche
- Sustainable Soils and Crops, Rothamsted Research, Harpenden AL5 2JQ, UK
| | - Fady Mohareb
- School of Water, Energy and Environment, Soil, Agrifood and Biosciences, Cranfield University, Bedford MK43 0AL, UK
| | - Daniel Simms
- School of Water, Energy and Environment, Soil, Agrifood and Biosciences, Cranfield University, Bedford MK43 0AL, UK
| | - Manal Mhada
- African Integrated Plant and Soil Science, Agro-Biosciences, University of Mohammed VI Polytechnic, Lot 660, Ben Guerir 43150, Morocco
| | | |
Collapse
|
39
|
Matsumoto S, Ishida S, Terayama K, Okuno Y. Quantitative analysis of protein dynamics using a deep learning technique combined with experimental cryo-EM density data and MD simulations. Biophys Physicobiol 2023; 20:e200022. [PMID: 38496243 PMCID: PMC10941960 DOI: 10.2142/biophysico.bppb-v20.0022] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2023] [Accepted: 05/12/2023] [Indexed: 03/19/2024] Open
Abstract
Protein functions associated with biological activity are precisely regulated by both tertiary structure and dynamic behavior. Thus, elucidating the high-resolution structures and quantitative information on in-solution dynamics is essential for understanding the molecular mechanisms. The main experimental approaches for determining tertiary structures include nuclear magnetic resonance (NMR), X-ray crystallography, and cryogenic electron microscopy (cryo-EM). Among these procedures, recent remarkable advances in the hardware and analytical techniques of cryo-EM have increasingly determined novel atomic structures of macromolecules, especially those with large molecular weights and complex assemblies. In addition to these experimental approaches, deep learning techniques, such as AlphaFold 2, accurately predict structures from amino acid sequences, accelerating structural biology research. Meanwhile, the quantitative analyses of the protein dynamics are conducted using experimental approaches, such as NMR and hydrogen-deuterium mass spectrometry, and computational approaches, such as molecular dynamics (MD) simulations. Although these procedures can quantitatively explore dynamic behavior at high resolution, the fundamental difficulties, such as signal crowding and high computational cost, greatly hinder their application to large and complex biological macromolecules. In recent years, machine learning techniques, especially deep learning techniques, have been actively applied to structural data to identify features that are difficult for humans to recognize from big data. Here, we review our approach to accurately estimate dynamic properties associated with local fluctuations from three-dimensional cryo-EM density data using a deep learning technique combined with MD simulations.
Collapse
Affiliation(s)
| | - Shoichi Ishida
- Graduate School of Medical Life Science, Yokohama City University, Yokohama, Kanagawa 230-0045, Japan
| | - Kei Terayama
- Graduate School of Medical Life Science, Yokohama City University, Yokohama, Kanagawa 230-0045, Japan
- RIKEN Center for Computational Science, Kobe, Hyogo 650-0047, Japan
| | - Yasuhshi Okuno
- Graduate School of Medicine, Kyoto University, Kyoto 606-8507, Japan
- RIKEN Center for Computational Science, Kobe, Hyogo 650-0047, Japan
| |
Collapse
|
40
|
De A, Wang X, Zhang Q, Wu J, Cong F. An efficient memory reserving-and-fading strategy for vector quantization based 3D brain segmentation and tumor extraction using an unsupervised deep learning network. Cogn Neurodyn 2023; 18:1-22. [PMID: 37362765 PMCID: PMC10132803 DOI: 10.1007/s11571-023-09965-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2022] [Revised: 02/24/2023] [Accepted: 03/08/2023] [Indexed: 06/28/2023] Open
Abstract
Deep learning networks are state-of-the-art approaches for 3D brain image segmentation, and the radiological characteristics extracted from tumors are of great significance for clinical diagnosis, treatment planning, and treatment outcome evaluation. However, two problems have been the hindering factors in brain image segmentation techniques. One is that deep learning networks require large amounts of manually annotated data. Another issue is the computational efficiency of 3D deep learning networks. In this study, we propose a vector quantization (VQ)-based 3D segmentation method that employs a novel unsupervised 3D deep embedding clustering (3D-DEC) network and an efficiency memory reserving-and-fading strategy. The VQ-based 3D-DEC network is trained on volume data in an unsupervised manner to avoid manual data annotation. The memory reserving-and-fading strategy beefs up model efficiency greatly. The designed methodology makes deep learning-based model feasible for biomedical image segmentation. The experiment is divided into two parts. First, we extensively evaluate the effectiveness and robustness of the proposed model on two authoritative MRI brain tumor databases (i.e., IBSR and BrainWeb). Second, we validate the model using real 3D brain tumor data collected from our institute for clinical practice significance. Results show that our method (without data manual annotation) has superior accuracy (0.74 ± 0.04 Tanimoto coefficient on IBSR, 97.5% TP and 97.7% TN on BrainWeb, and 91% Dice, 88% sensitivity and 87% specificity on real brain data) and remarkable efficiency (speedup ratio is 18.72 on IBSR, 31.16 on BrainWeb, 31.00 on real brain data) compared to the state-of-the-art methods. The results show that our proposed model can address the lacks of manual annotations, and greatly increase computation speedup with competitive segmentation accuracy compared to other state-of-the-art 3D CNN models. Moreover, the proposed model can be used for tumor treatment follow-ups every 6 months, providing critical details for surgical and postoperative treatment by correctly extracting numerical radiomic features of tumors.
Collapse
Affiliation(s)
- Ailing De
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
| | - Xiulin Wang
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
- School of Biomedical Engineering, Faculty of Medicine, Dalian University of Technology, Dalian, 116000 Liaoning China
| | - Qing Zhang
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
| | - Jianlin Wu
- Department of Radiology, Affiliated Zhongshan Hospital of Dalian University, Dalian, 116000 Liaoning China
| | - Fengyu Cong
- School of Biomedical Engineering, Faculty of Medicine, Dalian University of Technology, Dalian, 116000 Liaoning China
- Faculty of Information Technology, University of Jyväskylä, 40014 Jyväskylä, Finland
- School of Artificial Intelligence, Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, 116000 Liaoning China
- Key Laboratory of Integrated Circuit and Biomedical Electronic System, Dalian University of Technology, Dalian, 116000 Liaoning China
| |
Collapse
|
41
|
Bass C, Silva MD, Sudre C, Williams LZJ, Sousa HS, Tudosiu PD, Alfaro-Almagro F, Fitzgibbon SP, Glasser MF, Smith SM, Robinson EC. ICAM-Reg: Interpretable Classification and Regression With Feature Attribution for Mapping Neurological Phenotypes in Individual Scans. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:959-970. [PMID: 36374873 DOI: 10.1109/tmi.2022.3221890] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
An important goal of medical imaging is to be able to precisely detect patterns of disease specific to individual scans; however, this is challenged in brain imaging by the degree of heterogeneity of shape and appearance. Traditional methods, based on image registration, historically fail to detect variable features of disease, as they utilise population-based analyses, suited primarily to studying group-average effects. In this paper we therefore take advantage of recent developments in generative deep learning to develop a method for simultaneous classification, or regression, and feature attribution (FA). Specifically, we explore the use of a VAE-GAN (variational autoencoder - general adversarial network) for translation called ICAM, to explicitly disentangle class relevant features, from background confounds, for improved interpretability and regression of neurological phenotypes. We validate our method on the tasks of Mini-Mental State Examination (MMSE) cognitive test score prediction for the Alzheimer's Disease Neuroimaging Initiative (ADNI) cohort, as well as brain age prediction, for both neurodevelopment and neurodegeneration, using the developing Human Connectome Project (dHCP) and UK Biobank datasets. We show that the generated FA maps can be used to explain outlier predictions and demonstrate that the inclusion of a regression module improves the disentanglement of the latent space. Our code is freely available on GitHub https://github.com/CherBass/ICAM.
Collapse
|
42
|
Feature generation and multi-sequence fusion based deep convolutional network for breast tumor diagnosis with missing MR sequences. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104536] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/28/2022]
|
43
|
Wang C, Cui Z, Yang J, Han M, Carneiro G, Shen D. BowelNet: Joint Semantic-Geometric Ensemble Learning for Bowel Segmentation From Both Partially and Fully Labeled CT Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:1225-1236. [PMID: 36449590 DOI: 10.1109/tmi.2022.3225667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
Accurate bowel segmentation is essential for diagnosis and treatment of bowel cancers. Unfortunately, segmenting the entire bowel in CT images is quite challenging due to unclear boundary, large shape, size, and appearance variations, as well as diverse filling status within the bowel. In this paper, we present a novel two-stage framework, named BowelNet, to handle the challenging task of bowel segmentation in CT images, with two stages of 1) jointly localizing all types of the bowel, and 2) finely segmenting each type of the bowel. Specifically, in the first stage, we learn a unified localization network from both partially- and fully-labeled CT images to robustly detect all types of the bowel. To better capture unclear bowel boundary and learn complex bowel shapes, in the second stage, we propose to jointly learn semantic information (i.e., bowel segmentation mask) and geometric representations (i.e., bowel boundary and bowel skeleton) for fine bowel segmentation in a multi-task learning scheme. Moreover, we further propose to learn a meta segmentation network via pseudo labels to improve segmentation accuracy. By evaluating on a large abdominal CT dataset, our proposed BowelNet method can achieve Dice scores of 0.764, 0.848, 0.835, 0.774, and 0.824 in segmenting the duodenum, jejunum-ileum, colon, sigmoid, and rectum, respectively. These results demonstrate the effectiveness of our proposed BowelNet framework in segmenting the entire bowel from CT images.
Collapse
|
44
|
Wen M, Zhou Q, Tao B, Shcherbakov P, Xu Y, Zhang X. Short‐term and long‐term memory self‐attention network for segmentation of tumours in 3D medical images. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY 2023. [DOI: 10.1049/cit2.12179] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/19/2023] Open
Affiliation(s)
- Mingwei Wen
- Department of Biomedical Engineering College of Life Science and Technology Huazhong University of Science and Technology Wuhan China
| | - Quan Zhou
- Department of Biomedical Engineering College of Life Science and Technology Huazhong University of Science and Technology Wuhan China
| | - Bo Tao
- State Key Laboratory of Digital Manufacturing Equipment and Technology Huazhong University of Science and Technology Wuhan China
| | - Pavel Shcherbakov
- Institute for Control Science Russian Academy of Sciences Moscow Russia
| | - Yang Xu
- Hubei Medical Devices Quality Supervision and Test Institute Wuhan China
| | - Xuming Zhang
- Department of Biomedical Engineering College of Life Science and Technology Huazhong University of Science and Technology Wuhan China
| |
Collapse
|
45
|
Wu R, Xin Y, Qian J, Dong Y. A multi-scale interactive U-Net for pulmonary vessel segmentation method based on transfer learning. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104407] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
46
|
Zhang B, Wang Y, Ding C, Deng Z, Li L, Qin Z, Ding Z, Bian L, Yang C. Multi-scale feature pyramid fusion network for medical image segmentation. Int J Comput Assist Radiol Surg 2023; 18:353-365. [PMID: 36042149 DOI: 10.1007/s11548-022-02738-5] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2022] [Accepted: 08/11/2022] [Indexed: 02/03/2023]
Abstract
PURPOSE Medical image segmentation is the most widely used technique in diagnostic and clinical research. However, accurate segmentation of target organs from blurred border regions and low-contrast adjacent organs in Computed tomography (CT) imaging is crucial for clinical diagnosis and treatment. METHODS In this article, we propose a Multi-Scale Feature Pyramid Fusion Network (MS-Net) based on the codec structure formed by the combination of Multi-Scale Attention Module (MSAM) and Stacked Feature Pyramid Module (SFPM). Among them, MSAM is used to skip connections, which aims to extract different levels of context details by dynamically adjusting the receptive fields under different network depths; the SFPM including multi-scale strategies and multi-layer Feature Perception Module (FPM) is nested in the network at the deepest point, which aims to better focus the network's attention on the target organ by adaptively increasing the weight of the features of interest. RESULTS Experiments demonstrate that the proposed MS-Net significantly improved the Dice score from 91.74% to 94.54% on CHAOS, from 97.59% to 98.59% on Lung, and from 82.55% to 86.06% on ISIC 2018, compared with U-Net. Additionally, comparisons with other six state-of-the-art codec structures also show the presented network has great advantages on evaluation indicators such as Miou, Dice, ACC and AUC. CONCLUSION The experimental results show that both the MSAM and SFPM techniques proposed in this paper can assist the network to improve the segmentation effect, so that the proposed MS-Net method achieves better results in the CHAOS, Lung and ISIC 2018 segmentation tasks.
Collapse
Affiliation(s)
- Bing Zhang
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Yang Wang
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Caifu Ding
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Ziqing Deng
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Linwei Li
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Zesheng Qin
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Zhao Ding
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China
| | - Lifeng Bian
- Frontier Institute of Chip and System, Fudan University, Shanghai, 200433, China.
| | - Chen Yang
- Power Systems Engineering Research Center, Ministry of Education, College of Big Data and Information Engineering, Guizhou University, Guiyang, 550025, China.
| |
Collapse
|
47
|
Qureshi I, Yan J, Abbas Q, Shaheed K, Riaz AB, Wahid A, Khan MWJ, Szczuko P. Medical image segmentation using deep semantic-based methods: A review of techniques, applications and emerging trends. INFORMATION FUSION 2023. [DOI: 10.1016/j.inffus.2022.09.031] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
|
48
|
Chen Z, Zheng W, Pang K, Xia D, Guo L, Chen X, Wu F, Wang H. Weakly supervised learning analysis of Aβ plaque distribution in the whole rat brain. Front Neurosci 2023; 16:1097019. [PMID: 36741048 PMCID: PMC9892753 DOI: 10.3389/fnins.2022.1097019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2022] [Accepted: 12/30/2022] [Indexed: 01/20/2023] Open
Abstract
Alzheimer's disease (AD) is a great challenge for the world and hardly to be cured, partly because of the lack of animal models that fully mimic pathological progress. Recently, a rat model exhibiting the most pathological symptoms of AD has been reported. However, high-resolution imaging and accurate quantification of beta-amyloid (Aβ) plaques in the whole rat brain have not been fulfilled due to substantial technical challenges. In this paper, a high-efficiency data analysis pipeline is proposed to quantify Aβ plaques in whole rat brain through several terabytes of image data acquired by a high-speed volumetric imaging approach we have developed previously. A novel segmentation framework applying a high-performance weakly supervised learning method which can dramatically reduce the human labeling consumption is described in this study. The effectiveness of our segmentation framework is validated with different metrics. The segmented Aβ plaques were mapped to a standard rat brain atlas for quantitative analysis of the Aβ distribution in each brain area. This pipeline may also be applied to the segmentation and accurate quantification of other non-specific morphology objects.
Collapse
Affiliation(s)
- Zhiyi Chen
- National Engineering Laboratory for Brain-Inspired Intelligence Technology and Application, School of Information Science and Technology, University of Science and Technology of China, Hefei, China,Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China
| | - Weijie Zheng
- Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China,AHU-IAI AI Joint Laboratory, Anhui University, Hefei, China
| | - Keliang Pang
- School of Pharmaceutical Sciences, IDG/McGovern Institute for Brain Research, Tsinghua University-Peking University Joint Center for Life Sciences, Tsinghua University, Beijing, China,*Correspondence: Keliang Pang,
| | - Debin Xia
- National Engineering Laboratory for Brain-Inspired Intelligence Technology and Application, School of Information Science and Technology, University of Science and Technology of China, Hefei, China,Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China
| | - Lingxiao Guo
- Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China
| | - Xuejin Chen
- National Engineering Laboratory for Brain-Inspired Intelligence Technology and Application, School of Information Science and Technology, University of Science and Technology of China, Hefei, China,Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China
| | - Feng Wu
- National Engineering Laboratory for Brain-Inspired Intelligence Technology and Application, School of Information Science and Technology, University of Science and Technology of China, Hefei, China,Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China
| | - Hao Wang
- National Engineering Laboratory for Brain-Inspired Intelligence Technology and Application, School of Information Science and Technology, University of Science and Technology of China, Hefei, China,Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei, China,Hao Wang,
| |
Collapse
|
49
|
Cabeza-Ruiz R, Velázquez-Pérez L, Pérez-Rodríguez R, Reetz K. ConvNets for automatic detection of polyglutamine SCAs from brain MRIs: state of the art applications. Med Biol Eng Comput 2023; 61:1-24. [PMID: 36385616 DOI: 10.1007/s11517-022-02714-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2021] [Accepted: 10/26/2022] [Indexed: 11/17/2022]
Abstract
Polyglutamine spinocerebellar ataxias (polyQ SCAs) are a group of neurodegenerative diseases, clinically and genetically heterogeneous, characterized by loss of balance and motor coordination due to dysfunction of the cerebellum and its connections. The diagnosis of each type of polyQ SCA, alongside with genetic tests, includes medical images analysis, and its automation may help specialists to distinguish between each type. Convolutional neural networks (ConvNets or CNNs) have been recently used for medical image processing, with outstanding results. In this work, we present the main clinical and imaging features of polyglutamine SCAs, and the basics of CNNs. Finally, we review studies that have used this approach to automatically process brain medical images and may be applied to SCAs detection. We conclude by discussing the possible limitations and opportunities of using ConvNets for SCAs diagnose in the future.
Collapse
Affiliation(s)
| | - Luis Velázquez-Pérez
- Cuban Academy of Sciences, La Habana, Cuba
- Center for the Research and Rehabilitation of Hereditary Ataxias, Holguín, Cuba
| | - Roberto Pérez-Rodríguez
- CAD/CAM Study Center, University of Holguín, Holguín, Cuba
- Cuban Academy of Sciences, La Habana, Cuba
| | - Kathrin Reetz
- Department of Neurology, RWTH Aachen University, Aachen, Germany
| |
Collapse
|
50
|
Multi-target segmentation of pancreas and pancreatic tumor based on fusion of attention mechanism. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|