1
|
Kim G, Viswanathan AN, Bhatia R, Landman Y, Roumeliotis M, Erickson B, Schmidt EJ, Lee J. Dual convolution-transformer UNet (DCT-UNet) for organs at risk and clinical target volume segmentation in MRI for cervical cancer brachytherapy. Phys Med Biol 2024; 69:215014. [PMID: 39378904 DOI: 10.1088/1361-6560/ad84b2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2024] [Accepted: 10/08/2024] [Indexed: 10/10/2024]
Abstract
Objective. MRI is the standard imaging modality for high-dose-rate brachytherapy of cervical cancer. Precise contouring of organs at risk (OARs) and high-risk clinical target volume (HR-CTV) from MRI is a crucial step for radiotherapy planning and treatment. However, conventional manual contouring has limitations in terms of accuracy as well as procedural time. To overcome these, we propose a deep learning approach to automatically segment OARs (bladder, rectum, and sigmoid colon) and HR-CTV from female pelvic MRI.Approach. In the proposed pipeline, a coarse multi-organ segmentation model first segments all structures, from which a region of interest is computed for each structure. Then, each organ is segmented using an organ-specific fine segmentation model separately trained for each organ. To account for variable sizes of HR-CTV, a size-adaptive multi-model approach was employed. For coarse and fine segmentations, we designed a dual convolution-transformer UNet (DCT-UNet) which uses dual-path encoder consisting of convolution and transformer blocks. To evaluate our model, OAR segmentations were compared to the clinical contours drawn by the attending radiation oncologist. For HR-CTV, four sets of contours (clinical + three additional sets) were obtained to produce a consensus ground truth as well as for inter/intra-observer variability analysis.Main results. DCT-UNet achieved dice similarity coefficient (mean ± SD) of 0.932 ± 0.032 (bladder), 0.786 ± 0.090 (rectum), 0.663 ± 0.180 (sigmoid colon), and 0.741 ± 0.076 (HR-CTV), outperforming other state-of-the-art models. Notably, the size-adaptive multi-model significantly improved HR-CTV segmentation compared to a single-model. Furthermore, significant inter/intra-observer variability was observed, and our model showed comparable performance to all observers. Computation time for the entire pipeline per subject was 12.59 ± 0.79 s, which is significantly shorter than the typical manual contouring time of >15 min.Significance. These experimental results demonstrate that our model has great utility in cervical cancer brachytherapy by enabling fast and accurate automatic segmentation, and has potential in improving consistency in contouring. DCT-UNet source code is available athttps://github.com/JHU-MICA/DCT-UNet.
Collapse
Affiliation(s)
- Gayoung Kim
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| | - Akila N Viswanathan
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| | - Rohini Bhatia
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| | - Yosef Landman
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| | - Michael Roumeliotis
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| | - Beth Erickson
- Department of Radiation Oncology, Medical College of Wisconsin, Milwaukee, WI, United States of America
| | - Ehud J Schmidt
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
- Department of Cardiology, Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| | - Junghoon Lee
- Department of Radiation Oncology and Molecular Radiation Sciences, Johns Hopkins University, Baltimore, MD, United States of America
| |
Collapse
|
2
|
Al-Tahhan FE, Omar E. Fusion of Color Correction and HSV Segmentation Techniques for Automated Segmentation of Acute Lymphoblastic Leukemia. Microsc Res Tech 2024. [PMID: 39377153 DOI: 10.1002/jemt.24706] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2024] [Revised: 09/03/2024] [Accepted: 09/20/2024] [Indexed: 10/09/2024]
Abstract
This article presents an enhanced segmentation methodology for the accurate detection of acute lymphoblastic leukemia (ALL) in blood smear images. The proposed approach integrates color correction techniques with HSV color space segmentation to improve white blood cell analysis. Our method addresses common challenges in microscopic image processing, including sensor nonlinearity, uneven illumination, and color distortions. The key objectives of this study are to develop a robust preprocessing pipeline that normalizes blood smear images for consistent analysis, implement an HSV-based segmentation technique optimized for leukocyte detection, and validate the method's effectiveness across various ALL subtypes using clinical samples. The proposed technique was evaluated using real-world blood smear samples from ALL patients. Quantitative analysis demonstrates significant improvements in segmentation accuracy compared to traditional methods. Our approach shows strong capability in reliably detecting and segmenting ALL subtypes, offering the potential for enhanced diagnostic support in clinical settings.
Collapse
Affiliation(s)
- F E Al-Tahhan
- Department of Mathematics, Faculty of Science, Mansoura University, Mansoura, Egypt
| | - Emam Omar
- Department of Physics, Faculty of Science, Mansoura University, Mansoura, Egypt
| |
Collapse
|
3
|
Xue X, Sun L, Liang D, Zhu J, Liu L, Sun Q, Liu H, Gao J, Fu X, Ding J, Dai X, Tao L, Cheng J, Li T, Zhou F. Deep learning-based segmentation for high-dose-rate brachytherapy in cervical cancer using 3D Prompt-ResUNet. Phys Med Biol 2024; 69:195008. [PMID: 39270708 DOI: 10.1088/1361-6560/ad7ad1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 09/13/2024] [Indexed: 09/15/2024]
Abstract
Objective.To develop and evaluate a 3D Prompt-ResUNet module that utilized the prompt-based model combined with 3D nnUNet for rapid and consistent autosegmentation of high-risk clinical target volume (HRCTV) and organ at risk (OAR) in high-dose-rate brachytherapy for cervical cancer patients.Approach.We used 73 computed tomography scans and 62 magnetic resonance imaging scans from 135 (103 for training, 16 for validation, and 16 for testing) cervical cancer patients across two hospitals for HRCTV and OAR segmentation. A novel comparison of the deep learning neural networks 3D Prompt-ResUNet, nnUNet, and segment anything model-Med3D was applied for the segmentation. Evaluation was conducted in two parts: geometric and clinical assessments. Quantitative metrics included the Dice similarity coefficient (DSC), 95th percentile Hausdorff distance (HD95%), Jaccard index (JI), and Matthews correlation coefficient (MCC). Clinical evaluation involved interobserver comparison, 4-grade expert scoring, and a double-blinded Turing test.Main results.The Prompt-ResUNet model performed most similarly to experienced radiation oncologists, outperforming less experienced ones. During testing, the DSC, HD95% (mm), JI, and MCC value (mean ± SD) for HRCTV were 0.92 ± 0.03, 2.91 ± 0.69, 0.85 ± 0.04, and 0.92 ± 0.02, respectively. For the bladder, these values were 0.93 ± 0.05, 3.07 ± 1.05, 0.87 ± 0.08, and 0.93 ± 0.05, respectively. For the rectum, they were 0.87 ± 0.03, 3.54 ± 1.46, 0.78 ± 0.05, and 0.87 ± 0.03, respectively. For the sigmoid, they were 0.76 ± 0.11, 7.54 ± 5.54, 0.63 ± 0.14, and 0.78 ± 0.09, respectively. The Prompt-ResUNet achieved a clinical viability score of at least 2 in all evaluation cases (100%) for both HRCTV and bladder and exceeded the 30% positive rate benchmark for all evaluated structures in the Turing test.Significance.The Prompt-ResUNet architecture demonstrated high consistency with ground truth in autosegmentation of HRCTV and OARs, reducing interobserver variability and shortening treatment times.
Collapse
Affiliation(s)
- Xian Xue
- Key Laboratory of Radiological Protection and Nuclear Emergency, National Institute for Radiological Protection, Chinese Center for Disease Control and Prevention (CDC), Beijing 100088, People's Republic of China
| | - Lining Sun
- Department of radiation oncology, Fudan University Shanghai Cancer Center, Shanghai 200032, People's Republic of China
| | - Dazhu Liang
- Digital Health China Technologies Co., LTD, Beijing 100089, People's Republic of China
| | - Jingyang Zhu
- Department of radiation oncology, Zhongcheng Cancer center, Beijing 100160, People's Republic of China
| | - Lele Liu
- Department of radiation oncology, The First Affiliated Hospital of Zhengzhou University, Zhengzhou 450052, People's Republic of China
| | - Quanfu Sun
- Key Laboratory of Radiological Protection and Nuclear Emergency, National Institute for Radiological Protection, Chinese Center for Disease Control and Prevention (CDC), Beijing 100088, People's Republic of China
| | - Hefeng Liu
- Digital Health China Technologies Co., LTD, Beijing 100089, People's Republic of China
| | - Jianwei Gao
- Digital Health China Technologies Co., LTD, Beijing 100089, People's Republic of China
| | - Xiaosha Fu
- Biomedical Research Centre, Sheffield Hallam University, Sheffield S11WB, United Kingdom
| | - Jingjing Ding
- Department of radiation oncology, Chinese People's Liberation Army (PLA) General Hospital, Beijing 100853, People's Republic of China
| | - Xiangkun Dai
- Department of radiation oncology, Chinese People's Liberation Army (PLA) General Hospital, Beijing 100853, People's Republic of China
| | - Laiyuan Tao
- Digital Health China Technologies Co., LTD, Beijing 100089, People's Republic of China
| | - Jinsheng Cheng
- Key Laboratory of Radiological Protection and Nuclear Emergency, National Institute for Radiological Protection, Chinese Center for Disease Control and Prevention (CDC), Beijing 100088, People's Republic of China
| | - Tengxiang Li
- Department of Nuclear Science and Engineering, Nanhua University, Hunan 421001, People's Republic of China
| | - Fugen Zhou
- Department of Aero-space Information Engineering, Beihang University, Beijing 100191, People's Republic of China
| |
Collapse
|
4
|
Chen J, Qiu RL, Wang T, Momin S, Yang X. A Review of Artificial Intelligence in Brachytherapy. ARXIV 2024:arXiv:2409.16543v1. [PMID: 39398213 PMCID: PMC11469420] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Subscribe] [Scholar Register] [Indexed: 10/15/2024]
Abstract
Artificial intelligence (AI) has the potential to revolutionize brachytherapy's clinical workflow. This review comprehensively examines the application of AI, focusing on machine learning and deep learning, in facilitating various aspects of brachytherapy. We analyze AI's role in making brachytherapy treatments more personalized, efficient, and effective. The applications are systematically categorized into seven categories: imaging, preplanning, treatment planning, applicator reconstruction, quality assurance, outcome prediction, and real-time monitoring. Each major category is further subdivided based on cancer type or specific tasks, with detailed summaries of models, data sizes, and results presented in corresponding tables. This review offers insights into the current advancements, challenges, and the impact of AI on treatment paradigms, encouraging further research to expand its clinical utility.
Collapse
Affiliation(s)
- Jingchu Chen
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308
- School of Mechanical Engineering, Georgia Institute of Technology, GA, Atlanta, USA
| | - Richard L.J. Qiu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308
| | - Tonghe Wang
- Department of Medical Physics, Memorial Sloan Kettering Cancer Center, New York, NY 10065
| | - Shadab Momin
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30308
| |
Collapse
|
5
|
Ni R, Han K, Haibe-Kains B, Rink A. Generalizability of deep learning in organ-at-risk segmentation: A transfer learning study in cervical brachytherapy. Radiother Oncol 2024; 197:110332. [PMID: 38763356 DOI: 10.1016/j.radonc.2024.110332] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2024] [Revised: 05/02/2024] [Accepted: 05/03/2024] [Indexed: 05/21/2024]
Abstract
PURPOSE Deep learning can automate delineation in radiation therapy, reducing time and variability. Yet, its efficacy varies across different institutions, scanners, or settings, emphasizing the need for adaptable and robust models in clinical environments. Our study demonstrates the effectiveness of the transfer learning (TL) approach in enhancing the generalizability of deep learning models for auto-segmentation of organs-at-risk (OARs) in cervical brachytherapy. METHODS A pre-trained model was developed using 120 scans with ring and tandem applicator on a 3T magnetic resonance (MR) scanner (RT3). Four OARs were segmented and evaluated. Segmentation performance was evaluated by Volumetric Dice Similarity Coefficient (vDSC), 95 % Hausdorff Distance (HD95), surface DSC, and Added Path Length (APL). The model was fine-tuned on three out-of-distribution target groups. Pre- and post-TL outcomes, and influence of number of fine-tuning scans, were compared. A model trained with one group (Single) and a model trained with all four groups (Mixed) were evaluated on both seen and unseen data distributions. RESULTS TL enhanced segmentation accuracy across target groups, matching the pre-trained model's performance. The first five fine-tuning scans led to the most noticeable improvements, with performance plateauing with more data. TL outperformed training-from-scratch given the same training data. The Mixed model performed similarly to the Single model on RT3 scans but demonstrated superior performance on unseen data. CONCLUSIONS TL can improve a model's generalizability for OAR segmentation in MR-guided cervical brachytherapy, requiring less fine-tuning data and reduced training time. These results provide a foundation for developing adaptable models to accommodate clinical settings.
Collapse
Affiliation(s)
- Ruiyan Ni
- Department of Medical Biophysics, University of Toronto, Toronto, Canada
| | - Kathy Han
- Princess Margaret Cancer Center, University Health Network, Toronto, CA, Canada; Department of Radiation Oncology, University of Toronto, Toronto, CA, Canada
| | - Benjamin Haibe-Kains
- Department of Medical Biophysics, University of Toronto, Toronto, Canada; Princess Margaret Cancer Center, University Health Network, Toronto, CA, Canada; Vector Institute, Toronto, Toronto, CA, Canada.
| | - Alexandra Rink
- Department of Medical Biophysics, University of Toronto, Toronto, Canada; Princess Margaret Cancer Center, University Health Network, Toronto, CA, Canada; Department of Radiation Oncology, University of Toronto, Toronto, CA, Canada.
| |
Collapse
|
6
|
Zhu S, Chen G, Chen H, Lu Y, Wu M, Zheng B, Liu D, Qian C, Chen Y. Squeeze-and-excitation-attention-based mobile vision transformer for grading recognition of bladder prolapse in pelvic MRI images. Med Phys 2024; 51:5236-5249. [PMID: 38767532 DOI: 10.1002/mp.17171] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2023] [Revised: 03/22/2024] [Accepted: 03/31/2024] [Indexed: 05/22/2024] Open
Abstract
BACKGROUND Bladder prolapse is a common clinical disorder of pelvic floor dysfunction in women, and early diagnosis and treatment can help them recover. Pelvic magnetic resonance imaging (MRI) is one of the most important methods used by physicians to diagnose bladder prolapse; however, it is highly subjective and largely dependent on the clinical experience of physicians. The application of computer-aided diagnostic techniques to achieve a graded diagnosis of bladder prolapse can help improve its accuracy and shorten the learning curve. PURPOSE The purpose of this study is to combine convolutional neural network (CNN) and vision transformer (ViT) for grading bladder prolapse in place of traditional neural networks, and to incorporate attention mechanisms into mobile vision transformer (MobileViT) for assisting in the grading of bladder prolapse. METHODS This study focuses on the grading of bladder prolapse in pelvic organs using a combination of a CNN and a ViT. First, this study used MobileNetV2 to extract the local features of the images. Next, a ViT was used to extract the global features by modeling the non-local dependencies at a distance. Finally, a channel attention module (i.e., squeeze-and-excitation network) was used to improve the feature extraction network and enhance its feature representation capability. The final grading of the degree of bladder prolapse was thus achieved. RESULTS Using pelvic MRI images provided by a Huzhou Maternal and Child Health Care Hospital, this study used the proposed method to grade patients with bladder prolapse. The accuracy, Kappa value, sensitivity, specificity, precision, and area under the curve of our method were 86.34%, 78.27%, 83.75%, 95.43%, 85.70%, and 95.05%, respectively. In comparison with other CNN models, the proposed method performed better. CONCLUSIONS Thus, the model based on attention mechanisms exhibits better classification performance than existing methods for grading bladder prolapse in pelvic organs, and it can effectively assist physicians in achieving a more accurate bladder prolapse diagnosis.
Collapse
Affiliation(s)
- Shaojun Zhu
- School of information engineering, Huzhou University, Huzhou, Zhejiang, China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural 8 Resources, Huzhou University, Huzhou, China
| | - Guotao Chen
- School of information engineering, Huzhou University, Huzhou, Zhejiang, China
| | - Hongguang Chen
- School of information engineering, Huzhou University, Huzhou, Zhejiang, China
| | - Ying Lu
- School of information engineering, Huzhou University, Huzhou, Zhejiang, China
| | - Maonian Wu
- School of information engineering, Huzhou University, Huzhou, Zhejiang, China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural 8 Resources, Huzhou University, Huzhou, China
| | - Bo Zheng
- School of information engineering, Huzhou University, Huzhou, Zhejiang, China
- Zhejiang Province Key Laboratory of Smart Management & Application of Modern Agricultural 8 Resources, Huzhou University, Huzhou, China
| | - Dongquan Liu
- Ninghai First Hospital, Ninghai, Zhejiang, China
| | - Cheng Qian
- Department of Colon-rectal Surgery, Huzhou Maternity & Child Health Care Hospital, Huzhou, China
| | - Yun Chen
- Department of Colon-rectal Surgery, Huzhou Maternity & Child Health Care Hospital, Huzhou, China
| |
Collapse
|
7
|
Xue X, Liang D, Wang K, Gao J, Ding J, Zhou F, Xu J, Liu H, Sun Q, Jiang P, Tao L, Shi W, Cheng J. A deep learning-based 3D Prompt-nnUnet model for automatic segmentation in brachytherapy of postoperative endometrial carcinoma. J Appl Clin Med Phys 2024; 25:e14371. [PMID: 38682540 PMCID: PMC11244685 DOI: 10.1002/acm2.14371] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 02/07/2024] [Accepted: 03/25/2024] [Indexed: 05/01/2024] Open
Abstract
PURPOSE To create and evaluate a three-dimensional (3D) Prompt-nnUnet module that utilizes the prompts-based model combined with 3D nnUnet for producing the rapid and consistent autosegmentation of high-risk clinical target volume (HR CTV) and organ at risk (OAR) in high-dose-rate brachytherapy (HDR BT) for patients with postoperative endometrial carcinoma (EC). METHODS AND MATERIALS On two experimental batches, a total of 321 computed tomography (CT) scans were obtained for HR CTV segmentation from 321 patients with EC, and 125 CT scans for OARs segmentation from 125 patients. The numbers of training/validation/test were 257/32/32 and 87/13/25 for HR CTV and OARs respectively. A novel comparison of the deep learning neural network 3D Prompt-nnUnet and 3D nnUnet was applied for HR CTV and OARs segmentation. Three-fold cross validation and several quantitative metrics were employed, including Dice similarity coefficient (DSC), Hausdorff distance (HD), 95th percentile of Hausdorff distance (HD95%), and intersection over union (IoU). RESULTS The Prompt-nnUnet included two forms of parameters Predict-Prompt (PP) and Label-Prompt (LP), with the LP performing most similarly to the experienced radiation oncologist and outperforming the less experienced ones. During the testing phase, the mean DSC values for the LP were 0.96 ± 0.02, 0.91 ± 0.02, and 0.83 ± 0.07 for HR CTV, rectum and urethra, respectively. The mean HD values (mm) were 2.73 ± 0.95, 8.18 ± 4.84, and 2.11 ± 0.50, respectively. The mean HD95% values (mm) were 1.66 ± 1.11, 3.07 ± 0.94, and 1.35 ± 0.55, respectively. The mean IoUs were 0.92 ± 0.04, 0.84 ± 0.03, and 0.71 ± 0.09, respectively. A delineation time < 2.35 s per structure in the new model was observed, which was available to save clinician time. CONCLUSION The Prompt-nnUnet architecture, particularly the LP, was highly consistent with ground truth (GT) in HR CTV or OAR autosegmentation, reducing interobserver variability and shortening treatment time.
Collapse
Affiliation(s)
- Xian Xue
- Secondary Standard Dosimetry LaboratoryNational Institute for Radiological ProtectionChinese Center for Disease Control and Prevention (CDC)BeijingChina
| | - Dazhu Liang
- Digital Health China Technologies Co., LTDBeijingChina
| | - Kaiyue Wang
- Department of RadiotherapyPeking University Third HospitalBeijingChina
| | - Jianwei Gao
- Digital Health China Technologies Co., LTDBeijingChina
| | - Jingjing Ding
- Department of RadiotherapyChinese People's Liberation Army (PLA) General HospitalBeijingChina
| | - Fugen Zhou
- Department of Aero‐space Information EngineeringBeihang UniversityBeijingChina
| | - Juan Xu
- Digital Health China Technologies Co., LTDBeijingChina
| | - Hefeng Liu
- Digital Health China Technologies Co., LTDBeijingChina
| | - Quanfu Sun
- Secondary Standard Dosimetry LaboratoryNational Institute for Radiological ProtectionChinese Center for Disease Control and Prevention (CDC)BeijingChina
| | - Ping Jiang
- Department of RadiotherapyPeking University Third HospitalBeijingChina
| | - Laiyuan Tao
- Digital Health China Technologies Co., LTDBeijingChina
| | - Wenzhao Shi
- Digital Health China Technologies Co., LTDBeijingChina
| | - Jinsheng Cheng
- Secondary Standard Dosimetry LaboratoryNational Institute for Radiological ProtectionChinese Center for Disease Control and Prevention (CDC)BeijingChina
| |
Collapse
|
8
|
Wang X, Feng C, Huang M, Liu S, Ma H, Yu K. Cervical cancer segmentation based on medical images: a literature review. Quant Imaging Med Surg 2024; 14:5176-5204. [PMID: 39022282 PMCID: PMC11250284 DOI: 10.21037/qims-24-369] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Accepted: 05/20/2024] [Indexed: 07/20/2024]
Abstract
Background and Objective Cervical cancer clinical target volume (CTV) outlining and organs at risk segmentation are crucial steps in the diagnosis and treatment of cervical cancer. Manual segmentation is inefficient and subjective, leading to the development of automated or semi-automated methods. However, limitation of image quality, organ motion, and individual differences still pose significant challenges. Apart from numbers of studies on the medical images' segmentation, a comprehensive review within the field is lacking. The purpose of this paper is to comprehensively review the literatures on different types of medical image segmentation regarding cervical cancer and discuss the current level and challenges in segmentation process. Methods As of May 31, 2023, we conducted a comprehensive literature search on Google Scholar, PubMed, and Web of Science using the following term combinations: "cervical cancer images", "segmentation", and "outline". The included studies focused on the segmentation of cervical cancer utilizing computed tomography (CT), magnetic resonance (MR), and positron emission tomography (PET) images, with screening for eligibility by two independent investigators. Key Content and Findings This paper reviews representative papers on CTV and organs at risk segmentation in cervical cancer and classifies the methods into three categories based on image modalities. The traditional or deep learning methods are comprehensively described. The similarities and differences of related methods are analyzed, and their advantages and limitations are discussed in-depth. We have also included experimental results by using our private datasets to verify the performance of selected methods. The results indicate that the residual module and squeeze-and-excitation blocks module can significantly improve the performance of the model. Additionally, the segmentation method based on improved level set demonstrates better segmentation accuracy than other methods. Conclusions The paper provides valuable insights into the current state-of-the-art in cervical cancer CTV outlining and organs at risk segmentation, highlighting areas for future research.
Collapse
Affiliation(s)
- Xiu Wang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Chaolu Feng
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Shenyang, China
- School of Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Mingxu Huang
- School of Computer Science and Engineering, Northeastern University, Shenyang, China
| | - Shiqi Liu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - He Ma
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Shenyang, China
| | - Kun Yu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Shenyang, China
| |
Collapse
|
9
|
Fechter T, Sachpazidis I, Baltas D. The use of deep learning in interventional radiotherapy (brachytherapy): A review with a focus on open source and open data. Z Med Phys 2024; 34:180-196. [PMID: 36376203 PMCID: PMC11156786 DOI: 10.1016/j.zemedi.2022.10.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/13/2022] [Revised: 10/07/2022] [Accepted: 10/10/2022] [Indexed: 11/13/2022]
Abstract
Deep learning advanced to one of the most important technologies in almost all medical fields. Especially in areas, related to medical imaging it plays a big role. However, in interventional radiotherapy (brachytherapy) deep learning is still in an early phase. In this review, first, we investigated and scrutinised the role of deep learning in all processes of interventional radiotherapy and directly related fields. Additionally, we summarised the most recent developments. For better understanding, we provide explanations of key terms and approaches to solving common deep learning problems. To reproduce results of deep learning algorithms both source code and training data must be available. Therefore, a second focus of this work is on the analysis of the availability of open source, open data and open models. In our analysis, we were able to show that deep learning plays already a major role in some areas of interventional radiotherapy, but is still hardly present in others. Nevertheless, its impact is increasing with the years, partly self-propelled but also influenced by closely related fields. Open source, data and models are growing in number but are still scarce and unevenly distributed among different research groups. The reluctance in publishing code, data and models limits reproducibility and restricts evaluation to mono-institutional datasets. The conclusion of our analysis is that deep learning can positively change the workflow of interventional radiotherapy but there is still room for improvements when it comes to reproducible results and standardised evaluation methods.
Collapse
Affiliation(s)
- Tobias Fechter
- Division of Medical Physics, Department of Radiation Oncology, Medical Center University of Freiburg, Germany; Faculty of Medicine, University of Freiburg, Germany; German Cancer Consortium (DKTK), Partner Site Freiburg, Germany.
| | - Ilias Sachpazidis
- Division of Medical Physics, Department of Radiation Oncology, Medical Center University of Freiburg, Germany; Faculty of Medicine, University of Freiburg, Germany; German Cancer Consortium (DKTK), Partner Site Freiburg, Germany
| | - Dimos Baltas
- Division of Medical Physics, Department of Radiation Oncology, Medical Center University of Freiburg, Germany; Faculty of Medicine, University of Freiburg, Germany; German Cancer Consortium (DKTK), Partner Site Freiburg, Germany
| |
Collapse
|
10
|
Fransson S. Comparing multi-image and image augmentation strategies for deep learning-based prostate segmentation. Phys Imaging Radiat Oncol 2024; 29:100551. [PMID: 38444888 PMCID: PMC10912785 DOI: 10.1016/j.phro.2024.100551] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Revised: 01/29/2024] [Accepted: 02/06/2024] [Indexed: 03/07/2024] Open
Abstract
During MR-Linac-based adaptive radiotherapy, multiple images are acquired per patient. These can be applied in training deep learning networks to reduce annotation efforts. This study examined the advantage of using multiple versus single images for prostate treatment segmentation. Findings indicate minimal improvement in DICE and Hausdorff 95% metrics with multiple images. Maximum difference was seen for the rectum in the low data regime, training with images from five patients. Utilizing a 2D U-net resulted in DICE values of 0.80/0.83 when including 1/5 images per patient, respectively. Including more patients in training reduced the difference. Standard augmentation methods remained more effective.
Collapse
Affiliation(s)
- Samuel Fransson
- Department of Medical Physics, Uppsala University Hospital, Uppsala, Sweden
- Department of Surgical Sciences, Uppsala University, Uppsala, Sweden
| |
Collapse
|
11
|
Zhao JZ, Ni R, Chow R, Rink A, Weersink R, Croke J, Raman S. Artificial intelligence applications in brachytherapy: A literature review. Brachytherapy 2023; 22:429-445. [PMID: 37248158 DOI: 10.1016/j.brachy.2023.04.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Revised: 04/02/2023] [Accepted: 04/07/2023] [Indexed: 05/31/2023]
Abstract
PURPOSE Artificial intelligence (AI) has the potential to simplify and optimize various steps of the brachytherapy workflow, and this literature review aims to provide an overview of the work done in this field. METHODS AND MATERIALS We conducted a literature search in June 2022 on PubMed, Embase, and Cochrane for papers that proposed AI applications in brachytherapy. RESULTS A total of 80 papers satisfied inclusion/exclusion criteria. These papers were categorized as follows: segmentation (24), registration and image processing (6), preplanning (13), dose prediction and treatment planning (11), applicator/catheter/needle reconstruction (16), and quality assurance (10). AI techniques ranged from classical models such as support vector machines and decision tree-based learning to newer techniques such as U-Net and deep reinforcement learning, and were applied to facilitate small steps of a process (e.g., optimizing applicator selection) or even automate the entire step of the workflow (e.g., end-to-end preplanning). Many of these algorithms demonstrated human-level performance and offer significant improvements in speed. CONCLUSIONS AI has potential to augment, automate, and/or accelerate many steps of the brachytherapy workflow. We recommend that future studies adhere to standard reporting guidelines. We also stress the importance of using larger sample sizes and reporting results using clinically interpretable measures.
Collapse
Affiliation(s)
- Jonathan Zl Zhao
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Temerty Faculty of Medicine, University of Toronto, Toronto, Canada
| | - Ruiyan Ni
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Department of Medical Biophysics, University of Toronto, Toronto, Canada
| | - Ronald Chow
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Temerty Faculty of Medicine, University of Toronto, Toronto, Canada; Institute of Biomedical Engineering, University of Toronto, Toronto, Canada
| | - Alexandra Rink
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Department of Radiation Oncology, University of Toronto, Toronto, Canada; Department of Medical Biophysics, University of Toronto, Toronto, Canada
| | - Robert Weersink
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Department of Radiation Oncology, University of Toronto, Toronto, Canada; Department of Medical Biophysics, University of Toronto, Toronto, Canada; Institute of Biomedical Engineering, University of Toronto, Toronto, Canada
| | - Jennifer Croke
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Department of Radiation Oncology, University of Toronto, Toronto, Canada
| | - Srinivas Raman
- Princess Margaret Hospital Cancer Centre, Radiation Medicine Program, Toronto, Canada; Department of Radiation Oncology, University of Toronto, Toronto, Canada.
| |
Collapse
|
12
|
Kallis K, Moore LC, Cortes KG, Brown D, Mayadev J, Moore KL, Meyers SM. Automated treatment planning framework for brachytherapy of cervical cancer using 3D dose predictions. Phys Med Biol 2023; 68:10.1088/1361-6560/acc37c. [PMID: 36898161 PMCID: PMC10101723 DOI: 10.1088/1361-6560/acc37c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Accepted: 03/10/2023] [Indexed: 03/12/2023]
Abstract
Objective. To lay the foundation for automated knowledge-based brachytherapy treatment planning using 3D dose estimations, we describe an optimization framework to convert brachytherapy dose distributions directly into dwell times (DTs).Approach. A dose rate kernelḋ(r,θ,φ)was produced by exporting 3D dose for one dwell position from the treatment planning system and normalizing by DT. By translating and rotating this kernel to each dwell position, scaling by DT and summing over all dwell positions, dose was computed (Dcalc). We used a Python-coded COBYLA optimizer to iteratively determine the DTs that minimize the mean squared error betweenDcalcand reference doseDref, computed using voxels withDref80%-120% of prescription. As validation of the optimization, we showed that the optimizer replicates clinical plans whenDref= clinical dose in 40 patients treated with tandem-and-ovoid (T&O) or tandem-and-ring (T&R) and 0-3 needles. Then we demonstrated automated planning in 10 T&O usingDref= dose predicted from a convolutional neural network developed in past work. Validation and automated plans were compared to clinical plans using mean absolute differences (MAD=1N∑n=1Nabsxn-xn') over all voxels (xn= Dose,N= #voxels) and DTs (xn= DT,N= #dwell positions), mean differences (MD) in organD2ccand high-risk CTV D90 over all patients (where positive indicates higher clinical dose), and mean Dice similarity coefficients (DSC) for 100% isodose contours.Main results. Validation plans agreed well with clinical plans (MADdose= 1.1%, MADDT= 4 s or 0.8% of total plan time,D2ccMD = -0.2% to 0.2% and D90 MD = -0.6%, DSC = 0.99). For automated plans, MADdose= 6.5% and MADDT= 10.3 s (2.1%). The slightly higher clinical metrics in automated plans (D2ccMD = -3.8% to 1.3% and D90 MD = -5.1%) were due to higher neural network dose predictions. The overall shape of the automated dose distributions were similar to clinical doses (DSC = 0.91).Significance. Automated planning with 3D dose predictions could provide significant time savings and standardize treatment planning across practitioners, regardless of experience.
Collapse
Affiliation(s)
- Karoline Kallis
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| | - Lance C Moore
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| | - Katherina G Cortes
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| | - Derek Brown
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| | - Jyoti Mayadev
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| | - Kevin L Moore
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| | - Sandra M Meyers
- Department of Radiation Medicine and Applied Sciences, University of California, San Diego, La Jolla, CA, United States of America
| |
Collapse
|
13
|
Eidex Z, Ding Y, Wang J, Abouei E, Qiu RL, Liu T, Wang T, Yang X. Deep Learning in MRI-guided Radiation Therapy: A Systematic Review. ARXIV 2023:arXiv:2303.11378v2. [PMID: 36994167 PMCID: PMC10055493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Subscribe] [Scholar Register] [Indexed: 03/31/2023]
Abstract
MRI-guided radiation therapy (MRgRT) offers a precise and adaptive approach to treatment planning. Deep learning applications which augment the capabilities of MRgRT are systematically reviewed. MRI-guided radiation therapy offers a precise, adaptive approach to treatment planning. Deep learning applications which augment the capabilities of MRgRT are systematically reviewed with emphasis placed on underlying methods. Studies are further categorized into the areas of segmentation, synthesis, radiomics, and real time MRI. Finally, clinical implications, current challenges, and future directions are discussed.
Collapse
Affiliation(s)
- Zach Eidex
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA
- School of Mechanical Engineering, Georgia Institute of Technology, Atlanta, GA
| | - Yifu Ding
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA
| | - Jing Wang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA
| | - Elham Abouei
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA
| | - Richard L.J. Qiu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA
| | - Tian Liu
- Department of Radiation Oncology, Icahn School of Medicine at Mount Sinai, New York, NY
| | - Tonghe Wang
- Department of Medical Physics, Memorial Sloan Kettering Cancer Center, New York, NY
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA
- School of Mechanical Engineering, Georgia Institute of Technology, Atlanta, GA
| |
Collapse
|
14
|
Li Y, Lin C, Zhang Y, Feng S, Huang M, Bai Z. Automatic segmentation of prostate MRI based on 3D pyramid pooling Unet. Med Phys 2023; 50:906-921. [PMID: 35923153 DOI: 10.1002/mp.15895] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2021] [Revised: 06/23/2022] [Accepted: 06/25/2022] [Indexed: 01/01/2023] Open
Abstract
PURPOSE Automatic segmentation of prostate magnetic resonance (MR) images is crucial for the diagnosis, evaluation, and prognosis of prostate diseases (including prostate cancer). In recent years, the mainstream segmentation method for the prostate has been converted to convolutional neural networks. However, owing to the complexity of the tissue structure in MR images and the limitations of existing methods in spatial context modeling, the segmentation performance should be improved further. METHODS In this study, we proposed a novel 3D pyramid pool Unet that benefits from the pyramid pooling structure embedded in the skip connection (SC) and the deep supervision (DS) in the up-sampling of the 3D Unet. The parallel SC of the conventional 3D Unet network causes low-resolution information to be sent to the feature map repeatedly, resulting in blurred image features. To overcome the shortcomings of the conventional 3D Unet, we merge each decoder layer with the feature map of the same scale as the encoder and the smaller scale feature map of the pyramid pooling encoder. This SC combines the low-level details and high-level semantics at two different levels of feature maps. In addition, pyramid pooling performs multifaceted feature extraction on each image behind the convolutional layer, and DS learns hierarchical representations from comprehensive aggregated feature maps, which can improve the accuracy of the task. RESULTS Experiments on 3D prostate MR images of 78 patients demonstrated that our results were highly correlated with expert manual segmentation. The average relative volume difference and Dice similarity coefficient of the prostate volume area were 2.32% and 91.03%, respectively. CONCLUSION Quantitative experiments demonstrate that, compared with other methods, the results of our method are highly consistent with the expert manual segmentation.
Collapse
Affiliation(s)
- Yuchun Li
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of information and Communication Engineering, Hainan University, Haikou, China
| | - Cong Lin
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of information and Communication Engineering, Hainan University, Haikou, China.,College of Electronics and Information Engineering, Guangdong Ocean University, Zhanjiang, China
| | - Yu Zhang
- College of Computer science and Technology, Hainan University, Haikou, China
| | - Siling Feng
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of information and Communication Engineering, Hainan University, Haikou, China
| | - Mengxing Huang
- State Key Laboratory of Marine Resource Utilization in South China Sea, School of information and Communication Engineering, Hainan University, Haikou, China
| | - Zhiming Bai
- Haikou Municipal People's Hospital and Central South University Xiangya Medical College Affiliated Hospital, Haikou, China
| |
Collapse
|
15
|
Wong C, Fu Y, Li M, Mu S, Chu X, Fu J, Lin C, Zhang H. MRI-Based Artificial Intelligence in Rectal Cancer. J Magn Reson Imaging 2023; 57:45-56. [PMID: 35993550 DOI: 10.1002/jmri.28381] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2022] [Revised: 07/19/2022] [Accepted: 07/20/2022] [Indexed: 02/03/2023] Open
Abstract
Rectal cancer (RC) accounts for approximately one-third of colorectal cancer (CRC), with death rates increasing in patients younger than 50 years old. Magnetic resonance imaging (MRI) is routinely performed for tumor evaluation. However, the semantic features from images alone remain insufficient to guide treatment decisions. Functional MRIs are useful for revealing microstructural and functional abnormalities and nevertheless have low or modest repeatability and reproducibility. Therefore, during the preoperative evaluation and follow-up treatment of patients with RC, novel noninvasive imaging markers are needed to describe tumor characteristics to guide treatment strategies and achieve individualized diagnosis and treatment. In recent years, the development of artificial intelligence (AI) has created new tools for RC evaluation based on MRI. In this review, we summarize the research progress of AI in the evaluation of staging, prediction of high-risk factors, genotyping, response to therapy, recurrence, metastasis, prognosis, and segmentation with RC. We further discuss the challenges of clinical application, including improvement in imaging, model performance, and the biological meaning of features, which may also be major development directions in the future. EVIDENCE LEVEL: 5 TECHNICAL EFFICACY: Stage 2.
Collapse
Affiliation(s)
- Chinting Wong
- Department of Nuclear Medicine, The First Hospital of Jilin University, Changchun, China
| | - Yu Fu
- Department of Radiology, The First Hospital of Jilin University, Jilin Provincial Key Laboratory of Medical Imaging and Big Data, Changchun, China
| | - Mingyang Li
- Department of Radiology, The First Hospital of Jilin University, Jilin Provincial Key Laboratory of Medical Imaging and Big Data, Changchun, China
| | - Shengnan Mu
- Department of Radiology, The First Hospital of Jilin University, Jilin Provincial Key Laboratory of Medical Imaging and Big Data, Changchun, China
| | - Xiaotong Chu
- Department of Radiology, The First Hospital of Jilin University, Jilin Provincial Key Laboratory of Medical Imaging and Big Data, Changchun, China
| | - Jiahui Fu
- Department of Radiology, The First Hospital of Jilin University, Jilin Provincial Key Laboratory of Medical Imaging and Big Data, Changchun, China
| | - Chenghe Lin
- Department of Nuclear Medicine, The First Hospital of Jilin University, Changchun, China
| | - Huimao Zhang
- Department of Radiology, The First Hospital of Jilin University, Jilin Provincial Key Laboratory of Medical Imaging and Big Data, Changchun, China
| |
Collapse
|
16
|
Chen S, Zhong L, Qiu C, Zhang Z, Zhang X. Transformer-based multilevel region and edge aggregation network for magnetic resonance image segmentation. Comput Biol Med 2023; 152:106427. [PMID: 36543009 DOI: 10.1016/j.compbiomed.2022.106427] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2022] [Revised: 11/18/2022] [Accepted: 12/13/2022] [Indexed: 12/23/2022]
Abstract
To improve the quality of magnetic resonance (MR) image edge segmentation, some researchers applied additional edge labels to train the network to extract edge information and aggregate it with region information. They have made significant progress. However, due to the intrinsic locality of convolution operations, the convolution neural network-based region and edge aggregation has limitations in modeling long-range information. To solve this problem, we proposed a novel transformer-based multilevel region and edge aggregation network for MR image segmentation. To the best of our knowledge, this is the first literature on transformer-based region and edge aggregation. We first extract multilevel region and edge features using a dual-branch module. Then, the region and edge features at different levels are inferred and aggregated through multiple transformer-based inference modules to form multilevel complementary features. Finally, the attention feature selection module aggregates these complementary features with the corresponding level region and edge features to decode the region and edge features. We evaluated our method on a public MR dataset: Medical image computation and computer-assisted intervention atrial segmentation challenge (ASC). Meanwhile, the private MR dataset considered infrapatellar fat pad (IPFP). Our method achieved a dice score of 93.2% for ASC and 91.9% for IPFP. Compared with other 2D segmentation methods, our method improved a dice score by 0.6% for ASC and 3.0% for IPFP.
Collapse
Affiliation(s)
- Shaolong Chen
- School of Electronics and Communication Engineering, Sun Yat-sen University, Shenzhen, 518107, China
| | - Lijie Zhong
- Department of Medical Imaging, The Third Affiliated Hospital of Southern Medical University (Academy of Orthopedics·Guangdong Province), Guangzhou, 510630, China
| | - Changzhen Qiu
- School of Electronics and Communication Engineering, Sun Yat-sen University, Shenzhen, 518107, China
| | - Zhiyong Zhang
- School of Electronics and Communication Engineering, Sun Yat-sen University, Shenzhen, 518107, China.
| | - Xiaodong Zhang
- Department of Medical Imaging, The Third Affiliated Hospital of Southern Medical University (Academy of Orthopedics·Guangdong Province), Guangzhou, 510630, China.
| |
Collapse
|
17
|
Deep learning-based two-step organs at risk auto-segmentation model for brachytherapy planning in parotid gland carcinoma. J Contemp Brachytherapy 2022; 14:527-535. [PMID: 36819465 PMCID: PMC9924151 DOI: 10.5114/jcb.2022.123972] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2022] [Accepted: 12/02/2022] [Indexed: 01/18/2023] Open
Abstract
Purpose Delineation of organs at risk (OARs) represents a crucial step for both tailored delivery of radiation doses and prevention of radiation-induced toxicity in brachytherapy. Due to lack of studies on auto-segmentation methods in head and neck cancers, our study proposed a deep learning-based two-step approach for auto-segmentation of organs at risk in parotid carcinoma brachytherapy. Material and methods Computed tomography images of 200 patients with parotid gland carcinoma were used to train and evaluate our in-house developed two-step 3D nnU-Net-based model for OARs auto-segmentation. OARs during brachytherapy were defined as the auricula, condyle process, skin, mastoid process, external auditory canal, and mandibular ramus. Auto-segmentation results were compared to those of manual segmentation by expert oncologists. Accuracy was quantitatively evaluated in terms of dice similarity coefficient (DSC), Jaccard index, 95th-percentile Hausdorff distance (95HD), and precision and recall. Qualitative evaluation of auto-segmentation results was also performed. Results The mean DSC values of each OAR were 0.88, 0.91, 0.75, 0.89, 0.74, and 0.93, respectively, indicating close resemblance of auto-segmentation results to those of manual contouring. In addition, auto-segmentation could be completed within a minute, as compared with manual segmentation, which required over 20 minutes. All generated results were deemed clinically acceptable. Conclusions Our proposed deep learning-based two-step OARs auto-segmentation model demonstrated high efficiency and good agreement with gold standard manual contours. Thereby, this novel approach carries the potential in expediting the treatment planning process of brachytherapy for parotid gland cancers, while allowing for more accurate radiation delivery to minimize toxicity.
Collapse
|
18
|
Shiny Irene D, Indra Priyadharshini S, Tamizh Kuzhali R, Nancy P. An IoT based smart menstrual cup using optimized adaptive CNN model for effective menstrual hygiene management. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10308-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/05/2022]
|
19
|
Dong C, Xu S, Li Z. A novel end-to-end deep learning solution for coronary artery segmentation from CCTA. Med Phys 2022; 49:6945-6959. [PMID: 35770676 DOI: 10.1002/mp.15842] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2021] [Revised: 04/07/2022] [Accepted: 06/10/2022] [Indexed: 12/13/2022] Open
Abstract
PURPOSE Coronary computed tomographic angiography (CCTA) plays a vital role in the diagnosis of cardiovascular diseases, among which automatic coronary artery segmentation (CAS) serves as one of the most challenging tasks. To computationally assist the task, this paper proposes a novel end-to-end deep learning-based (DL) solution for automatic CAS. METHODS Inspired by the Di-Vnet network, a fully automatic multistage DL solution is proposed. The new solution aims to preserve the integrity of blood vessels in terms of both their shape details and continuity. The solution is developed using 338 CCTA cases, among which 133 cases (33865 axial images) have their ground-truth cardiac masks pre-annotated and 205 cases (53365 axial images) have their ground-truth coronary artery (CA) masks pre-annotated. The solution's accuracy is measured using dice similarity coefficient (DSC), 95th percentile Hausdorff Distance (95% HD), Recall, and Precision scores for CAS. RESULTS The proposed solution attains 90.29% in DSC, 2.11 mm in 95% HD, 97.02% in Recall, and 92.17% in Precision, respectively, which consumes 0.112 s per image and 30 s per case on average. Such performance of our method is superior to other state-of-the-art segmentation methods. CONCLUSIONS The novel DL solution is able to automatically learn to perform CAS in an end-to-end fashion, attaining a high accuracy, efficiency and robustness simultaneously.
Collapse
Affiliation(s)
- Caixia Dong
- Institute of Medical Artificial Intelligence, The Second Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China
| | - Songhua Xu
- Institute of Medical Artificial Intelligence, The Second Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China
| | - Zongfang Li
- Institute of Medical Artificial Intelligence, The Second Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China
| |
Collapse
|
20
|
Zhang Y, Yang J, Liu Y, Tian J, Wang S, Zhong C, Shi Z, Zhang Y, He Z. Decoupled pyramid correlation network for liver tumor segmentation from CT images. Med Phys 2022; 49:7207-7221. [PMID: 35620834 DOI: 10.1002/mp.15723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2021] [Revised: 03/24/2022] [Accepted: 04/16/2022] [Indexed: 12/13/2022] Open
Abstract
PURPOSE Automated liver tumor segmentation from computed tomography (CT) images is a necessary prerequisite in the interventions of hepatic abnormalities and surgery planning. However, accurate liver tumor segmentation remains challenging due to the large variability of tumor sizes and inhomogeneous texture. Recent advances based on fully convolutional network (FCN) for medical image segmentation drew on the success of learning discriminative pyramid features. In this paper, we propose a decoupled pyramid correlation network (DPC-Net) that exploits attention mechanisms to fully leverage both low- and high-level features embedded in FCN to segment liver tumor. METHODS We first design a powerful pyramid feature encoder (PFE) to extract multilevel features from input images. Then we decouple the characteristics of features concerning spatial dimension (i.e., height, width, depth) and semantic dimension (i.e., channel). On top of that, we present two types of attention modules, spatial correlation (SpaCor) and semantic correlation (SemCor) modules, to recursively measure the correlation of multilevel features. The former selectively emphasizes global semantic information in low-level features with the guidance of high-level ones. The latter adaptively enhance spatial details in high-level features with the guidance of low-level ones. RESULTS We evaluate the DPC-Net on MICCAI 2017 LiTS Liver Tumor Segmentation (LiTS) challenge data set. Dice similarity coefficient (DSC) and average symmetric surface distance (ASSD) are employed for evaluation. The proposed method obtains a DSC of 76.4% and an ASSD of 0.838 mm for liver tumor segmentation, outperforming the state-of-the-art methods. It also achieves a competitive result with a DSC of 96.0% and an ASSD of 1.636 mm for liver segmentation. CONCLUSIONS The experimental results show promising performance of DPC-Net for liver and tumor segmentation from CT images. Furthermore, the proposed SemCor and SpaCor can effectively model the multilevel correlation from both semantic and spatial dimensions. The proposed attention modules are lightweight and can be easily extended to other multilevel methods in an end-to-end manner.
Collapse
Affiliation(s)
- Yao Zhang
- Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China.,Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, China
| | - Jiawei Yang
- Electrical and Computer Engineering, University of California, Los Angeles, California, USA
| | - Yang Liu
- Institute of Computing Technology, Chinese Academy of Sciences, Beijing, China.,Computer Science and Technology, University of Chinese Academy of Sciences, Beijing, China
| | | | - Siyun Wang
- Dornsife College of Letters, Arts and Sciences, University of Southern California, Los Angeles, California, USA
| | | | | | | | | |
Collapse
|
21
|
|
22
|
Zabihollahy F, Viswanathan AN, Schmidt EJ, Lee J. Fully automated segmentation of clinical target volume in cervical cancer from magnetic resonance imaging with convolutional neural network. J Appl Clin Med Phys 2022; 23:e13725. [PMID: 35894782 PMCID: PMC9512359 DOI: 10.1002/acm2.13725] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Accepted: 06/25/2022] [Indexed: 01/14/2023] Open
Abstract
PURPOSE Contouring clinical target volume (CTV) from medical images is an essential step for radiotherapy (RT) planning. Magnetic resonance imaging (MRI) is used as a standard imaging modality for CTV segmentation in cervical cancer due to its superior soft-tissue contrast. However, the delineation of CTV is challenging as CTV contains microscopic extensions that are not clearly visible even in MR images, resulting in significant contour variability among radiation oncologists depending on their knowledge and experience. In this study, we propose a fully automated deep learning-based method to segment CTV from MR images. METHODS Our method begins with the bladder segmentation, from which the CTV position is estimated in the axial view. The superior-inferior CTV span is then detected using an Attention U-Net. A CTV-specific region of interest (ROI) is determined, and three-dimensional (3-D) blocks are extracted from the ROI volume. Finally, a CTV segmentation map is computed using a 3-D U-Net from the extracted 3-D blocks. RESULTS We developed and evaluated our method using 213 MRI scans obtained from 125 patients (183 for training, 30 for test). Our method achieved (mean ± SD) Dice similarity coefficient of 0.85 ± 0.03 and the 95th percentile Hausdorff distance of 3.70 ± 0.35 mm on test cases, outperforming other state-of-the-art methods significantly (p-value < 0.05). Our method also produces an uncertainty map along with the CTV segmentation by employing the Monte Carlo dropout technique to draw physician's attention to the regions with high uncertainty, where careful review and manual correction may be needed. CONCLUSIONS Experimental results show that the developed method is accurate, fast, and reproducible for contouring CTV from MRI, demonstrating its potential to assist radiation oncologists in alleviating the burden of tedious contouring for RT planning in cervical cancer.
Collapse
Affiliation(s)
- Fatemeh Zabihollahy
- Department of Radiation Oncology and Molecular Radiation SciencesJohns Hopkins University School of MedicineBaltimoreMarylandUSA
| | - Akila N. Viswanathan
- Department of Radiation Oncology and Molecular Radiation SciencesJohns Hopkins University School of MedicineBaltimoreMarylandUSA
| | - Ehud J. Schmidt
- Division of Cardiology, Department of MedicineJohns Hopkins University School of MedicineBaltimoreMarylandUSA
| | - Junghoon Lee
- Department of Radiation Oncology and Molecular Radiation SciencesJohns Hopkins University School of MedicineBaltimoreMarylandUSA
| |
Collapse
|