1
|
Saeed SU, Ramalhinho J, Pinnock M, Shen Z, Fu Y, Montaña-Brown N, Bonmati E, Barratt DC, Pereira SP, Davidson B, Clarkson MJ, Hu Y. Active learning using adaptable task-based prioritisation. Med Image Anal 2024; 95:103181. [PMID: 38640779 DOI: 10.1016/j.media.2024.103181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2023] [Revised: 04/03/2024] [Accepted: 04/12/2024] [Indexed: 04/21/2024]
Abstract
Supervised machine learning-based medical image computing applications necessitate expert label curation, while unlabelled image data might be relatively abundant. Active learning methods aim to prioritise a subset of available image data for expert annotation, for label-efficient model training. We develop a controller neural network that measures priority of images in a sequence of batches, as in batch-mode active learning, for multi-class segmentation tasks. The controller is optimised by rewarding positive task-specific performance gain, within a Markov decision process (MDP) environment that also optimises the task predictor. In this work, the task predictor is a segmentation network. A meta-reinforcement learning algorithm is proposed with multiple MDPs, such that the pre-trained controller can be adapted to a new MDP that contains data from different institutes and/or requires segmentation of different organs or structures within the abdomen. We present experimental results using multiple CT datasets from more than one thousand patients, with segmentation tasks of nine different abdominal organs, to demonstrate the efficacy of the learnt prioritisation controller function and its cross-institute and cross-organ adaptability. We show that the proposed adaptable prioritisation metric yields converging segmentation accuracy for a new kidney segmentation task, unseen in training, using between approximately 40% to 60% of labels otherwise required with other heuristic or random prioritisation metrics. For clinical datasets of limited size, the proposed adaptable prioritisation offers a performance improvement of 22.6% and 10.2% in Dice score, for tasks of kidney and liver vessel segmentation, respectively, compared to random prioritisation and alternative active sampling strategies.
Collapse
Affiliation(s)
- Shaheer U Saeed
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK.
| | - João Ramalhinho
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Mark Pinnock
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Ziyi Shen
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Yunguan Fu
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK; InstaDeep, London, UK
| | - Nina Montaña-Brown
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Ester Bonmati
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK; School of Computer Science and Engineering, University of Westminster, London, UK
| | - Dean C Barratt
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Stephen P Pereira
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK; Institute for Liver and Digestive Health, University College London, London, UK
| | - Brian Davidson
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK; Centre for Surgical Innovation, Organ Regeneration and Transplantation (CISORT), Division of Surgery & Interventional Science, University College London, London, UK
| | - Matthew J Clarkson
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| | - Yipeng Hu
- Centre for Medical Image Computing, Wellcome/EPSRC Centre for Interventional & Surgical Sciences, and Department of Medical Physics & Biomedical Engineering, University College London, London, UK
| |
Collapse
|
2
|
Martín-Saladich Q, Pericàs JM, Ciudin A, Ramirez-Serra C, Escobar M, Rivera-Esteban J, Aguadé-Bruix S, González Ballester MA, Herance JR. Metabolic-associated fatty liver voxel-based quantification on CT images using a contrast adapted automatic tool. Med Image Anal 2024; 95:103185. [PMID: 38718716 DOI: 10.1016/j.media.2024.103185] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Revised: 12/22/2023] [Accepted: 04/19/2024] [Indexed: 06/01/2024]
Abstract
BACKGROUND & AIMS Metabolic-dysfunction associated fatty liver disease (MAFLD) is highly prevalent and can lead to liver complications and comorbidities, with non-invasive tests such as vibration-controlled transient elastography (VCTE) and invasive liver biopsies being used for diagnosis The aim of the present study was to develop a new fully automatized method for quantifying the percentage of fat in the liver based on a voxel analysis on computed tomography (CT) images to solve previously unconcluded diagnostic deficiencies either in contrast (CE) or non-contrast enhanced (NCE) assessments. METHODS Liver and spleen were segmented using nn-UNet on CE- and NCE-CT images. Radiodensity values were obtained for both organs for defining the key benchmarks for fatty liver assessment: liver mean, liver-to-spleen ratio, liver-spleen difference, and their average. VCTE was used for validation. A classification task method was developed for detection of suitable patients to fulfill maximum reproducibility across cohorts and highlight subjects with other potential radiodensity-related diseases. RESULTS Best accuracy was attained using the average of all proposed benchmarks being the liver-to-spleen ratio highly useful for CE and the liver-to-spleen difference for NCE. The proposed whole-organ automatic segmentation displayed superior potential when compared to the typically used manual region-of-interest drawing as it allows to accurately obtain the percent of fat in liver, among other improvements. Atypical patients were successfully stratified through a function based on biochemical data. CONCLUSIONS The developed method tackles the current drawbacks including biopsy invasiveness, and CT-related weaknesses such as lack of automaticity, dependency on contrast agent, no quantification of the percentage of fat in liver, and limited information on region-to-organ affectation. We propose this tool as an alternative for individualized MAFLD evaluation by an early detection of abnormal CT patterns based in radiodensity whilst abording detection of non-suitable patients to avoid unnecessary exposure to CT radiation. Furthermore, this work presents a surrogate aid for assessing fatty liver at a primary assessment of MAFLD using elastography data.
Collapse
Affiliation(s)
- Queralt Martín-Saladich
- Nuclear Medicine, Radiology and Cardiology Departments, Medical Molecular Imaging Research Group, Vall d'Hebron Research Institute (VHIR), Vall d'Hebron University Hospital, Autonomous University Barcelona, Barcelona 08035, Spain; Department of Information and Communication Technologies, BCN MedTech, Universitat Pompeu Fabra, Barcelona 08018, Spain
| | - Juan M Pericàs
- Vall d'Hebron Institute for Research, Liver Unit, Vall d'Hebron University Hospital, Barcelona 08035, Spain; Centro de Investigación Biomédica en Red de Enfermedades Hepáticas y Digestivas (CIBEREHD), Instituto de Salud Carlos III, Madrid 28029, Spain
| | - Andreea Ciudin
- Endocrinology Department, Diabetes and Metabolism Research Group, VHIR, Vall d'Hebron University Hospital, Autonomous University Barcelona, Barcelona 08035, Spain; Centro de Investigación Biomédica en Red de Diabetes y Enfermedades Metabólicas (CIBERDEM), Instituto de Salud Carlos III, Madrid 28029, Spain
| | - Clara Ramirez-Serra
- Clinical Biochemistry Research Group, Vall d'Hebron Research Institute (VHIR), Biochemical Core Facilities, Vall d'Hebron University Hospital, Autonomous University Barcelona, Barcelona 08035, Spain
| | - Manuel Escobar
- Nuclear Medicine, Radiology and Cardiology Departments, Medical Molecular Imaging Research Group, Vall d'Hebron Research Institute (VHIR), Vall d'Hebron University Hospital, Autonomous University Barcelona, Barcelona 08035, Spain
| | - Jesús Rivera-Esteban
- Vall d'Hebron Institute for Research, Liver Unit, Vall d'Hebron University Hospital, Barcelona 08035, Spain; Centro de Investigación Biomédica en Red de Enfermedades Hepáticas y Digestivas (CIBEREHD), Instituto de Salud Carlos III, Madrid 28029, Spain
| | - Santiago Aguadé-Bruix
- Nuclear Medicine, Radiology and Cardiology Departments, Medical Molecular Imaging Research Group, Vall d'Hebron Research Institute (VHIR), Vall d'Hebron University Hospital, Autonomous University Barcelona, Barcelona 08035, Spain
| | - Miguel A González Ballester
- Department of Information and Communication Technologies, BCN MedTech, Universitat Pompeu Fabra, Barcelona 08018, Spain; Catalan Institution for Research and Advanced Studies (ICREA), Barcelona 08010, Spain
| | - José Raul Herance
- Centro de Investigación Biomédica en Red de Enfermedades Hepáticas y Digestivas (CIBEREHD), Instituto de Salud Carlos III, Madrid 28029, Spain; Centro de Investigación Biomédica en Red de Bioingeniería, Biomateriales y Nanomedicina, Instituto de Salud Carlos III, Madrid 28029, Spain.
| |
Collapse
|
3
|
Lu A, Huang H, Hu Y, Zbijewski W, Unberath M, Siewerdsen JH, Weiss CR, Sisniega A. Vessel-targeted compensation of deformable motion in interventional cone-beam CT. Med Image Anal 2024; 97:103254. [PMID: 38968908 DOI: 10.1016/j.media.2024.103254] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Revised: 06/01/2024] [Accepted: 06/24/2024] [Indexed: 07/07/2024]
Abstract
The present standard of care for unresectable liver cancer is transarterial chemoembolization (TACE), which involves using chemotherapeutic particles to selectively embolize the arteries supplying hepatic tumors. Accurate volumetric identification of intricate fine vascularity is crucial for selective embolization. Three-dimensional imaging, particularly cone-beam CT (CBCT), aids in visualization and targeting of small vessels in such highly variable anatomy, but long image acquisition time results in intra-scan patient motion, which distorts vascular structures and tissue boundaries. To improve clarity of vascular anatomy and intra-procedural utility, this work proposes a targeted motion estimation and compensation framework that removes the need for any prior information or external tracking and for user interaction. Motion estimation is performed in two stages: (i) a target identification stage that segments arteries and catheters in the projection domain using a multi-view convolutional neural network to construct a coarse 3D vascular mask; and (ii) a targeted motion estimation stage that iteratively solves for the time-varying motion field via optimization of a vessel-enhancing objective function computed over the target vascular mask. The vessel-enhancing objective is derived through eigenvalues of the local image Hessian to emphasize bright tubular structures. Motion compensation is achieved via spatial transformer operators that apply time-dependent deformations to partial angle reconstructions, allowing efficient minimization via gradient backpropagation. The framework was trained and evaluated in anatomically realistic simulated motion-corrupted CBCTs mimicking TACE of hepatic tumors, at intermediate (3.0 mm) and large (6.0 mm) motion magnitudes. Motion compensation substantially improved median vascular DICE score (from 0.30 to 0.59 for large motion), image SSIM (from 0.77 to 0.93 for large motion), and vessel sharpness (0.189 mm-1 to 0.233 mm-1 for large motion) in simulated cases. Motion compensation also demonstrated increased vessel sharpness (0.188 mm-1 before to 0.205 mm-1 after) and reconstructed vessel length (median increased from 37.37 to 41.00 mm) on a clinical interventional CBCT. The proposed anatomy-aware motion compensation framework presented a promising approach for improving the utility of CBCT for intra-procedural vascular imaging, facilitating selective embolization procedures.
Collapse
Affiliation(s)
- Alexander Lu
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, Traylor Research Building, #622 720 Rutland Avenue Baltimore MD 21205, USA
| | - Heyuan Huang
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, Traylor Research Building, #622 720 Rutland Avenue Baltimore MD 21205, USA
| | - Yicheng Hu
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Wojciech Zbijewski
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, Traylor Research Building, #622 720 Rutland Avenue Baltimore MD 21205, USA
| | - Mathias Unberath
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Jeffrey H Siewerdsen
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, Traylor Research Building, #622 720 Rutland Avenue Baltimore MD 21205, USA; Departments of Imaging Physics, Radiation Physics, and Neurosurgery, The University of Texas M.D. Anderson Cancer Center, TX, USA
| | - Clifford R Weiss
- Department of Radiology, Johns Hopkins University, Baltimore, MD, USA
| | - Alejandro Sisniega
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD, Traylor Research Building, #622 720 Rutland Avenue Baltimore MD 21205, USA.
| |
Collapse
|
4
|
dos Santos PV, Scoczynski Ribeiro Martins M, Amorim Nogueira S, Gonçalves C, Maffei Loureiro R, Pacheco Calixto W. Unsupervised model for structure segmentation applied to brain computed tomography. PLoS One 2024; 19:e0304017. [PMID: 38870119 PMCID: PMC11175403 DOI: 10.1371/journal.pone.0304017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Accepted: 05/03/2024] [Indexed: 06/15/2024] Open
Abstract
This article presents an unsupervised method for segmenting brain computed tomography scans. The proposed methodology involves image feature extraction and application of similarity and continuity constraints to generate segmentation maps of the anatomical head structures. Specifically designed for real-world datasets, this approach applies a spatial continuity scoring function tailored to the desired number of structures. The primary objective is to assist medical experts in diagnosis by identifying regions with specific abnormalities. Results indicate a simplified and accessible solution, reducing computational effort, training time, and financial costs. Moreover, the method presents potential for expediting the interpretation of abnormal scans, thereby impacting clinical practice. This proposed approach might serve as a practical tool for segmenting brain computed tomography scans, and make a significant contribution to the analysis of medical images in both research and clinical settings.
Collapse
Affiliation(s)
- Paulo Victor dos Santos
- Electrical, Mechanical & Computer Engineering School, Federal University of Goias, Goiania, Brazil
- Department of Radiology, Hospital Israelita Albert Einstein, Sao Paulo, Sao Paulo, Brazil
- Technology Research and Development Center (GCITE), Federal Institute of Goias, Goiania, Brazil
| | - Marcella Scoczynski Ribeiro Martins
- Electrical, Mechanical & Computer Engineering School, Federal University of Goias, Goiania, Brazil
- Federal University of Technology - Parana, Ponta Grossa, Parana, Brazil
| | - Solange Amorim Nogueira
- Electrical, Mechanical & Computer Engineering School, Federal University of Goias, Goiania, Brazil
- Department of Radiology, Hospital Israelita Albert Einstein, Sao Paulo, Sao Paulo, Brazil
| | | | - Rafael Maffei Loureiro
- Department of Radiology, Hospital Israelita Albert Einstein, Sao Paulo, Sao Paulo, Brazil
| | - Wesley Pacheco Calixto
- Electrical, Mechanical & Computer Engineering School, Federal University of Goias, Goiania, Brazil
- Technology Research and Development Center (GCITE), Federal Institute of Goias, Goiania, Brazil
| |
Collapse
|
5
|
Liu X, Qu L, Xie Z, Zhao J, Shi Y, Song Z. Towards more precise automatic analysis: a systematic review of deep learning-based multi-organ segmentation. Biomed Eng Online 2024; 23:52. [PMID: 38851691 PMCID: PMC11162022 DOI: 10.1186/s12938-024-01238-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2023] [Accepted: 04/11/2024] [Indexed: 06/10/2024] Open
Abstract
Accurate segmentation of multiple organs in the head, neck, chest, and abdomen from medical images is an essential step in computer-aided diagnosis, surgical navigation, and radiation therapy. In the past few years, with a data-driven feature extraction approach and end-to-end training, automatic deep learning-based multi-organ segmentation methods have far outperformed traditional methods and become a new research topic. This review systematically summarizes the latest research in this field. We searched Google Scholar for papers published from January 1, 2016 to December 31, 2023, using keywords "multi-organ segmentation" and "deep learning", resulting in 327 papers. We followed the PRISMA guidelines for paper selection, and 195 studies were deemed to be within the scope of this review. We summarized the two main aspects involved in multi-organ segmentation: datasets and methods. Regarding datasets, we provided an overview of existing public datasets and conducted an in-depth analysis. Concerning methods, we categorized existing approaches into three major classes: fully supervised, weakly supervised and semi-supervised, based on whether they require complete label information. We summarized the achievements of these methods in terms of segmentation accuracy. In the discussion and conclusion section, we outlined and summarized the current trends in multi-organ segmentation.
Collapse
Affiliation(s)
- Xiaoyu Liu
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, 138 Yixueyuan Road, Shanghai, 200032, People's Republic of China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Linhao Qu
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, 138 Yixueyuan Road, Shanghai, 200032, People's Republic of China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Ziyue Xie
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, 138 Yixueyuan Road, Shanghai, 200032, People's Republic of China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Jiayue Zhao
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, 138 Yixueyuan Road, Shanghai, 200032, People's Republic of China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Yonghong Shi
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, 138 Yixueyuan Road, Shanghai, 200032, People's Republic of China.
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China.
| | - Zhijian Song
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, 138 Yixueyuan Road, Shanghai, 200032, People's Republic of China.
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China.
| |
Collapse
|
6
|
Liu J, Zhang Y, Wang K, Yavuz MC, Chen X, Yuan Y, Li H, Yang Y, Yuille A, Tang Y, Zhou Z. Universal and extensible language-vision models for organ segmentation and tumor detection from abdominal computed tomography. Med Image Anal 2024; 97:103226. [PMID: 38852215 DOI: 10.1016/j.media.2024.103226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2023] [Revised: 03/30/2024] [Accepted: 05/27/2024] [Indexed: 06/11/2024]
Abstract
The advancement of artificial intelligence (AI) for organ segmentation and tumor detection is propelled by the growing availability of computed tomography (CT) datasets with detailed, per-voxel annotations. However, these AI models often struggle with flexibility for partially annotated datasets and extensibility for new classes due to limitations in the one-hot encoding, architectural design, and learning scheme. To overcome these limitations, we propose a universal, extensible framework enabling a single model, termed Universal Model, to deal with multiple public datasets and adapt to new classes (e.g., organs/tumors). Firstly, we introduce a novel language-driven parameter generator that leverages language embeddings from large language models, enriching semantic encoding compared with one-hot encoding. Secondly, the conventional output layers are replaced with lightweight, class-specific heads, allowing Universal Model to simultaneously segment 25 organs and six types of tumors and ease the addition of new classes. We train our Universal Model on 3410 CT volumes assembled from 14 publicly available datasets and then test it on 6173 CT volumes from four external datasets. Universal Model achieves first place on six CT tasks in the Medical Segmentation Decathlon (MSD) public leaderboard and leading performance on the Beyond The Cranial Vault (BTCV) dataset. In summary, Universal Model exhibits remarkable computational efficiency (6× faster than other dataset-specific models), demonstrates strong generalization across different hospitals, transfers well to numerous downstream tasks, and more importantly, facilitates the extensibility to new classes while alleviating the catastrophic forgetting of previously learned classes. Codes, models, and datasets are available at https://github.com/ljwztc/CLIP-Driven-Universal-Model.
Collapse
Affiliation(s)
- Jie Liu
- City University of Hong Kong, Hong Kong
| | - Yixiao Zhang
- Johns Hopkins University, United States of America
| | - Kang Wang
- University of California, San Francisco, United States of America
| | - Mehmet Can Yavuz
- University of California, San Francisco, United States of America
| | - Xiaoxi Chen
- University of Illinois Urbana-Champaign, United States of America
| | | | | | - Yang Yang
- University of California, San Francisco, United States of America
| | - Alan Yuille
- Johns Hopkins University, United States of America
| | | | - Zongwei Zhou
- Johns Hopkins University, United States of America.
| |
Collapse
|
7
|
Huang H, Liu Y, Siewerdsen JH, Lu A, Hu Y, Zbijewski W, Unberath M, Weiss CR, Sisniega A. Deformable motion compensation in interventional cone-beam CT with a context-aware learned autofocus metric. Med Phys 2024; 51:4158-4180. [PMID: 38733602 DOI: 10.1002/mp.17125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/25/2023] [Revised: 04/02/2024] [Accepted: 05/03/2024] [Indexed: 05/13/2024] Open
Abstract
PURPOSE Interventional Cone-Beam CT (CBCT) offers 3D visualization of soft-tissue and vascular anatomy, enabling 3D guidance of abdominal interventions. However, its long acquisition time makes CBCT susceptible to patient motion. Image-based autofocus offers a suitable platform for compensation of deformable motion in CBCT, but it relies on handcrafted motion metrics based on first-order image properties and that lack awareness of the underlying anatomy. This work proposes a data-driven approach to motion quantification via a learned, context-aware, deformable metric,VI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ , that quantifies the amount of motion degradation as well as the realism of the structural anatomical content in the image. METHODS The proposedVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ was modeled as a deep convolutional neural network (CNN) trained to recreate a reference-based structural similarity metric-visual information fidelity (VIF). The deep CNN acted on motion-corrupted images, providing an estimation of the spatial VIF map that would be obtained against a motion-free reference, capturing motion distortion, and anatomic plausibility. The deep CNN featured a multi-branch architecture with a high-resolution branch for estimation of voxel-wise VIF on a small volume of interest. A second contextual, low-resolution branch provided features associated to anatomical context for disentanglement of motion effects and anatomical appearance. The deep CNN was trained on paired motion-free and motion-corrupted data obtained with a high-fidelity forward projection model for a protocol involving 120 kV and 9.90 mGy. The performance ofVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ was evaluated via metrics of correlation with ground truth VIF ${\bm{VIF}}$ and with the underlying deformable motion field in simulated data with deformable motion fields with amplitude ranging from 5 to 20 mm and frequency from 2.4 up to 4 cycles/scan. Robustness to variation in tissue contrast and noise levels was assessed in simulation studies with varying beam energy (90-120 kV) and dose (1.19-39.59 mGy). Further validation was obtained on experimental studies with a deformable phantom. Final validation was obtained via integration ofVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ on an autofocus compensation framework, applied to motion compensation on experimental datasets and evaluated via metric of spatial resolution on soft-tissue boundaries and sharpness of contrast-enhanced vascularity. RESULTS The magnitude and spatial map ofVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ showed consistent and high correlation levels with the ground truth in both simulation and real data, yielding average normalized cross correlation (NCC) values of 0.95 and 0.88, respectively. Similarly,VI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ achieved good correlation values with the underlying motion field, with average NCC of 0.90. In experimental phantom studies,VI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ properly reflects the change in motion amplitudes and frequencies: voxel-wise averaging of the localVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ across the full reconstructed volume yielded an average value of 0.69 for the case with mild motion (2 mm, 12 cycles/scan) and 0.29 for the case with severe motion (12 mm, 6 cycles/scan). Autofocus motion compensation usingVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ resulted in noticeable mitigation of motion artifacts and improved spatial resolution of soft tissue and high-contrast structures, resulting in reduction of edge spread function width of 8.78% and 9.20%, respectively. Motion compensation also increased the conspicuity of contrast-enhanced vascularity, reflected in an increase of 9.64% in vessel sharpness. CONCLUSION The proposedVI F D L ${\bm{VI}}{{\bm{F}}}_{DL}$ , featuring a novel context-aware architecture, demonstrated its capacity as a reference-free surrogate of structural similarity to quantify motion-induced degradation of image quality and anatomical plausibility of image content. The validation studies showed robust performance across motion patterns, x-ray techniques, and anatomical instances. The proposed anatomy- and context-aware metric poses a powerful alternative to conventional motion estimation metrics, and a step forward for application of deep autofocus motion compensation for guidance in clinical interventional procedures.
Collapse
Affiliation(s)
- Heyuan Huang
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland, USA
| | - Yixuan Liu
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland, USA
| | - Jeffrey H Siewerdsen
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland, USA
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Alexander Lu
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland, USA
| | - Yicheng Hu
- Department of Computer Science, Johns Hopkins University, Baltimore, Maryland, USA
| | - Wojciech Zbijewski
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland, USA
| | - Mathias Unberath
- Department of Computer Science, Johns Hopkins University, Baltimore, Maryland, USA
| | - Clifford R Weiss
- Department of Radiology, Johns Hopkins University, Baltimore, Maryland, USA
| | - Alejandro Sisniega
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore, Maryland, USA
| |
Collapse
|
8
|
Singh SB, Sarrami AH, Gatidis S, Varniab ZS, Chaudhari A, Daldrup-Link HE. Applications of Artificial Intelligence for Pediatric Cancer Imaging. AJR Am J Roentgenol 2024. [PMID: 38809123 DOI: 10.2214/ajr.24.31076] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/30/2024]
Abstract
Artificial intelligence (AI) is transforming medical imaging of adult patients. However, its utilization in pediatric oncology imaging remains constrained, in part due to the inherent data scarcity associated with childhood cancers. Pediatric cancers are rare, and imaging technologies are evolving rapidly, leading to insufficient data of a particular type to effectively train these algorithms. The small market size of pediatrics compared to adults could also contribute to this challenge, as market size is a driver of commercialization. This article provides an overview of the current state of AI applications for pediatric cancer imaging, including applications for medical image acquisition, processing, reconstruction, segmentation, diagnosis, staging, and treatment response monitoring. While current developments are promising, impediments due to diverse anatomies of growing children and nonstandardized imaging protocols have led to limited clinical translation thus far. Opportunities include leveraging reconstruction algorithms to achieve accelerated low-dose imaging and automating the generation of metric-based staging and treatment monitoring scores. Transfer-learning of adult-based AI models to pediatric cancers, multi-institutional data sharing, and ethical data privacy practices for pediatric patients with rare cancers will be keys to unlocking AI's full potential for clinical translation and improved outcomes for these young patients.
Collapse
Affiliation(s)
- Shashi B Singh
- Department of Radiology, Division of Pediatric Radiology, Stanford University School of Medicine, Stanford, CA, 94305
| | - Amir H Sarrami
- Department of Radiology, Division of Pediatric Radiology, Stanford University School of Medicine, Stanford, CA, 94305
| | - Sergios Gatidis
- Department of Radiology, Division of Pediatric Radiology, Stanford University School of Medicine, Stanford, CA, 94305
| | - Zahra S Varniab
- Department of Radiology, Division of Pediatric Radiology, Stanford University School of Medicine, Stanford, CA, 94305
| | - Akshay Chaudhari
- Department of Radiology, Integrative Biomedical Imaging Informatics (IBIIS), Stanford University School of Medicine, Stanford University, Stanford, CA, 94304
- Department of Biomedical Data Science, Stanford University School of Medicine, Stanford University, Stanford, CA, 94304
| | - Heike E Daldrup-Link
- Department of Radiology, Division of Pediatric Radiology, Stanford University School of Medicine, Stanford, CA, 94305
- Department of Pediatrics, Pediatric Hematology/Oncology, Lucile Packard Children's Hospital, Stanford University, Stanford, CA, 94305
| |
Collapse
|
9
|
Koitka S, Baldini G, Kroll L, van Landeghem N, Pollok OB, Haubold J, Pelka O, Kim M, Kleesiek J, Nensa F, Hosch R. SAROS: A dataset for whole-body region and organ segmentation in CT imaging. Sci Data 2024; 11:483. [PMID: 38729970 PMCID: PMC11087485 DOI: 10.1038/s41597-024-03337-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Accepted: 05/01/2024] [Indexed: 05/12/2024] Open
Abstract
The Sparsely Annotated Region and Organ Segmentation (SAROS) dataset was created using data from The Cancer Imaging Archive (TCIA) to provide a large open-access CT dataset with high-quality annotations of body landmarks. In-house segmentation models were employed to generate annotation proposals on randomly selected cases from TCIA. The dataset includes 13 semantic body region labels (abdominal/thoracic cavity, bones, brain, breast implant, mediastinum, muscle, parotid/submandibular/thyroid glands, pericardium, spinal cord, subcutaneous tissue) and six body part labels (left/right arm/leg, head, torso). Case selection was based on the DICOM series description, gender, and imaging protocol, resulting in 882 patients (438 female) for a total of 900 CTs. Manual review and correction of proposals were conducted in a continuous quality control cycle. Only every fifth axial slice was annotated, yielding 20150 annotated slices from 28 data collections. For the reproducibility on downstream tasks, five cross-validation folds and a test set were pre-defined. The SAROS dataset serves as an open-access resource for training and evaluating novel segmentation models, covering various scanner vendors and diseases.
Collapse
Affiliation(s)
- Sven Koitka
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
| | - Giulia Baldini
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
| | - Lennard Kroll
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
| | - Natalie van Landeghem
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
| | - Olivia B Pollok
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
| | - Johannes Haubold
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
| | - Obioma Pelka
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
- Data Integration Center, Central IT Department, University Hospital Essen, Essen, Germany
| | - Moon Kim
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
| | - Jens Kleesiek
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
| | - Felix Nensa
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany
| | - René Hosch
- Institute of Interventional and Diagnostic Radiology and Neuroradiology, University Hospital Essen, Essen, Germany.
- Institute for Artificial Intelligence in Medicine, University Hospital Essen, Essen, Germany.
| |
Collapse
|
10
|
Hresko DJ, Drotar P. BucketAugment: Reinforced Domain Generalisation in Abdominal CT Segmentation. IEEE OPEN JOURNAL OF ENGINEERING IN MEDICINE AND BIOLOGY 2024; 5:353-361. [PMID: 38899027 PMCID: PMC11186658 DOI: 10.1109/ojemb.2024.3397623] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2023] [Revised: 02/28/2024] [Accepted: 05/03/2024] [Indexed: 06/21/2024] Open
Abstract
Goal: In recent years, deep neural networks have consistently outperformed previously proposed methods in the domain of medical segmentation. However, due to their nature, these networks often struggle to delineate desired structures in data that fall outside their training distribution. The goal of this study is to address the challenges associated with domain generalization in CT segmentation by introducing a novel method called BucketAugment for deep neural networks. Methods: BucketAugment leverages principles from the Q-learning algorithm and employs validation loss to search for an optimal policy within a search space comprised of distributed stacks of 3D volumetric augmentations, termed 'buckets.' These buckets have tunable parameters and can be seamlessly integrated into existing neural network architectures, offering flexibility for customization. Results: In our experiments, we focus on segmenting kidney and liver structures across three distinct medical datasets, each containing CT scans of the abdominal region collected from various clinical institutions and scanner vendors. Our results indicate that BucketAugment significantly enhances domain generalization across diverse medical datasets, requiring only minimal modifications to existing network architectures. Conclusions: The introduction of BucketAugment provides a promising solution to the challenges of domain generalization in CT segmentation. By leveraging Q-learning principles and distributed stacks of 3D augmentations, this method improves the performance of deep neural networks on medical segmentation tasks, demonstrating its potential to enhance the applicability of such models across different datasets and clinical scenarios.
Collapse
Affiliation(s)
| | - Peter Drotar
- Technical University of Kosice040 01KosiceSlovakia
| |
Collapse
|
11
|
Kim J, Li Y, Shin BS. 3D-DGGAN: A Data-Guided Generative Adversarial Network for High Fidelity in Medical Image Generation. IEEE J Biomed Health Inform 2024; 28:2904-2915. [PMID: 38416610 DOI: 10.1109/jbhi.2024.3367375] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/01/2024]
Abstract
Three-dimensional images are frequently used in medical imaging research for classification, segmentation, and detection. However, the limited availability of 3D images hinders research progress due to network training difficulties. Generative methods have been proposed to create medical images using AI techniques. Nevertheless, 2D approaches have difficulty dealing with 3D anatomical structures, which can result in discontinuities between slices. To mitigate these discontinuities, several 3D generative networks have been proposed. However, the scarcity of available 3D images makes training these networks with limited samples inadequate for producing high-fidelity 3D images. We propose a data-guided generative adversarial network to provide high fidelity in 3D image generation. The generator creates fake images with noise using reference code obtained by extracting features from real images. The generator also creates decoded images using reference code without noise. These decoded images are compared to the real images to evaluate fidelity in the reference code. This generation process can create high-fidelity 3D images from only a small amount of real training data. Additionally, our method employs three types of discriminator: volume (evaluates all the slices), slab (evaluates a set of consecutive slices), and slice (evaluates randomly selected slices). The proposed discriminator enhances fidelity by differentiating between real and fake images based on detailed characteristics. Results from our method are compared with existing methods by using quantitative analysis such as Fréchet inception distance and maximum mean discrepancy. The results demonstrate that our method produces more realistic 3D images than existing methods.
Collapse
|
12
|
Rong Y, Chen Q, Fu Y, Yang X, Al-Hallaq HA, Wu QJ, Yuan L, Xiao Y, Cai B, Latifi K, Benedict SH, Buchsbaum JC, Qi XS. NRG Oncology Assessment of Artificial Intelligence Deep Learning-Based Auto-segmentation for Radiation Therapy: Current Developments, Clinical Considerations, and Future Directions. Int J Radiat Oncol Biol Phys 2024; 119:261-280. [PMID: 37972715 PMCID: PMC11023777 DOI: 10.1016/j.ijrobp.2023.10.033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2023] [Revised: 09/16/2023] [Accepted: 10/14/2023] [Indexed: 11/19/2023]
Abstract
Deep learning neural networks (DLNN) in Artificial intelligence (AI) have been extensively explored for automatic segmentation in radiotherapy (RT). In contrast to traditional model-based methods, data-driven AI-based models for auto-segmentation have shown high accuracy in early studies in research settings and controlled environment (single institution). Vendor-provided commercial AI models are made available as part of the integrated treatment planning system (TPS) or as a stand-alone tool that provides streamlined workflow interacting with the main TPS. These commercial tools have drawn clinics' attention thanks to their significant benefit in reducing the workload from manual contouring and shortening the duration of treatment planning. However, challenges occur when applying these commercial AI-based segmentation models to diverse clinical scenarios, particularly in uncontrolled environments. Contouring nomenclature and guideline standardization has been the main task undertaken by the NRG Oncology. AI auto-segmentation holds the potential clinical trial participants to reduce interobserver variations, nomenclature non-compliance, and contouring guideline deviations. Meanwhile, trial reviewers could use AI tools to verify contour accuracy and compliance of those submitted datasets. In recognizing the growing clinical utilization and potential of these commercial AI auto-segmentation tools, NRG Oncology has formed a working group to evaluate the clinical utilization and potential of commercial AI auto-segmentation tools. The group will assess in-house and commercially available AI models, evaluation metrics, clinical challenges, and limitations, as well as future developments in addressing these challenges. General recommendations are made in terms of the implementation of these commercial AI models, as well as precautions in recognizing the challenges and limitations.
Collapse
Affiliation(s)
- Yi Rong
- Mayo Clinic Arizona, Phoenix, AZ
| | - Quan Chen
- City of Hope Comprehensive Cancer Center Duarte, CA
| | - Yabo Fu
- Memorial Sloan Kettering Cancer Center, Commack, NY
| | | | | | | | - Lulin Yuan
- Virginia Commonwealth University, Richmond, VA
| | - Ying Xiao
- University of Pennsylvania/Abramson Cancer Center, Philadelphia, PA
| | - Bin Cai
- The University of Texas Southwestern Medical Center, Dallas, TX
| | | | - Stanley H Benedict
- University of California Davis Comprehensive Cancer Center, Sacramento, CA
| | | | - X Sharon Qi
- University of California Los Angeles, Los Angeles, CA
| |
Collapse
|
13
|
Liu C, Klein L, Huang Y, Baader E, Lell M, Kachelrieß M, Maier A. Two-view topogram-based anatomy-guided CT reconstruction for prospective risk minimization. Sci Rep 2024; 14:9373. [PMID: 38653993 DOI: 10.1038/s41598-024-59731-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2024] [Accepted: 04/15/2024] [Indexed: 04/25/2024] Open
Abstract
To facilitate a prospective estimation of the effective dose of an CT scan prior to the actual scanning in order to use sophisticated patient risk minimizing methods, a prospective spatial dose estimation and the known anatomical structures are required. To this end, a CT reconstruction method is required to reconstruct CT volumes from as few projections as possible, i.e. by using the topograms, with anatomical structures as correct as possible. In this work, an optimized CT reconstruction model based on a generative adversarial network (GAN) is proposed. The GAN is trained to reconstruct 3D volumes from an anterior-posterior and a lateral CT projection. To enhance anatomical structures, a pre-trained organ segmentation network and the 3D perceptual loss are applied during the training phase, so that the model can then generate both organ-enhanced CT volume and organ segmentation masks. The proposed method can reconstruct CT volumes with PSNR of 26.49, RMSE of 196.17, and SSIM of 0.64, compared to 26.21, 201.55 and 0.63 using the baseline method. In terms of the anatomical structure, the proposed method effectively enhances the organ shapes and boundaries and allows for a straight-forward identification of the relevant anatomical structures. We note that conventional reconstruction metrics fail to indicate the enhancement of anatomical structures. In addition to such metrics, the evaluation is expanded with assessing the organ segmentation performance. The average organ dice of the proposed method is 0.71 compared with 0.63 for the baseline model, indicating the enhancement of anatomical structures.
Collapse
Affiliation(s)
- Chang Liu
- Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany.
| | - Laura Klein
- Division of X-Ray Imaging and Computed Tomography, German Cancer Research Center (DKFZ), Heidelberg, Germany
- Department of Physics and Astronomy, Ruprecht-Karls-University Heidelberg, Heidelberg, Germany
| | - Yixing Huang
- Department of Radiation Oncology, Universitätsklinikum Erlangen, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany
| | - Edith Baader
- Division of X-Ray Imaging and Computed Tomography, German Cancer Research Center (DKFZ), Heidelberg, Germany
- Department of Physics and Astronomy, Ruprecht-Karls-University Heidelberg, Heidelberg, Germany
| | - Michael Lell
- Department of Radiology and Nuclear Medicine, Klinikum Nürnberg, Paracelsus Medical University, Nuremberg, Germany
| | - Marc Kachelrieß
- Division of X-Ray Imaging and Computed Tomography, German Cancer Research Center (DKFZ), Heidelberg, Germany
- Medical Faculty, Ruprecht-Karls-University Heidelberg, Heidelberg, Germany
| | - Andreas Maier
- Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), Erlangen, Germany
| |
Collapse
|
14
|
Selle M, Kircher M, Schwennen C, Visscher C, Jung K. Dimension reduction and outlier detection of 3-D shapes derived from multi-organ CT images. BMC Med Inform Decis Mak 2024; 24:49. [PMID: 38355504 PMCID: PMC10865689 DOI: 10.1186/s12911-024-02457-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2023] [Accepted: 02/08/2024] [Indexed: 02/16/2024] Open
Abstract
BACKGROUND Unsupervised clustering and outlier detection are important in medical research to understand the distributional composition of a collective of patients. A number of clustering methods exist, also for high-dimensional data after dimension reduction. Clustering and outlier detection may, however, become less robust or contradictory if multiple high-dimensional data sets per patient exist. Such a scenario is given when the focus is on 3-D data of multiple organs per patient, and a high-dimensional feature matrix per organ is extracted. METHODS We use principal component analysis (PCA), t-distributed stochastic neighbor embedding (t-SNE) and multiple co-inertia analysis (MCIA) combined with bagplots to study the distribution of multi-organ 3-D data taken by computed tomography scans. After point-set registration of multiple organs from two public data sets, multiple hundred shape features are extracted per organ. While PCA and t-SNE can only be applied to each organ individually, MCIA can project the data of all organs into the same low-dimensional space. RESULTS MCIA is the only approach, here, with which data of all organs can be projected into the same low-dimensional space. We studied how frequently (i.e., by how many organs) a patient was classified to belong to the inner or outer 50% of the population, or as an outlier. Outliers could only be detected with MCIA and PCA. MCIA and t-SNE were more robust in judging the distributional location of a patient in contrast to PCA. CONCLUSIONS MCIA is more appropriate and robust in judging the distributional location of a patient in the case of multiple high-dimensional data sets per patient. It is still recommendable to apply PCA or t-SNE in parallel to MCIA to study the location of individual organs.
Collapse
Affiliation(s)
- Michael Selle
- Institute of Animal Genomics, University of Veterinary Medicine Hannover, Hannover, Germany.
| | - Magdalena Kircher
- Institute of Animal Genomics, University of Veterinary Medicine Hannover, Hannover, Germany
| | - Cornelia Schwennen
- Institute for Animal Nutrition, University of Veterinary Medicine Hannover, Hannover, Germany
| | - Christian Visscher
- Institute for Animal Nutrition, University of Veterinary Medicine Hannover, Hannover, Germany
| | - Klaus Jung
- Institute of Animal Genomics, University of Veterinary Medicine Hannover, Hannover, Germany.
| |
Collapse
|
15
|
Kim J, Li Y, Shin BS. Volumetric Imitation Generative Adversarial Networks for Anatomical Human Body Modeling. Bioengineering (Basel) 2024; 11:163. [PMID: 38391649 PMCID: PMC10886047 DOI: 10.3390/bioengineering11020163] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Revised: 02/02/2024] [Accepted: 02/06/2024] [Indexed: 02/24/2024] Open
Abstract
Volumetric representation is a technique used to express 3D objects in various fields, such as medical applications. On the other hand, tomography images for reconstructing volumetric data have limited utilization because they contain personal information. Existing GAN-based medical image generation techniques can produce virtual tomographic images for volume reconstruction while preserving the patient's privacy. Nevertheless, these images often do not consider vertical correlations between the adjacent slices, leading to erroneous results in 3D reconstruction. Furthermore, while volume generation techniques have been introduced, they often focus on surface modeling, making it challenging to represent the internal anatomical features accurately. This paper proposes volumetric imitation GAN (VI-GAN), which imitates a human anatomical model to generate volumetric data. The primary goal of this model is to capture the attributes and 3D structure, including the external shape, internal slices, and the relationship between the vertical slices of the human anatomical model. The proposed network consists of a generator for feature extraction and up-sampling based on a 3D U-Net and ResNet structure and a 3D-convolution-based LFFB (local feature fusion block). In addition, a discriminator utilizes 3D convolution to evaluate the authenticity of the generated volume compared to the ground truth. VI-GAN also devises reconstruction loss, including feature and similarity losses, to converge the generated volumetric data into a human anatomical model. In this experiment, the CT data of 234 people were used to assess the reliability of the results. When using volume evaluation metrics to measure similarity, VI-GAN generated a volume that realistically represented the human anatomical model compared to existing volume generation methods.
Collapse
Affiliation(s)
- Jion Kim
- Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
| | - Yan Li
- Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
| | - Byeong-Seok Shin
- Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
| |
Collapse
|
16
|
Liao W, Luo X, He Y, Dong Y, Li C, Li K, Zhang S, Zhang S, Wang G, Xiao J. Comprehensive Evaluation of a Deep Learning Model for Automatic Organs-at-Risk Segmentation on Heterogeneous Computed Tomography Images for Abdominal Radiation Therapy. Int J Radiat Oncol Biol Phys 2023; 117:994-1006. [PMID: 37244625 DOI: 10.1016/j.ijrobp.2023.05.034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2022] [Revised: 03/13/2023] [Accepted: 05/18/2023] [Indexed: 05/29/2023]
Abstract
PURPOSE Our purpose was to develop a deep learning model (AbsegNet) that produces accurate contours of 16 organs at risk (OARs) for abdominal malignancies as an essential part of fully automated radiation treatment planning. METHODS AND MATERIALS Three data sets with 544 computed tomography scans were retrospectively collected. Data set 1 was split into 300 training cases and 128 test cases (cohort 1) for AbsegNet. Data set 2, including cohort 2 (n = 24) and cohort 3 (n = 20), were used to validate AbsegNet externally. Data set 3, including cohort 4 (n = 40) and cohort 5 (n = 32), were used to clinically assess the accuracy of AbsegNet-generated contours. Each cohort was from a different center. The Dice similarity coefficient and 95th-percentile Hausdorff distance were calculated to evaluate the delineation quality for each OAR. Clinical accuracy evaluation was classified into 4 levels: no revision, minor revisions (0% < volumetric revision degrees [VRD] ≤ 10%), moderate revisions (10% ≤ VRD < 20%), and major revisions (VRD ≥20%). RESULTS For all OARs, AbsegNet achieved a mean Dice similarity coefficient of 86.73%, 85.65%, and 88.04% in cohorts 1, 2, and 3, respectively, and a mean 95th-percentile Hausdorff distance of 8.92, 10.18, and 12.40 mm, respectively. The performance of AbsegNet outperformed SwinUNETR, DeepLabV3+, Attention-UNet, UNet, and 3D-UNet. When experts evaluated contours from cohorts 4 and 5, 4 OARs (liver, kidney_L, kidney_R, and spleen) of all patients were scored as having no revision, and over 87.5% of patients with contours of the stomach, esophagus, adrenals, or rectum were considered as having no or minor revisions. Only 15.0% of patients with colon and small bowel contours required major revisions. CONCLUSIONS We propose a novel deep-learning model to delineate OARs on diverse data sets. Most contours produced by AbsegNet are accurate and robust and are, therefore, clinically applicable and helpful to facilitate radiation therapy workflow.
Collapse
Affiliation(s)
- Wenjun Liao
- Department of Radiation Oncology, Radiation Oncology Key Laboratory of Sichuan Province, Sichuan Clinical Research Center for Cancer, Sichuan Cancer Hospital & Institute, Sichuan Cancer Center, Affiliated Cancer Hospital of University of Electronic Science and Technology of China, Chengdu, China
| | - Xiangde Luo
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China; Shanghai AI Laboratory, Shanghai, China
| | - Yuan He
- Department of Radiation Oncology, The First Affiliated Hospital of USTC, Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, Anhui, China
| | - Ye Dong
- Department of NanFang PET Center, Nanfang Hospital, Southern Medical University, Guangzhou, China
| | - Churong Li
- Department of Radiation Oncology, Radiation Oncology Key Laboratory of Sichuan Province, Sichuan Clinical Research Center for Cancer, Sichuan Cancer Hospital & Institute, Sichuan Cancer Center, Affiliated Cancer Hospital of University of Electronic Science and Technology of China, Chengdu, China
| | - Kang Li
- West China Biomedical Big Data Center
| | - Shichuan Zhang
- Department of Radiation Oncology, Radiation Oncology Key Laboratory of Sichuan Province, Sichuan Clinical Research Center for Cancer, Sichuan Cancer Hospital & Institute, Sichuan Cancer Center, Affiliated Cancer Hospital of University of Electronic Science and Technology of China, Chengdu, China
| | - Shaoting Zhang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China; Shanghai AI Laboratory, Shanghai, China
| | - Guotai Wang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China; Shanghai AI Laboratory, Shanghai, China
| | - Jianghong Xiao
- Radiotherapy Physics & Technology Center, Department of Radiation Oncology, Cancer Center, West China Hospital, Sichuan University, Chengdu, China.
| |
Collapse
|
17
|
Li C, Bagher-Ebadian H, Sultan RI, Elshaikh M, Movsas B, Zhu D, Chetty IJ. A new architecture combining convolutional and transformer-based networks for automatic 3D multi-organ segmentation on CT images. Med Phys 2023; 50:6990-7002. [PMID: 37738468 DOI: 10.1002/mp.16750] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Revised: 08/08/2023] [Accepted: 08/13/2023] [Indexed: 09/24/2023] Open
Abstract
PURPOSE Deep learning-based networks have become increasingly popular in the field of medical image segmentation. The purpose of this research was to develop and optimize a new architecture for automatic segmentation of the prostate gland and normal organs in the pelvic, thoracic, and upper gastro-intestinal (GI) regions. METHODS We developed an architecture which combines a shifted-window (Swin) transformer with a convolutional U-Net. The network includes a parallel encoder, a cross-fusion block, and a CNN-based decoder to extract local and global information and merge related features on the same scale. A skip connection is applied between the cross-fusion block and decoder to integrate low-level semantic features. Attention gates (AGs) are integrated within the CNN to suppress features in image background regions. Our network is termed "SwinAttUNet." We optimized the architecture for automatic image segmentation. Training datasets consisted of planning-CT datasets from 300 prostate cancer patients from an institutional database and 100 CT datasets from a publicly available dataset (CT-ORG). Images were linearly interpolated and resampled to a spatial resolution of (1.0 × 1.0× 1.5) mm3 . A volume patch (192 × 192 × 96) was used for training and inference, and the dataset was split into training (75%), validation (10%), and test (15%) cohorts. Data augmentation transforms were applied consisting of random flip, rotation, and intensity scaling. The loss function comprised Dice and cross-entropy equally weighted and summed. We evaluated Dice coefficients (DSC), 95th percentile Hausdorff Distances (HD95), and Average Surface Distances (ASD) between results of our network and ground truth data. RESULTS SwinAttUNet, DSC values were 86.54 ± 1.21, 94.15 ± 1.17, and 87.15 ± 1.68% and HD95 values were 5.06 ± 1.42, 3.16 ± 0.93, and 5.54 ± 1.63 mm for the prostate, bladder, and rectum, respectively. Respective ASD values were 1.45 ± 0.57, 0.82 ± 0.12, and 1.42 ± 0.38 mm. For the lung, liver, kidneys and pelvic bones, respective DSC values were: 97.90 ± 0.80, 96.16 ± 0.76, 93.74 ± 2.25, and 89.31 ± 3.87%. Respective HD95 values were: 5.13 ± 4.11, 2.73 ± 1.19, 2.29 ± 1.47, and 5.31 ± 1.25 mm. Respective ASD values were: 1.88 ± 1.45, 1.78 ± 1.21, 0.71 ± 0.43, and 1.21 ± 1.11 mm. Our network outperformed several existing deep learning approaches using only attention-based convolutional or Transformer-based feature strategies, as detailed in the results section. CONCLUSIONS We have demonstrated that our new architecture combining Transformer- and convolution-based features is able to better learn the local and global context for automatic segmentation of multi-organ, CT-based anatomy.
Collapse
Affiliation(s)
- Chengyin Li
- College of Engineering - Dept. of Computer Science, Wayne State University, Detroit, Michigan, USA
| | - Hassan Bagher-Ebadian
- Department of Radiation Oncology, Henry Ford Cancer Institute, Detroit, Michigan, USA
- Department of Radiology, Michigan State University, East Lansing, Michigan, USA
- Department of Osteopathic Medicine, Michigan State University, East Lansing, Michigan, USA
- Department of Physics, Oakland University, Rochester, Michigan, USA
| | - Rafi Ibn Sultan
- College of Engineering - Dept. of Computer Science, Wayne State University, Detroit, Michigan, USA
| | - Mohamed Elshaikh
- Department of Radiation Oncology, Henry Ford Cancer Institute, Detroit, Michigan, USA
| | - Benjamin Movsas
- Department of Radiation Oncology, Henry Ford Cancer Institute, Detroit, Michigan, USA
| | - Dongxiao Zhu
- College of Engineering - Dept. of Computer Science, Wayne State University, Detroit, Michigan, USA
| | - Indrin J Chetty
- Department of Radiation Oncology, Henry Ford Cancer Institute, Detroit, Michigan, USA
- Department of Radiation Oncology, Cedars Sinai Medical Center, Los Angeles, CA, USA
| |
Collapse
|
18
|
Weisman AJ, Huff DT, Govindan RM, Chen S, Perk TG. Multi-organ segmentation of CT via convolutional neural network: impact of training setting and scanner manufacturer. Biomed Phys Eng Express 2023; 9:065021. [PMID: 37725928 DOI: 10.1088/2057-1976/acfb06] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Accepted: 09/19/2023] [Indexed: 09/21/2023]
Abstract
Objective. Automated organ segmentation on CT images can enable the clinical use of advanced quantitative software devices, but model performance sensitivities must be understood before widespread adoption can occur. The goal of this study was to investigate performance differences between Convolutional Neural Networks (CNNs) trained to segment one (single-class) versus multiple (multi-class) organs, and between CNNs trained on scans from a single manufacturer versus multiple manufacturers.Methods. The multi-class CNN was trained on CT images obtained from 455 whole-body PET/CT scans (413 for training, 42 for testing) taken with Siemens, GE, and Phillips PET/CT scanners where 16 organs were segmented. The multi-class CNN was compared to 16 smaller single-class CNNs trained using the same data, but with segmentations of only one organ per model. In addition, CNNs trained on Siemens-only (N = 186) and GE-only (N = 219) scans (manufacturer-specific) were compared with CNNs trained on data from both Siemens and GE scanners (manufacturer-mixed). Segmentation performance was quantified using five performance metrics, including the Dice Similarity Coefficient (DSC).Results. The multi-class CNN performed well compared to previous studies, even in organs usually considered difficult auto-segmentation targets (e.g., pancreas, bowel). Segmentations from the multi-class CNN were significantly superior to those from smaller single-class CNNs in most organs, and the 16 single-class models took, on average, six times longer to segment all 16 organs compared to the single multi-class model. The manufacturer-mixed approach achieved minimally higher performance over the manufacturer-specific approach.Significance. A CNN trained on contours of multiple organs and CT data from multiple manufacturers yielded high-quality segmentations. Such a model is an essential enabler of image processing in a software device that quantifies and analyzes such data to determine a patient's treatment response. To date, this activity of whole organ segmentation has not been adopted due to the intense manual workload and time required.
Collapse
Affiliation(s)
- Amy J Weisman
- AIQ Solutions, Madison, WI, United States of America
| | - Daniel T Huff
- AIQ Solutions, Madison, WI, United States of America
| | | | - Song Chen
- Department of Nuclear Medicine, The First Hospital of China Medical University, Shenyang, Liaoning, People's Republic of China
| | | |
Collapse
|
19
|
Mai DVC, Drami I, Pring ET, Gould LE, Lung P, Popuri K, Chow V, Beg MF, Athanasiou T, Jenkins JT. A systematic review of automated segmentation of 3D computed-tomography scans for volumetric body composition analysis. J Cachexia Sarcopenia Muscle 2023; 14:1973-1986. [PMID: 37562946 PMCID: PMC10570079 DOI: 10.1002/jcsm.13310] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Revised: 05/03/2023] [Accepted: 07/11/2023] [Indexed: 08/12/2023] Open
Abstract
Automated computed tomography (CT) scan segmentation (labelling of pixels according to tissue type) is now possible. This technique is being adapted to achieve three-dimensional (3D) segmentation of CT scans, opposed to single L3-slice alone. This systematic review evaluates feasibility and accuracy of automated segmentation of 3D CT scans for volumetric body composition (BC) analysis, as well as current limitations and pitfalls clinicians and researchers should be aware of. OVID Medline, Embase and grey literature databases up to October 2021 were searched. Original studies investigating automated skeletal muscle, visceral and subcutaneous AT segmentation from CT were included. Seven of the 92 studies met inclusion criteria. Variation existed in expertise and numbers of humans performing ground-truth segmentations used to train algorithms. There was heterogeneity in patient characteristics, pathology and CT phases that segmentation algorithms were developed upon. Reporting of anatomical CT coverage varied, with confusing terminology. Six studies covered volumetric regional slabs rather than the whole body. One study stated the use of whole-body CT, but it was not clear whether this truly meant head-to-fingertip-to-toe. Two studies used conventional computer algorithms. The latter five used deep learning (DL), an artificial intelligence technique where algorithms are similarly organized to brain neuronal pathways. Six of seven reported excellent segmentation performance (Dice similarity coefficients > 0.9 per tissue). Internal testing on unseen scans was performed for only four of seven algorithms, whilst only three were tested externally. Trained DL algorithms achieved full CT segmentation in 12 to 75 s versus 25 min for non-DL techniques. DL enables opportunistic, rapid and automated volumetric BC analysis of CT performed for clinical indications. However, most CT scans do not cover head-to-fingertip-to-toe; further research must validate using common CT regions to estimate true whole-body BC, with direct comparison to single lumbar slice. Due to successes of DL, we expect progressive numbers of algorithms to materialize in addition to the seven discussed in this paper. Researchers and clinicians in the field of BC must therefore be aware of pitfalls. High Dice similarity coefficients do not inform the degree to which BC tissues may be under- or overestimated and nor does it inform on algorithm precision. Consensus is needed to define accuracy and precision standards for ground-truth labelling. Creation of a large international, multicentre common CT dataset with BC ground-truth labels from multiple experts could be a robust solution.
Collapse
Affiliation(s)
- Dinh Van Chi Mai
- Department of SurgerySt Mark's Academic Institute, St Mark's HospitalLondonUK
- Department of Surgery and CancerImperial CollegeLondonUK
| | - Ioanna Drami
- Department of SurgerySt Mark's Academic Institute, St Mark's HospitalLondonUK
- Department of Metabolism, Digestion and ReproductionImperial CollegeLondonUK
| | - Edward T. Pring
- Department of SurgerySt Mark's Academic Institute, St Mark's HospitalLondonUK
- Department of Surgery and CancerImperial CollegeLondonUK
| | - Laura E. Gould
- Department of SurgerySt Mark's Academic Institute, St Mark's HospitalLondonUK
- School of Cancer Sciences, College of Medical, Veterinary & Life SciencesUniverstiy of GlasgowGlasgowUK
| | - Phillip Lung
- Department of SurgerySt Mark's Academic Institute, St Mark's HospitalLondonUK
- Department of Surgery and CancerImperial CollegeLondonUK
| | - Karteek Popuri
- Department of Computer ScienceMemorial University of NewfoundlandSt JohnsCanada
| | - Vincent Chow
- School of Engineering ScienceSimon Fraser UniversityBurnabyCanada
| | - Mirza F. Beg
- School of Engineering ScienceSimon Fraser UniversityBurnabyCanada
| | | | - John T. Jenkins
- Department of SurgerySt Mark's Academic Institute, St Mark's HospitalLondonUK
- Department of Surgery and CancerImperial CollegeLondonUK
| | | |
Collapse
|
20
|
Wasserthal J, Breit HC, Meyer MT, Pradella M, Hinck D, Sauter AW, Heye T, Boll DT, Cyriac J, Yang S, Bach M, Segeroth M. TotalSegmentator: Robust Segmentation of 104 Anatomic Structures in CT Images. Radiol Artif Intell 2023; 5:e230024. [PMID: 37795137 PMCID: PMC10546353 DOI: 10.1148/ryai.230024] [Citation(s) in RCA: 76] [Impact Index Per Article: 76.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2023] [Revised: 05/16/2023] [Accepted: 06/14/2023] [Indexed: 10/06/2023]
Abstract
Purpose To present a deep learning segmentation model that can automatically and robustly segment all major anatomic structures on body CT images. Materials and Methods In this retrospective study, 1204 CT examinations (from 2012, 2016, and 2020) were used to segment 104 anatomic structures (27 organs, 59 bones, 10 muscles, and eight vessels) relevant for use cases such as organ volumetry, disease characterization, and surgical or radiation therapy planning. The CT images were randomly sampled from routine clinical studies and thus represent a real-world dataset (different ages, abnormalities, scanners, body parts, sequences, and sites). The authors trained an nnU-Net segmentation algorithm on this dataset and calculated Dice similarity coefficients to evaluate the model's performance. The trained algorithm was applied to a second dataset of 4004 whole-body CT examinations to investigate age-dependent volume and attenuation changes. Results The proposed model showed a high Dice score (0.943) on the test set, which included a wide range of clinical data with major abnormalities. The model significantly outperformed another publicly available segmentation model on a separate dataset (Dice score, 0.932 vs 0.871; P < .001). The aging study demonstrated significant correlations between age and volume and mean attenuation for a variety of organ groups (eg, age and aortic volume [rs = 0.64; P < .001]; age and mean attenuation of the autochthonous dorsal musculature [rs = -0.74; P < .001]). Conclusion The developed model enables robust and accurate segmentation of 104 anatomic structures. The annotated dataset (https://doi.org/10.5281/zenodo.6802613) and toolkit (https://www.github.com/wasserth/TotalSegmentator) are publicly available.Keywords: CT, Segmentation, Neural Networks Supplemental material is available for this article. © RSNA, 2023See also commentary by Sebro and Mongan in this issue.
Collapse
Affiliation(s)
- Jakob Wasserthal
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Hanns-Christian Breit
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Manfred T. Meyer
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Maurice Pradella
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Daniel Hinck
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Alexander W. Sauter
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Tobias Heye
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Daniel T. Boll
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Joshy Cyriac
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Shan Yang
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Michael Bach
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| | - Martin Segeroth
- From the Clinic of Radiology and Nuclear Medicine, University Hospital Basel, Basel, Switzerland, Petersgraben 4, 4031 Basel, Switzerland
| |
Collapse
|
21
|
Kwee TC, Roest C, Kasalak Ö, Pennings JP, de Jong IJ, Yakar D. A new medical imaging postprocessing and interpretation concept to investigate the clinical relevance of incidentalomas: can we keep Pandora's box closed? Acta Radiol 2023; 64:2170-2179. [PMID: 37116890 DOI: 10.1177/02841851231158769] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/30/2023]
Abstract
BACKGROUND Incidental imaging findings (incidentalomas) are common, but there is currently no effective means to investigate their clinical relevance. PURPOSE To introduce a new concept to postprocess a medical imaging examination in a way that incidentalomas are concealed while its diagnostic potential is maintained to answer the referring physician's clinical questions. MATERIAL AND METHODS A deep learning algorithm was developed to automatically eliminate liver, gallbladder, pancreas, spleen, adrenal glands, lungs, and bone from unenhanced computed tomography (CT). This deep learning algorithm was applied to a separately held set of unenhanced CT scans of 27 patients who underwent CT to evaluate for urolithiasis, and who had a total of 32 incidentalomas in one of the aforementioned organs. RESULTS Median visual scores for organ elimination on modified CT were 100% for the liver, gallbladder, spleen, and right adrenal gland, 90%-99% for the pancreas, lungs, and bones, and 80%-89% for the left adrenal gland. In 26 out of 27 cases (96.3%), the renal calyces and pelves, ureters, and urinary bladder were completely visible on modified CT. In one case, a short (<1 cm) trajectory of the left ureter was not clearly visible due to adjacent atherosclerosis that was mistaken for bone by the algorithm. Of 32 incidentalomas, 28 (87.5%) were completely concealed on modified CT. CONCLUSION This preliminary technical report demonstrated the feasibility of a new approach to postprocess and evaluate medical imaging examinations that can be used by future prospective research studies with long-term follow-up to investigate the clinical relevance of incidentalomas.
Collapse
Affiliation(s)
- Thomas C Kwee
- Medical Imaging Center, Departments of Radiology, University of Groningen, University Medical Center Groningen, Groningen, the Netherlands
| | - Christian Roest
- Medical Imaging Center, Departments of Radiology, University of Groningen, University Medical Center Groningen, Groningen, the Netherlands
| | - Ömer Kasalak
- Medical Imaging Center, Departments of Radiology, University of Groningen, University Medical Center Groningen, Groningen, the Netherlands
| | - Jan P Pennings
- Medical Imaging Center, Departments of Radiology, University of Groningen, University Medical Center Groningen, Groningen, the Netherlands
| | - Igle Jan de Jong
- Department of Urology, University of Groningen, University Medical Center Groningen, Groningen, the Netherlands
| | - Derya Yakar
- Medical Imaging Center, Departments of Radiology, University of Groningen, University Medical Center Groningen, Groningen, the Netherlands
| |
Collapse
|
22
|
Role of Machine Learning-Based CT Body Composition in Risk Prediction and Prognostication: Current State and Future Directions. Diagnostics (Basel) 2023; 13:diagnostics13050968. [PMID: 36900112 PMCID: PMC10000509 DOI: 10.3390/diagnostics13050968] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Revised: 02/11/2023] [Accepted: 02/18/2023] [Indexed: 03/08/2023] Open
Abstract
CT body composition analysis has been shown to play an important role in predicting health and has the potential to improve patient outcomes if implemented clinically. Recent advances in artificial intelligence and machine learning have led to high speed and accuracy for extracting body composition metrics from CT scans. These may inform preoperative interventions and guide treatment planning. This review aims to discuss the clinical applications of CT body composition in clinical practice, as it moves towards widespread clinical implementation.
Collapse
|
23
|
Li H, Nan Y, Del Ser J, Yang G. Large-Kernel Attention for 3D Medical Image Segmentation. Cognit Comput 2023; 16:2063-2077. [PMID: 38974012 PMCID: PMC11226511 DOI: 10.1007/s12559-023-10126-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Accepted: 02/09/2023] [Indexed: 03/03/2023]
Abstract
Automated segmentation of multiple organs and tumors from 3D medical images such as magnetic resonance imaging (MRI) and computed tomography (CT) scans using deep learning methods can aid in diagnosing and treating cancer. However, organs often overlap and are complexly connected, characterized by extensive anatomical variation and low contrast. In addition, the diversity of tumor shape, location, and appearance, coupled with the dominance of background voxels, makes accurate 3D medical image segmentation difficult. In this paper, a novel 3D large-kernel (LK) attention module is proposed to address these problems to achieve accurate multi-organ segmentation and tumor segmentation. The advantages of biologically inspired self-attention and convolution are combined in the proposed LK attention module, including local contextual information, long-range dependencies, and channel adaptation. The module also decomposes the LK convolution to optimize the computational cost and can be easily incorporated into CNNs such as U-Net. Comprehensive ablation experiments demonstrated the feasibility of convolutional decomposition and explored the most efficient and effective network design. Among them, the best Mid-type 3D LK attention-based U-Net network was evaluated on CT-ORG and BraTS 2020 datasets, achieving state-of-the-art segmentation performance when compared to avant-garde CNN and Transformer-based methods for medical image segmentation. The performance improvement due to the proposed 3D LK attention module was statistically validated.
Collapse
Affiliation(s)
- Hao Li
- National Heart and Lung Institute, Faculty of Medicine, Imperial College London, London, UK
- Department of Bioengineering, Faculty of Engineering, Imperial College London, London, UK
| | - Yang Nan
- National Heart and Lung Institute, Faculty of Medicine, Imperial College London, London, UK
| | - Javier Del Ser
- TECNALIA, Basque Research & Technology Alliance (BRTA), Derio, Spain
- University of the Basque Country (UPV/EHU), Bilbao, Spain
| | - Guang Yang
- National Heart and Lung Institute, Faculty of Medicine, Imperial College London, London, UK
- Royal Brompton Hospital, London, UK
| |
Collapse
|
24
|
Somers P, Holdenried-Krafft S, Zahn J, Schüle J, Veil C, Harland N, Walz S, Stenzl A, Sawodny O, Tarín C, Lensch HPA. Cystoscopic depth estimation using gated adversarial domain adaptation. Biomed Eng Lett 2023; 13:141-151. [PMID: 37124116 PMCID: PMC10130294 DOI: 10.1007/s13534-023-00261-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2022] [Revised: 12/20/2022] [Accepted: 01/09/2023] [Indexed: 01/22/2023] Open
Abstract
AbstractMonocular depth estimation from camera images is very important for surrounding scene evaluation in many technical fields from automotive to medicine. However, traditional triangulation methods using stereo cameras or multiple views with the assumption of a rigid environment are not applicable for endoscopic domains. Particularly in cystoscopies it is not possible to produce ground truth depth information to directly train machine learning algorithms for using a monocular image directly for depth prediction. This work considers first creating a synthetic cystoscopic environment for initial encoding of depth information from synthetically rendered images. Next, the task of predicting pixel-wise depth values for real images is constrained to a domain adaption between the synthetic and real image domains. This adaptation is done through added gated residual blocks in order to simplify the network task and maintain training stability during adversarial training. Training is done on an internally collected cystoscopy dataset from human patients. The results after training demonstrate the ability to predict reasonable depth estimations from actual cystoscopic videos and added stability from using gated residual blocks is shown to prevent mode collapse during adversarial training.
Collapse
|
25
|
Huang SY, Hsu WL, Hsu RJ, Liu DW. Fully Convolutional Network for the Semantic Segmentation of Medical Images: A Survey. Diagnostics (Basel) 2022; 12:diagnostics12112765. [PMID: 36428824 PMCID: PMC9689961 DOI: 10.3390/diagnostics12112765] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2022] [Revised: 10/19/2022] [Accepted: 11/04/2022] [Indexed: 11/16/2022] Open
Abstract
There have been major developments in deep learning in computer vision since the 2010s. Deep learning has contributed to a wealth of data in medical image processing, and semantic segmentation is a salient technique in this field. This study retrospectively reviews recent studies on the application of deep learning for segmentation tasks in medical imaging and proposes potential directions for future development, including model development, data augmentation processing, and dataset creation. The strengths and deficiencies of studies on models and data augmentation, as well as their application to medical image segmentation, were analyzed. Fully convolutional network developments have led to the creation of the U-Net and its derivatives. Another noteworthy image segmentation model is DeepLab. Regarding data augmentation, due to the low data volume of medical images, most studies focus on means to increase the wealth of medical image data. Generative adversarial networks (GAN) increase data volume via deep learning. Despite the increasing types of medical image datasets, there is still a deficiency of datasets on specific problems, which should be improved moving forward. Considering the wealth of ongoing research on the application of deep learning processing to medical image segmentation, the data volume and practical clinical application problems must be addressed to ensure that the results are properly applied.
Collapse
Affiliation(s)
- Sheng-Yao Huang
- Institute of Medical Science, Tzu Chi University, Hualien 97071, Taiwan
- Department of Radiation Oncology, Hualien Tzu Chi General Hospital, Buddhist Tzu Chi Medical Foundation, Hualien 97071, Taiwan
| | - Wen-Lin Hsu
- Department of Radiation Oncology, Hualien Tzu Chi General Hospital, Buddhist Tzu Chi Medical Foundation, Hualien 97071, Taiwan
- Cancer Center, Hualien Tzu Chi Hospital, Buddhist Tzu Chi Medical Foundation, Hualien 97071, Taiwan
- School of Medicine, Tzu Chi University, Hualien 97071, Taiwan
| | - Ren-Jun Hsu
- Institute of Medical Science, Tzu Chi University, Hualien 97071, Taiwan
- Cancer Center, Hualien Tzu Chi Hospital, Buddhist Tzu Chi Medical Foundation, Hualien 97071, Taiwan
- School of Medicine, Tzu Chi University, Hualien 97071, Taiwan
- Correspondence: (R.-J.H.); (D.-W.L.); Tel. & Fax: +886-3-8561825 (R.-J.H. & D.-W.L.)
| | - Dai-Wei Liu
- Institute of Medical Science, Tzu Chi University, Hualien 97071, Taiwan
- Department of Radiation Oncology, Hualien Tzu Chi General Hospital, Buddhist Tzu Chi Medical Foundation, Hualien 97071, Taiwan
- Cancer Center, Hualien Tzu Chi Hospital, Buddhist Tzu Chi Medical Foundation, Hualien 97071, Taiwan
- School of Medicine, Tzu Chi University, Hualien 97071, Taiwan
- Correspondence: (R.-J.H.); (D.-W.L.); Tel. & Fax: +886-3-8561825 (R.-J.H. & D.-W.L.)
| |
Collapse
|
26
|
Pan S, Lei Y, Wang T, Wynne J, Chang CW, Roper J, Jani AB, Patel P, Bradley JD, Liu T, Yang X. Male pelvic multi-organ segmentation using token-based transformer Vnet. Phys Med Biol 2022; 67:10.1088/1361-6560/ac95f7. [PMID: 36170872 PMCID: PMC9671083 DOI: 10.1088/1361-6560/ac95f7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2022] [Accepted: 09/28/2022] [Indexed: 11/12/2022]
Abstract
Objective. This work aims to develop an automated segmentation method for the prostate and its surrounding organs-at-risk in pelvic computed tomography to facilitate prostate radiation treatment planning.Approach. In this work, we propose a novel deep learning algorithm combining a U-shaped convolutional neural network (CNN) and vision transformer (VIT) for multi-organ (i.e. bladder, prostate, rectum, left and right femoral heads) segmentation in male pelvic CT images. The U-shaped model consists of three components: a CNN-based encoder for local feature extraction, a token-based VIT for capturing global dependencies from the CNN features, and a CNN-based decoder for predicting the segmentation outcome from the VIT's output. The novelty of our network is a token-based multi-head self-attention mechanism used in the transformer, which encourages long-range dependencies and forwards informative high-resolution feature maps from the encoder to the decoder. In addition, a knowledge distillation strategy is deployed to further enhance the learning capability of the proposed network.Main results. We evaluated the network using: (1) a dataset collected from 94 patients with prostate cancer; (2) and a public dataset CT-ORG. A quantitative evaluation of the proposed network's performance was performed on each organ based on (1) volume similarity between the segmented contours and ground truth using Dice score, segmentation sensitivity, and precision, (2) surface similarity evaluated by Hausdorff distance (HD), mean surface distance (MSD) and residual mean square distance (RMS), (3) and percentage volume difference (PVD). The performance was then compared against other state-of-art methods. Average volume similarity measures obtained by the network overall organs were Dice score = 0.91, sensitivity = 0.90, precision = 0.92, average surface similarities were HD = 3.78 mm, MSD = 1.24 mm, RMS = 2.03 mm; average percentage volume difference was PVD = 9.9% on the first dataset. The network also obtained Dice score = 0.93, sensitivity = 0.93, precision = 0.93, average surface similarities were HD = 5.82 mm, MSD = 1.16 mm, RMS = 1.24 mm; average percentage volume difference was PVD = 6.6% on the CT-ORG dataset.Significance. In summary, we propose a token-based transformer network with knowledge distillation for multi-organ segmentation using CT images. This method provides accurate and reliable segmentation results for each organ using CT imaging, facilitating the prostate radiation clinical workflow.
Collapse
Affiliation(s)
- Shaoyan Pan
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Yang Lei
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Tonghe Wang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Jacob Wynne
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Chih-Wei Chang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Justin Roper
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Ashesh B Jani
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Pretesh Patel
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Jeffrey D Bradley
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Tian Liu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
- Department of Biomedical Informatics, Emory University, Atlanta, GA 30322, United States of America
| |
Collapse
|
27
|
WORD: A large scale dataset, benchmark and clinical applicable study for abdominal organ segmentation from CT image. Med Image Anal 2022; 82:102642. [DOI: 10.1016/j.media.2022.102642] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2022] [Revised: 08/18/2022] [Accepted: 09/20/2022] [Indexed: 11/22/2022]
|
28
|
Berzoini R, Colombo AA, Bardini S, Conelli A, D'Arnese E, Santambrogio MD. An Optimized U-Net for Unbalanced Multi-Organ Segmentation. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:3764-3767. [PMID: 36085901 DOI: 10.1109/embc48229.2022.9871288] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Medical practice is shifting towards the automation and standardization of the most repetitive procedures to speed up the time-to-diagnosis. Semantic segmentation repre-sents a critical stage in identifying a broad spectrum of regions of interest within medical images. Indeed, it identifies relevant objects by attributing to each image pixels a value representing pre-determined classes. Despite the relative ease of visually locating organs in the human body, automated multi-organ segmentation is hindered by the variety of shapes and dimensions of organs and computational resources. Within this context, we propose BIONET, a U-Net-based Fully Convolutional Net-work for efficiently semantically segmenting abdominal organs. BIONET deals with unbalanced data distribution related to the physiological conformation of the considered organs, reaching good accuracy for variable organs dimension with low variance, and a Weighted Global Dice Score score of 93.74 ± 1.1%, and an inference performance of 138 frames per second. Clinical Relevance - This work established a starting point for developing an automatic tool for semantic segmentation of variable-sized organs within the abdomen, reaching considerable accuracy on small and large organs with low variability, reaching a 93.74 ± 1.1 % of Weighted Global Dice Score.
Collapse
|
29
|
Predicting narrow ureters before ureteroscopic lithotripsy with a neural network: a retrospective bicenter study. Urolithiasis 2022; 50:599-610. [PMID: 35737011 DOI: 10.1007/s00240-022-01341-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2022] [Accepted: 05/30/2022] [Indexed: 10/17/2022]
Abstract
In some patients, the passage of semi-rigid ureteroscopes up the ureter is impossible due to narrow ureteral lumen. We established a neural network to predict the inability of the ureter to accommodate the semi-rigid ureteroscope and the need for active or passive dilatation using non-contrast computed tomography (CT) images. Data were collected retrospectively from two centers of 1989 eligible patients who underwent ureteroscopic lithotripsy with ureteral stones. Patients were categorized into two groups: control and narrow ureter. The network was designed and trained for predicting a narrow ureter during initial ureteroscopic lithotripsy, which integrated multi-scale features of the ureter. The predictive efficacy of neural networks DenseNet3D, ResNet3D, ResNet3D MC, and TimeSformer was compared. Furthermore, a previous ureteroscopy or a history of double-J stent placement, ureteral wall thickness and Hounsfield unit (HU) density of the ureter under the stone were compared. Model performance was assessed based on the accuracy, area under the receiver operating characteristic curve (AUC ROC), etc. The DenseNet3D-based network achieved an AUC ROC score of 0.884 and an accuracy of 85.29%, followed by the ResNet3D-based network, the ResNet3D MC-based network, and the TimeSformer-based network. The DenseNet3D-based network significantly outperformed other candidate predictors. Furthermore, the networks were validated in an external test set. Decision curve analysis showed the clinical utility of the neural network. The neural network provides an individualized preoperative prediction of narrow ureter based on non-contrast CT images, which could be employed as part of a surgical decision-making support system.
Collapse
|
30
|
Cong W, Li M, Guo X, Wang G. Estimating optical parameters of biological tissues with photon-counting micro-CT. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2022; 39:841-846. [PMID: 36215445 PMCID: PMC9552592 DOI: 10.1364/josaa.451319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/16/2021] [Accepted: 03/16/2022] [Indexed: 06/16/2023]
Abstract
Wavelength-dependent absorption and scattering properties determine the fluorescence photon transport in biological tissues and image resolution of optical molecular tomography. Currently, these parameters are computed from optically measured data. For small animal imaging, estimation of optical parameters is a large-scale optimization problem, which is highly ill-posed. In this paper, we propose a new, to the best of our knowledge, approach to estimate optical parameters of biological tissues with photon-counting micro-computed tomography (micro-CT). From photon-counting x-ray data, multi-energy micro-CT images can be reconstructed to perform multi-organ segmentation and material decomposition in terms of tissue constituents. The concentration and characteristics of major tissue constituents can be utilized to calculate the optical absorption and scattering coefficients of the involved tissues. In our study, we perform numerical simulation, phantom experiments, and in vivo animal studies to calculate the optical parameters using our proposed approach. The results show that our approach can estimate optical parameters of tissues with a relative error of <10%, accurately mapping the optical parameter distributions in a small animal.
Collapse
|
31
|
Looe HK, Blum I, Schönfeld AB, Tekin T, Delfs B, Poppe B. Model-based machine learning for the recovery of lateral dose profiles of small photon fields in magnetic field. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac5bfa] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2021] [Accepted: 03/09/2022] [Indexed: 11/12/2022]
Abstract
Abstract
Objective. To investigate the feasibility to train artificial neural networks (NN) to recover lateral dose profiles from detector measurements in a magnetic field. Approach. A novel framework based on a mathematical convolution model has been proposed to generate measurement-less training dataset. 2D dose deposition kernels and detector lateral fluence response functions of two air-filled ionization chambers and two diode-type detectors have been simulated without magnetic field and for magnetic field B = 0.35 and 1.5 T. Using these convolution kernels, training dataset consisting pairs of dose profiles
D
x
,
y
and signal profiles
M
x
,
y
were computed for a total of 108 2D photon fluence profiles
ψ
(
x
,
y
)
(80% training/20% validation). The NN were tested using three independent datasets, where the second test dataset has been obtained from simulations using realistic phase space files of clinical linear accelerator and the third test dataset was measured at a conventional linac equipped with electromagnets. Main results. The convolution kernels show magnetic field dependence due to the influence of the Lorentz force on the electron transport in the water phantom and detectors. The NN show good performance during training and validation with mean square error reaching a value of 1e-6 or smaller. The corresponding correlation coefficients R reached the value of 1 for all models indicating an excellent agreement between expected
D
x
,
y
and predicted
D
pred
x
,
y
.
The comparisons between
D
x
,
y
and
D
pred
x
,
y
using the three test datasets resulted in gamma indices (1 mm/1% global) <1 for all evaluated data points. Significance. Two verification approaches have been proposed to warrant the mathematical consistencies of the NN outputs. Besides offering a correction strategy not existed so far for relative dosimetry in a magnetic field, this work could help to raise awareness and to improve understanding on the distortion of detector’s signal profiles by a magnetic field.
Collapse
|
32
|
Deep Learning and Domain-Specific Knowledge to Segment the Liver from Synthetic Dual Energy CT Iodine Scans. Diagnostics (Basel) 2022; 12:diagnostics12030672. [PMID: 35328225 PMCID: PMC8947702 DOI: 10.3390/diagnostics12030672] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2022] [Revised: 02/24/2022] [Accepted: 03/03/2022] [Indexed: 11/23/2022] Open
Abstract
We map single energy CT (SECT) scans to synthetic dual-energy CT (synth-DECT) material density iodine (MDI) scans using deep learning (DL) and demonstrate their value for liver segmentation. A 2D pix2pix (P2P) network was trained on 100 abdominal DECT scans to infer synth-DECT MDI scans from SECT scans. The source and target domain were paired with DECT monochromatic 70 keV and MDI scans. The trained P2P algorithm then transformed 140 public SECT scans to synth-DECT scans. We split 131 scans into 60% train, 20% tune, and 20% held-out test to train four existing liver segmentation frameworks. The remaining nine low-dose SECT scans tested system generalization. Segmentation accuracy was measured with the dice coefficient (DSC). The DSC per slice was computed to identify sources of error. With synth-DECT (and SECT) scans, an average DSC score of 0.93±0.06 (0.89±0.01) and 0.89±0.01 (0.81±0.02) was achieved on the held-out and generalization test sets. Synth-DECT-trained systems required less data to perform as well as SECT-trained systems. Low DSC scores were primarily observed around the scan margin or due to non-liver tissue or distortions within ground-truth annotations. In general, training with synth-DECT scans resulted in improved segmentation performance with less data.
Collapse
|
33
|
Aiello M, Baldi D, Esposito G, Valentino M, Randon M, Salvatore M, Cavaliere C. Evaluation of AI-Based Segmentation Tools for COVID-19 Lung Lesions on Conventional and Ultra-low Dose CT Scans. Dose Response 2022; 20:15593258221082896. [PMID: 35422680 PMCID: PMC9002358 DOI: 10.1177/15593258221082896] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2021] [Accepted: 02/04/2022] [Indexed: 11/16/2022] Open
Abstract
A reliable diagnosis and accurate monitoring are pivotal steps for treatment and prevention of COVID-19. Chest computed tomography (CT) has been considered a crucial diagnostic imaging technique for the injury assessment of the viral pneumonia. Furthermore, the automatization of the segmentation methods for lung alterations helps to speed up the diagnosis and lighten radiologists' workload. Considering the assiduous pathology monitoring, ultra-low dose (ULD) chest CT protocols have been implemented to drastically reduce the radiation burden. Unfortunately, the available AI technologies have not been trained on ULD-CT data and validated and their applicability deserves careful evaluation. Therefore, this work aims to compare the results of available AI tools (BCUnet, CORADS AI, NVIDIA CLARA Train SDK and CT Pneumonia Analysis) on a dataset of 73 CT examinations acquired both with conventional dose (CD) and ULD protocols. COVID-19 volume percentage, resulting from each tool, was statistically compared. This study demonstrated high comparability of the results on CD-CT and ULD-CT data among the four AI tools, with high correlation between the results obtained on both protocols (R > .68, P < .001, for all AI tools).
Collapse
Affiliation(s)
| | | | | | - Marika Valentino
- Istituto di Scienze Applicate e
Sistemi Intelligenti “Eduardo Caianiello” (ISASI-CNR), Pozzuoli, Italy
- Università Degli Studi di Napoli
Federico II, Dip. di Ingegneria Elettrica e Delle Tecnologie
Dell'Informazione, Italy
| | | | | | | |
Collapse
|
34
|
Paravastu SS, Hasani N, Farhadi F, Collins MT, Edenbrandt L, Summers RM, Saboury B. Applications of Artificial Intelligence in 18F-Sodium Fluoride Positron Emission Tomography/Computed Tomography:: Current State and Future Directions. PET Clin 2021; 17:115-135. [PMID: 34809861 DOI: 10.1016/j.cpet.2021.09.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
This review discusses the current state of artificial intelligence (AI) in 18F-NaF-PET/CT imaging and the potential applications to come in diagnosis, prognostication, and improvement of care in patients with bone diseases, with emphasis on the role of AI algorithms in CT bone segmentation, relying on their prevalence in medical imaging and utility in the extraction of spatial information in combined PET/CT studies.
Collapse
Affiliation(s)
- Sriram S Paravastu
- Department of Radiology and Imaging Sciences, Clinical Center, National Institutes of Health (NIH), 9000 Rockville Pike, Building 10, Room 1C455, Bethesda, MD 20892, USA; Skeletal Disorders and Mineral Homeostasis Section, National Institute of Dental and Craniofacial Research, National Institutes of Health (NIH), 30 Convent Dr., Building 30, Room 228 MSC 4320, Bethesda, MD 20892, USA
| | - Navid Hasani
- Department of Radiology and Imaging Sciences, Clinical Center, National Institutes of Health (NIH), 9000 Rockville Pike, Building 10, Room 1C455, Bethesda, MD 20892, USA; University of Queensland Faculty of Medicine, Ochsner Clinical School, New Orleans, LA 70121, USA
| | - Faraz Farhadi
- Department of Radiology and Imaging Sciences, Clinical Center, National Institutes of Health (NIH), 9000 Rockville Pike, Building 10, Room 1C455, Bethesda, MD 20892, USA; Geisel School of Medicine at Dartmouth, Hanover, NH 03755, USA
| | - Michael T Collins
- Skeletal Disorders and Mineral Homeostasis Section, National Institute of Dental and Craniofacial Research, National Institutes of Health (NIH), 30 Convent Dr., Building 30, Room 228 MSC 4320, Bethesda, MD 20892, USA
| | - Lars Edenbrandt
- Department of Clinical Physiology, Sahlgrenska University Hospital, Göteborg, Sweden
| | - Ronald M Summers
- Department of Radiology and Imaging Sciences, Clinical Center, National Institutes of Health (NIH), 9000 Rockville Pike, Building 10, Room 1C455, Bethesda, MD 20892, USA
| | - Babak Saboury
- Department of Radiology and Imaging Sciences, Clinical Center, National Institutes of Health (NIH), 9000 Rockville Pike, Building 10, Room 1C455, Bethesda, MD 20892, USA; Department of Computer Science and Electrical Engineering, University of Maryland- Baltimore County, Baltimore, MD, USA; Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, USA.
| |
Collapse
|
35
|
Yousefirizi F, Jha AK, Brosch-Lenz J, Saboury B, Rahmim A. Toward High-Throughput Artificial Intelligence-Based Segmentation in Oncological PET Imaging. PET Clin 2021; 16:577-596. [PMID: 34537131 DOI: 10.1016/j.cpet.2021.06.001] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/15/2022]
Abstract
Artificial intelligence (AI) techniques for image-based segmentation have garnered much attention in recent years. Convolutional neural networks have shown impressive results and potential toward fully automated segmentation in medical imaging, and particularly PET imaging. To cope with the limited access to annotated data needed in supervised AI methods, given tedious and prone-to-error manual delineations, semi-supervised and unsupervised AI techniques have also been explored for segmentation of tumors or normal organs in single- and bimodality scans. This work reviews existing AI techniques for segmentation tasks and the evaluation criteria for translational AI-based segmentation efforts toward routine adoption in clinical workflows.
Collapse
Affiliation(s)
- Fereshteh Yousefirizi
- Department of Integrative Oncology, BC Cancer Research Institute, 675 West 10th Avenue, Vancouver, British Columbia V5Z 1L3, Canada.
| | - Abhinav K Jha
- Department of Biomedical Engineering, Washington University in St. Louis, St Louis, MO 63130, USA; Mallinckrodt Institute of Radiology, Washington University School of Medicine, St Louis, MO 63110, USA
| | - Julia Brosch-Lenz
- Department of Integrative Oncology, BC Cancer Research Institute, 675 West 10th Avenue, Vancouver, British Columbia V5Z 1L3, Canada
| | - Babak Saboury
- Department of Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, 9000 Rockville Pike, Bethesda, MD 20892, USA; Department of Computer Science and Electrical Engineering, University of Maryland Baltimore County, Baltimore, MD, USA; Department of Radiology, Hospital of the University of Pennsylvania, 3400 Spruce Street, Philadelphia, PA 19104, USA
| | - Arman Rahmim
- Department of Radiology, University of British Columbia, BC Cancer, BC Cancer Research Institute, 675 West 10th Avenue, Office 6-112, Vancouver, British Columbia V5Z 1L3, Canada; Department of Physics, University of British Columbia, Senior Scientist & Provincial Medical Imaging Physicist, BC Cancer, BC Cancer Research Institute, 675 West 10th Avenue, Office 6-112, Vancouver, British Columbia V5Z 1L3, Canada
| |
Collapse
|
36
|
Going to Extremes: Weakly Supervised Medical Image Segmentation. MACHINE LEARNING AND KNOWLEDGE EXTRACTION 2021. [DOI: 10.3390/make3020026] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Medical image annotation is a major hurdle for developing precise and robust machine-learning models. Annotation is expensive, time-consuming, and often requires expert knowledge, particularly in the medical field. Here, we suggest using minimal user interaction in the form of extreme point clicks to train a segmentation model which, in effect, can be used to speed up medical image annotation. An initial segmentation is generated based on the extreme points using the random walker algorithm. This initial segmentation is then used as a noisy supervision signal to train a fully convolutional network that can segment the organ of interest, based on the provided user clicks. Through experimentation on several medical imaging datasets, we show that the predictions of the network can be refined using several rounds of training with the prediction from the same weakly annotated data. Further improvements are shown using the clicked points within a custom-designed loss and attention mechanism. Our approach has the potential to speed up the process of generating new training datasets for the development of new machine-learning and deep-learning-based models for, but not exclusively, medical image analysis.
Collapse
|
37
|
Cadrin-Chênevert A. Toward a More Quantitative and Specific Representation of Normality. Radiol Artif Intell 2021; 3:e210005. [PMID: 33939776 PMCID: PMC8035574 DOI: 10.1148/ryai.2021210005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2021] [Revised: 01/08/2021] [Accepted: 01/11/2021] [Indexed: 11/11/2022]
Affiliation(s)
- Alexandre Cadrin-Chênevert
- From the Department of Medical Imaging, CISSS Lanaudière, affiliated with Laval University, 1000 Blvd St-Anne, Saint-Charles-Borromée, QC, Canada J6E 6J2
| |
Collapse
|