1
|
Zhou Z, Gong H, Hsieh S, McCollough CH, Yu L. Image quality evaluation in deep-learning-based CT noise reduction using virtual imaging trial methods: Contrast-dependent spatial resolution. Med Phys 2024; 51:5399-5413. [PMID: 38555876 PMCID: PMC11321944 DOI: 10.1002/mp.17029] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 02/19/2024] [Accepted: 02/26/2024] [Indexed: 04/02/2024] Open
Abstract
BACKGROUND Deep-learning-based image reconstruction and noise reduction methods (DLIR) have been increasingly deployed in clinical CT. Accurate image quality assessment of these methods is challenging as the performance measured using physical phantoms may not represent the true performance of DLIR in patients since DLIR is trained mostly on patient images. PURPOSE In this work, we aim to develop a patient-data-based virtual imaging trial framework and, as a first application, use it to measure the spatial resolution properties of a DLIR method. METHODS The patient-data-based virtual imaging trial framework consists of five steps: (1) insertion of lesions into projection domain data using the acquisition geometry of the patient exam to simulate different lesion characteristics; (2) insertion of noise into projection domain data using a realistic photon statistical model of the CT system to simulate different dose levels; (3) creation of DLIR-processed images from projection or image data; (4) creation of ensembles of DLIR-processed patient images from a large number of noise and lesion realizations; and (5) evaluation of image quality using ensemble DLIR images. This framework was applied to measure the spatial resolution of a ResNet based deep convolutional neural network (DCNN) trained on patient images. Lesions in a cylindrical shape and different contrast levels (-500, -100, -50, -20, -10 HU) were inserted to the lower right lobe of the liver in a patient case. Multiple dose levels were simulated (50%, 25%, 12.5%). Each lesion and dose condition had 600 noise realizations. Multiple reconstruction and denoising methods were used on all the noise realizations, including the original filtered-backprojection (FBP), iterative reconstruction (IR), and the DCNN method with three different strength setting (DCNN-weak, DCNN-medium, and DCNN-strong). Mean lesion signal was calculated by performing ensemble averaging of all the noise realizations for each lesion and dose condition and then subtracting the lesion-present images from the lesion absent images. Modulation transfer functions (MTFs) both in-plane and along the z-axis were calculated based on the mean lesion signals. The standard deviations of MTFs at each condition were estimated with bootstrapping: randomly sampling (with replacement) all the DLIR/FBP/IR images from the ensemble data (600 samples) at each condition. The impact of varying lesion contrast, dose levels, and denoising strengths were evaluated. Statistical analysis with paired t-test was used to compare the z-axis and in-plane spatial resolution of five algorithms for five different contrasts and three dose levels. RESULTS The in-plane and z-axis spatial resolution degradation of DCNN becomes more severe as the contrast or radiation dose decreased, or DCNN denoising strength increased. In comparison with FBP, a 59.5% and 4.1% reduction of in-plane and z-axis MTF (in terms of spatial frequencies at 50% MTF), respectively, was observed at low contrast (-10 HU) for DCNN with the highest denoising strength at 25% routine dose level. When the dose level reduces from 50% to 12.5% of routine dose, the in-plane and z-axis MTFs reduces from 92.1% to 76.3%, and from 98.9% to 95.5%, respectively, at contrast of -100 HU, using FBP as the reference. For most conditions of contrasts and dose levels, significant differences were found among the five algorithms, with the following relationship in both in-plane and cross-plane spatial resolution: FBP > DCNN-Weak > IR > DCNN-Medium > DCNN-Strong. The spatial resolution difference among algorithms decreases at higher contrast or dose levels. CONCLUSIONS A patient-data-based virtual imaging trial framework was developed and applied to measuring the spatial resolution properties of a DCNN noise reduction method at different contrast and dose levels using real patient data. As with other non-linear image reconstruction and post-processing techniques, the evaluated DCNN method degraded the in-plane and z-axis spatial resolution at lower contrast levels, lower radiation dose, and higher denoising strength.
Collapse
Affiliation(s)
| | - Hao Gong
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| | - Scott Hsieh
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| | | | - Lifeng Yu
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| |
Collapse
|
2
|
Zhu B, Yang Y. Quality assessment of abdominal CT images: an improved ResNet algorithm with dual-attention mechanism. Am J Transl Res 2024; 16:3099-3107. [PMID: 39114678 PMCID: PMC11301486 DOI: 10.62347/wkns8633] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 05/19/2024] [Indexed: 08/10/2024]
Abstract
OBJECTIVES To enhance medical image classification using a Dual-attention ResNet model and investigate the impact of attention mechanisms on model performance in a clinical setting. METHODS We utilized a dataset of medical images and implemented a Dual-attention ResNet model, integrating self-attention and spatial attention mechanisms. The model was trained and evaluated using binary and five-level quality classification tasks, leveraging standard evaluation metrics. RESULTS Our findings demonstrated substantial performance improvements with the Dual-attention ResNet model in both classification tasks. In the binary classification task, the model achieved an accuracy of 0.940, outperforming the conventional ResNet model. Similarly, in the five-level quality classification task, the Dual-attention ResNet model attained an accuracy of 0.757, highlighting its efficacy in capturing nuanced distinctions in image quality. CONCLUSIONS The integration of attention mechanisms within the ResNet model resulted in significant performance enhancements, showcasing its potential for improving medical image classification tasks. These results underscore the promising role of attention mechanisms in facilitating more accurate and discriminative analysis of medical images, thus holding substantial promise for clinical applications in radiology and diagnostics.
Collapse
Affiliation(s)
- Boying Zhu
- Shanghai Institute of Technical Physics, Chinese Academy of SciencesShanghai 200083, China
- University of Chinese Academy of SciencesBeijing 100049, China
| | - Yuanyuan Yang
- Shanghai Institute of Technical Physics, Chinese Academy of SciencesShanghai 200083, China
- University of Chinese Academy of SciencesBeijing 100049, China
| |
Collapse
|
3
|
Mukherjee S, Korfiatis P, Khasawneh H, Rajamohan N, Patra A, Suman G, Singh A, Thakkar J, Patnam NG, Trivedi KH, Karbhari A, Chari ST, Truty MJ, Halfdanarson TR, Bolan CW, Sandrasegaran K, Majumder S, Goenka AH. Bounding box-based 3D AI model for user-guided volumetric segmentation of pancreatic ductal adenocarcinoma on standard-of-care CTs. Pancreatology 2023; 23:522-529. [PMID: 37296006 PMCID: PMC10676442 DOI: 10.1016/j.pan.2023.05.008] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Revised: 05/19/2023] [Accepted: 05/20/2023] [Indexed: 06/12/2023]
Abstract
OBJECTIVES To develop a bounding-box-based 3D convolutional neural network (CNN) for user-guided volumetric pancreas ductal adenocarcinoma (PDA) segmentation. METHODS Reference segmentations were obtained on CTs (2006-2020) of treatment-naïve PDA. Images were algorithmically cropped using a tumor-centered bounding box for training a 3D nnUNet-based-CNN. Three radiologists independently segmented tumors on test subset, which were combined with reference segmentations using STAPLE to derive composite segmentations. Generalizability was evaluated on Cancer Imaging Archive (TCIA) (n = 41) and Medical Segmentation Decathlon (MSD) (n = 152) datasets. RESULTS Total 1151 patients [667 males; age:65.3 ± 10.2 years; T1:34, T2:477, T3:237, T4:403; mean (range) tumor diameter:4.34 (1.1-12.6)-cm] were randomly divided between training/validation (n = 921) and test subsets (n = 230; 75% from other institutions). Model had a high DSC (mean ± SD) against reference segmentations (0.84 ± 0.06), which was comparable to its DSC against composite segmentations (0.84 ± 0.11, p = 0.52). Model-predicted versus reference tumor volumes were comparable (mean ± SD) (29.1 ± 42.2-cc versus 27.1 ± 32.9-cc, p = 0.69, CCC = 0.93). Inter-reader variability was high (mean DSC 0.69 ± 0.16), especially for smaller and isodense tumors. Conversely, model's high performance was comparable between tumor stages, volumes and densities (p > 0.05). Model was resilient to different tumor locations, status of pancreatic/biliary ducts, pancreatic atrophy, CT vendors and slice thicknesses, as well as to the epicenter and dimensions of the bounding-box (p > 0.05). Performance was generalizable on MSD (DSC:0.82 ± 0.06) and TCIA datasets (DSC:0.84 ± 0.08). CONCLUSION A computationally efficient bounding box-based AI model developed on a large and diverse dataset shows high accuracy, generalizability, and robustness to clinically encountered variations for user-guided volumetric PDA segmentation including for small and isodense tumors. CLINICAL RELEVANCE AI-driven bounding box-based user-guided PDA segmentation offers a discovery tool for image-based multi-omics models for applications such as risk-stratification, treatment response assessment, and prognostication, which are urgently needed to customize treatment strategies to the unique biological profile of each patient's tumor.
Collapse
Affiliation(s)
- Sovanlal Mukherjee
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Panagiotis Korfiatis
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Hala Khasawneh
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Naveen Rajamohan
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Anurima Patra
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Garima Suman
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Aparna Singh
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Jay Thakkar
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Nandakumar G Patnam
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Kamaxi H Trivedi
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Aashna Karbhari
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Suresh T Chari
- Department of Gastroenterology, Hepatology and Nutrition, The University of Texas MD Anderson Cancer Center, 1515 Holcombe Blvd, Houston, TX, 77030, USA.
| | - Mark J Truty
- Department of Surgery, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | | | - Candice W Bolan
- Department of Radiology, Mayo Clinic, 4500 San Pablo Rd S, Jacksonville, FL, 32224, USA.
| | - Kumar Sandrasegaran
- Department of Radiology, Mayo Clinic, 13400 E Shea Blvd, Scottsdale, AZ, 85259, USA.
| | - Shounak Majumder
- Department of Gastroenterology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| | - Ajit H Goenka
- Department of Radiology, Mayo Clinic, 200 First Street SW, Rochester, MN, 55905, USA.
| |
Collapse
|
4
|
Zhou Z, Inoue A, McCollough CH, Yu L. Self-trained deep convolutional neural network for noise reduction in CT. J Med Imaging (Bellingham) 2023; 10:044008. [PMID: 37636895 PMCID: PMC10449263 DOI: 10.1117/1.jmi.10.4.044008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2023] [Revised: 08/04/2023] [Accepted: 08/08/2023] [Indexed: 08/29/2023] Open
Abstract
Purpose Supervised deep convolutional neural network (CNN)-based methods have been actively used in clinical CT to reduce image noise. The networks of these methods are typically trained using paired high- and low-quality data from a massive number of patients and/or phantom images. This training process is tedious, and the network trained under a given condition may not be generalizable to patient images acquired and reconstructed under different conditions. We propose a self-trained deep CNN (ST_CNN) method for noise reduction in CT that does not rely on pre-existing training datasets. Approach The ST_CNN training was accomplished using extensive data augmentation in the projection domain, and the inference was applied to the data itself. Specifically, multiple independent noise insertions were applied to the original patient projection data to generate multiple realizations of low-quality projection data. Then, rotation augmentation was adopted for both the original and low-quality projection data by applying the rotation angle directly on the projection data so that images were rotated at arbitrary angles without introducing additional bias. A large number of paired low- and high-quality images from the same patient were reconstructed and paired for training the ST_CNN model. Results No significant difference was found between the ST_CNN and conventional CNN models in terms of the peak signal-to-noise ratio and structural similarity index measure. The ST_CNN model outperformed the conventional CNN model in terms of noise texture and homogeneity in liver parenchyma as well as better subjective visualization of liver lesions. The ST_CNN may sacrifice the sharpness of vessels slightly compared to the conventional CNN model but without affecting the visibility of peripheral vessels or diagnosis of vascular pathology. Conclusions The proposed ST_CNN method trained from the data itself may achieve similar image quality in comparison with conventional deep CNN denoising methods pre-trained on external datasets.
Collapse
Affiliation(s)
- Zhongxing Zhou
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | - Akitoshi Inoue
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | | | - Lifeng Yu
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| |
Collapse
|