1
|
Lee S, Jung JH, Choi Y, Seok E, Jung J, Lim H, Kim D, Yun M. Cross-Modality Image Translation From Brain 18F-FDG PET/CT Images to Fluid-Attenuated Inversion Recovery Images Using the CypixGAN Framework. Clin Nucl Med 2024:00003072-990000000-01280. [PMID: 39325494 DOI: 10.1097/rlu.0000000000005441] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/27/2024]
Abstract
PURPOSE PET/CT and MRI can accurately diagnose dementia but are expensive and inconvenient for patients. Therefore, we aimed to generate synthetic fluid-attenuated inversion recovery (FLAIR) images from 18F-FDG PET and CT images of the human brain using a generative adversarial network (GAN)-based deep learning framework called the CypixGAN, which combined the CycleGAN framework with the L1 loss function of the pix2pix. PATIENTS AND METHODS Data from 143 patients who underwent PET/CT and MRI were used for training (n = 79), validation (n = 20), and testing (n = 44) the deep learning frameworks. Synthetic FLAIR images were generated using the pix2pix, CycleGAN, and CypixGAN, and white matter hyperintensities (WMHs) were then segmented. The performance of CypixGAN was compared with that of the other frameworks. RESULTS The CypixGAN outperformed the pix2pix and CycleGAN in generating synthetic FLAIR images with superior visual quality. Peak signal-to-noise ratio and structural similarity index (mean ± standard deviation) estimated using the CypixGAN (20.23 ± 1.31 and 0.80 ± 0.02, respectively) were significantly higher than those estimated using the pix2pix (19.35 ± 1.43 and 0.79 ± 0.02, respectively) and CycleGAN (18.74 ± 1.49 and 0.78 ± 0.02, respectively) (P < 0.001). WMHs in synthetic FLAIR images generated using the CypixGAN closely resembled those in ground-truth images, as indicated by the low absolute percentage volume differences and high dice similarity coefficients. CONCLUSIONS The CypixGAN generated high-quality FLAIR images owing to the preservation of spatial information despite using unpaired images. This framework may help improve diagnostic performance and cost-effectiveness of PET/CT when MRI scan is unavailable.
Collapse
Affiliation(s)
- Sangwon Lee
- From the Department of Electronic Engineering, Sogang University, Seoul, Republic of Korea
| | - Jin Ho Jung
- From the Department of Electronic Engineering, Sogang University, Seoul, Republic of Korea
| | - Yong Choi
- From the Department of Electronic Engineering, Sogang University, Seoul, Republic of Korea
| | - Eunyeong Seok
- From the Department of Electronic Engineering, Sogang University, Seoul, Republic of Korea
| | - Jiwoong Jung
- From the Department of Electronic Engineering, Sogang University, Seoul, Republic of Korea
| | - Hyunkeong Lim
- Department of Nuclear Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| | | | - Mijin Yun
- Department of Nuclear Medicine, Yonsei University College of Medicine, Seoul, Republic of Korea
| |
Collapse
|
2
|
Pan S, Abouei E, Peng J, Qian J, Wynne JF, Wang T, Chang CW, Roper J, Nye JA, Mao H, Yang X. Full-dose whole-body PET synthesis from low-dose PET using high-efficiency denoising diffusion probabilistic model: PET consistency model. Med Phys 2024; 51:5468-5478. [PMID: 38588512 PMCID: PMC11321936 DOI: 10.1002/mp.17068] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2023] [Revised: 03/26/2024] [Accepted: 03/26/2024] [Indexed: 04/10/2024] Open
Abstract
PURPOSE Positron Emission Tomography (PET) has been a commonly used imaging modality in broad clinical applications. One of the most important tradeoffs in PET imaging is between image quality and radiation dose: high image quality comes with high radiation exposure. Improving image quality is desirable for all clinical applications while minimizing radiation exposure is needed to reduce risk to patients. METHODS We introduce PET Consistency Model (PET-CM), an efficient diffusion-based method for generating high-quality full-dose PET images from low-dose PET images. It employs a two-step process, adding Gaussian noise to full-dose PET images in the forward diffusion, and then denoising them using a PET Shifted-window Vision Transformer (PET-VIT) network in the reverse diffusion. The PET-VIT network learns a consistency function that enables direct denoising of Gaussian noise into clean full-dose PET images. PET-CM achieves state-of-the-art image quality while requiring significantly less computation time than other methods. Evaluation with normalized mean absolute error (NMAE), peak signal-to-noise ratio (PSNR), multi-scale structure similarity index (SSIM), normalized cross-correlation (NCC), and clinical evaluation including Human Ranking Score (HRS) and Standardized Uptake Value (SUV) Error analysis shows its superiority in synthesizing full-dose PET images from low-dose inputs. RESULTS In experiments comparing eighth-dose to full-dose images, PET-CM demonstrated impressive performance with NMAE of 1.278 ± 0.122%, PSNR of 33.783 ± 0.824 dB, SSIM of 0.964 ± 0.009, NCC of 0.968 ± 0.011, HRS of 4.543, and SUV Error of 0.255 ± 0.318%, with an average generation time of 62 s per patient. This is a significant improvement compared to the state-of-the-art diffusion-based model with PET-CM reaching this result 12× faster. Similarly, in the quarter-dose to full-dose image experiments, PET-CM delivered competitive outcomes, achieving an NMAE of 0.973 ± 0.066%, PSNR of 36.172 ± 0.801 dB, SSIM of 0.984 ± 0.004, NCC of 0.990 ± 0.005, HRS of 4.428, and SUV Error of 0.151 ± 0.192% using the same generation process, which underlining its high quantitative and clinical precision in both denoising scenario. CONCLUSIONS We propose PET-CM, the first efficient diffusion-model-based method, for estimating full-dose PET images from low-dose images. PET-CM provides comparable quality to the state-of-the-art diffusion model with higher efficiency. By utilizing this approach, it becomes possible to maintain high-quality PET images suitable for clinical use while mitigating the risks associated with radiation. The code is availble at https://github.com/shaoyanpan/Full-dose-Whole-body-PET-Synthesis-from-Low-dose-PET-Using-Consistency-Model.
Collapse
Affiliation(s)
- Shaoyan Pan
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
- Department of Biomedical Informatics, Emory University, Atlanta, GA 30322, USA
| | - Elham Abouei
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Junbo Peng
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Joshua Qian
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Jacob F Wynne
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Tonghe Wang
- Department of Medical Physics, Memorial Sloan Kettering Cancer Center, New York, NY 10065, USA
| | - Chih-Wei Chang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Justin Roper
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Jonathon A Nye
- Radiology and Radiological Science, Medical University of South Carolina, Charleston, SC 29425, USA
| | - Hui Mao
- Department of Radiology and Imaging Science, and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, USA
- Department of Biomedical Informatics, Emory University, Atlanta, GA 30322, USA
| |
Collapse
|
3
|
Li Y, Li Y. PETformer network enables ultra-low-dose total-body PET imaging without structural prior. Phys Med Biol 2024; 69:075030. [PMID: 38417180 DOI: 10.1088/1361-6560/ad2e6f] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 02/28/2024] [Indexed: 03/01/2024]
Abstract
Objective.Positron emission tomography (PET) is essential for non-invasive imaging of metabolic processes in healthcare applications. However, the use of radiolabeled tracers exposes patients to ionizing radiation, raising concerns about carcinogenic potential, and warranting efforts to minimize doses without sacrificing diagnostic quality.Approach.In this work, we present a novel neural network architecture, PETformer, designed for denoising ultra-low-dose PET images without requiring structural priors such as computed tomography (CT) or magnetic resonance imaging. The architecture utilizes a U-net backbone, synergistically combining multi-headed transposed attention blocks with kernel-basis attention and channel attention mechanisms for both short- and long-range dependencies and enhanced feature extraction. PETformer is trained and validated on a dataset of 317 patients imaged on a total-body uEXPLORER PET/CT scanner.Main results.Quantitative evaluations using structural similarity index measure and liver signal-to-noise ratio showed PETformer's significant superiority over other established denoising algorithms across different dose-reduction factors.Significance.Its ability to identify and recover intrinsic anatomical details from background noise with dose reductions as low as 2% and its capacity in maintaining high target-to-background ratios while preserving the integrity of uptake values of small lesions enables PET-only fast and accurate disease diagnosis. Furthermore, PETformer exhibits computational efficiency with only 37 M trainable parameters, making it well-suited for commercial integration.
Collapse
Affiliation(s)
- Yuxiang Li
- United Imaging Healthcare America, Houston, TX, 77054, United States of America
- Division of Otolaryngology-Head and Neck Surgery, Department of Surgery, University of California, San Diego, CA 92093, United States of America
- Research Service, VA San Diego Healthcare System, San Diego, CA 92161, United States of America
| | - Yusheng Li
- United Imaging Healthcare America, Houston, TX, 77054, United States of America
| |
Collapse
|
4
|
Wang Y, Luo Y, Zu C, Zhan B, Jiao Z, Wu X, Zhou J, Shen D, Zhou L. 3D multi-modality Transformer-GAN for high-quality PET reconstruction. Med Image Anal 2024; 91:102983. [PMID: 37926035 DOI: 10.1016/j.media.2023.102983] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2022] [Revised: 08/06/2023] [Accepted: 09/28/2023] [Indexed: 11/07/2023]
Abstract
Positron emission tomography (PET) scans can reveal abnormal metabolic activities of cells and provide favorable information for clinical patient diagnosis. Generally, standard-dose PET (SPET) images contain more diagnostic information than low-dose PET (LPET) images but higher-dose scans can also bring higher potential radiation risks. To reduce the radiation risk while acquiring high-quality PET images, in this paper, we propose a 3D multi-modality edge-aware Transformer-GAN for high-quality SPET reconstruction using the corresponding LPET images and T1 acquisitions from magnetic resonance imaging (T1-MRI). Specifically, to fully excavate the metabolic distributions in LPET and anatomical structural information in T1-MRI, we first use two separate CNN-based encoders to extract local spatial features from the two modalities, respectively, and design a multimodal feature integration module to effectively integrate the two kinds of features given the diverse contributions of features at different locations. Then, as CNNs can describe local spatial information well but have difficulty in modeling long-range dependencies in images, we further apply a Transformer-based encoder to extract global semantic information in the input images and use a CNN decoder to transform the encoded features into SPET images. Finally, a patch-based discriminator is applied to ensure the similarity of patch-wise data distribution between the reconstructed and real images. Considering the importance of edge information in anatomical structures for clinical disease diagnosis, besides voxel-level estimation error and adversarial loss, we also introduce an edge-aware loss to retain more edge detail information in the reconstructed SPET images. Experiments on the phantom dataset and clinical dataset validate that our proposed method can effectively reconstruct high-quality SPET images and outperform current state-of-the-art methods in terms of qualitative and quantitative metrics.
Collapse
Affiliation(s)
- Yan Wang
- School of Computer Science, Sichuan University, Chengdu, China
| | - Yanmei Luo
- School of Computer Science, Sichuan University, Chengdu, China
| | - Chen Zu
- Department of Risk Controlling Research, JD.COM, China
| | - Bo Zhan
- School of Computer Science, Sichuan University, Chengdu, China
| | - Zhengyang Jiao
- School of Computer Science, Sichuan University, Chengdu, China
| | - Xi Wu
- School of Computer Science, Chengdu University of Information Technology, China
| | - Jiliu Zhou
- School of Computer Science, Sichuan University, Chengdu, China
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China; Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China.
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Australia.
| |
Collapse
|
5
|
Sun H, Jiang Y, Yuan J, Wang H, Liang D, Fan W, Hu Z, Zhang N. High-quality PET image synthesis from ultra-low-dose PET/MRI using bi-task deep learning. Quant Imaging Med Surg 2022; 12:5326-5342. [PMID: 36465830 PMCID: PMC9703111 DOI: 10.21037/qims-22-116] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Accepted: 08/04/2022] [Indexed: 01/25/2023]
Abstract
BACKGROUND Lowering the dose for positron emission tomography (PET) imaging reduces patients' radiation burden but decreases the image quality by increasing noise and reducing imaging detail and quantifications. This paper introduces a method for acquiring high-quality PET images from an ultra-low-dose state to achieve both high-quality images and a low radiation burden. METHODS We developed a two-task-based end-to-end generative adversarial network, named bi-c-GAN, that incorporated the advantages of PET and magnetic resonance imaging (MRI) modalities to synthesize high-quality PET images from an ultra-low-dose input. Moreover, a combined loss, including the mean absolute error, structural loss, and bias loss, was created to improve the trained model's performance. Real integrated PET/MRI data from 67 patients' axial heads (each with 161 slices) were used for training and validation purposes. Synthesized images were quantified by the peak signal-to-noise ratio (PSNR), normalized mean square error (NMSE), structural similarity (SSIM), and contrast noise ratio (CNR). The improvement ratios of these four selected quantitative metrics were used to compare the images produced by bi-c-GAN with other methods. RESULTS In the four-fold cross-validation, the proposed bi-c-GAN outperformed the other three selected methods (U-net, c-GAN, and multiple input c-GAN). With the bi-c-GAN, in a 5% low-dose PET, the image quality was higher than that of the other three methods by at least 6.7% in the PSNR, 0.6% in the SSIM, 1.3% in the NMSE, and 8% in the CNR. In the hold-out validation, bi-c-GAN improved the image quality compared to U-net and c-GAN in both 2.5% and 10% low-dose PET. For example, the PSNR using bi-C-GAN was at least 4.46% in the 2.5% low-dose PET and at most 14.88% in the 10% low-dose PET. Visual examples also showed a higher quality of images generated from the proposed method, demonstrating the denoising and improving ability of bi-c-GAN. CONCLUSIONS By taking advantage of integrated PET/MR images and multitask deep learning (MDL), the proposed bi-c-GAN can efficiently improve the image quality of ultra-low-dose PET and reduce radiation exposure.
Collapse
Affiliation(s)
- Hanyu Sun
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Yongluo Jiang
- Department of Nuclear Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Jianmin Yuan
- Central Research Institute, Shanghai United Imaging Healthcare, Shanghai, China
| | - Haining Wang
- United Imaging Research Institute of Innovative Medical Equipment, Shenzhen, China
| | - Dong Liang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Wei Fan
- Department of Nuclear Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Zhanli Hu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China;,United Imaging Research Institute of Innovative Medical Equipment, Shenzhen, China
| | - Na Zhang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China;,United Imaging Research Institute of Innovative Medical Equipment, Shenzhen, China
| |
Collapse
|
6
|
Castillo-Flores S, Gonzalez MR, Bryce-Alberti M, de Souza F, Subhawong TK, Kuker R, Pretell-Mazzini J. PET-CT in the Evaluation of Neoadjuvant/Adjuvant Treatment Response of Soft-tissue Sarcomas: A Comprehensive Review of the Literature. JBJS Rev 2022; 10:01874474-202212000-00003. [PMID: 36639875 DOI: 10.2106/jbjs.rvw.22.00131] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
Abstract
➢ In soft-tissue sarcomas (STSs), the use of positron emission tomography-computed tomography (PET-CT) through a standardized uptake value reduction rate correlates well with histopathological response to neoadjuvant treatment and survival. ➢ PET-CT has shown a better sensitivity to diagnose systemic involvement compared with magnetic resonance imaging and CT; therefore, it has an important role in detecting recurrent systemic disease. However, delaying the use of PET-CT scan, to differentiate tumor recurrence from benign fluorodeoxyglucose uptake changes after surgical treatment and radiotherapy, is essential. ➢ PET-CT limitations such as difficult differentiation between benign inflammatory and malignant processes, inefficient discrimination between benign soft-tissue tumors and STSs, and low sensitivity when evaluating small pulmonary metastases must be of special consideration.
Collapse
Affiliation(s)
- Samy Castillo-Flores
- Medical Student at Facultad de Medicina Alberto Hurtado, Universidad Peruana Cayetano Heredia, Lima, Peru
| | - Marcos R Gonzalez
- Medical Student at Facultad de Medicina Alberto Hurtado, Universidad Peruana Cayetano Heredia, Lima, Peru
| | - Mayte Bryce-Alberti
- Medical Student at Facultad de Medicina Alberto Hurtado, Universidad Peruana Cayetano Heredia, Lima, Peru
| | - Felipe de Souza
- Division of Musculoskeletal Radiology, Department of Radiology, University of Miami Miller School of Medicine, Miami, Florida
| | - Ty K Subhawong
- Division of Musculoskeletal Radiology, Department of Radiology, University of Miami Miller School of Medicine, Miami, Florida
| | - Russ Kuker
- Division of Musculoskeletal Radiology, Department of Radiology, University of Miami Miller School of Medicine, Miami, Florida
| | - Juan Pretell-Mazzini
- Division of Orthopedic Oncology, Miami Cancer Institute, Baptist Health System South Florida, Plantation, Florida
| |
Collapse
|
7
|
FDG-PET to T1 Weighted MRI Translation with 3D Elicit Generative Adversarial Network (E-GAN). SENSORS 2022; 22:s22124640. [PMID: 35746422 PMCID: PMC9227640 DOI: 10.3390/s22124640] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/28/2022] [Revised: 05/30/2022] [Accepted: 06/07/2022] [Indexed: 02/01/2023]
Abstract
Objective: With the strengths of deep learning, computer-aided diagnosis (CAD) is a hot topic for researchers in medical image analysis. One of the main requirements for training a deep learning model is providing enough data for the network. However, in medical images, due to the difficulties of data collection and data privacy, finding an appropriate dataset (balanced, enough samples, etc.) is quite a challenge. Although image synthesis could be beneficial to overcome this issue, synthesizing 3D images is a hard task. The main objective of this paper is to generate 3D T1 weighted MRI corresponding to FDG-PET. In this study, we propose a separable convolution-based Elicit generative adversarial network (E-GAN). The proposed architecture can reconstruct 3D T1 weighted MRI from 2D high-level features and geometrical information retrieved from a Sobel filter. Experimental results on the ADNI datasets for healthy subjects show that the proposed model improves the quality of images compared with the state of the art. In addition, the evaluation of E-GAN and the state of art methods gives a better result on the structural information (13.73% improvement for PSNR and 22.95% for SSIM compared to Pix2Pix GAN) and textural information (6.9% improvements for homogeneity error in Haralick features compared to Pix2Pix GAN).
Collapse
|
8
|
Bonardel G, Dupont A, Decazes P, Queneau M, Modzelewski R, Coulot J, Le Calvez N, Hapdey S. Clinical and phantom validation of a deep learning based denoising algorithm for F-18-FDG PET images from lower detection counting in comparison with the standard acquisition. EJNMMI Phys 2022; 9:36. [PMID: 35543894 PMCID: PMC9095795 DOI: 10.1186/s40658-022-00465-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2021] [Accepted: 04/20/2022] [Indexed: 11/21/2022] Open
Abstract
Background PET/CT image quality is directly influenced by the F-18-FDG injected activity. The higher the injected activity, the less noise in the reconstructed images but the more radioactive staff exposition. A new FDA cleared software has been introduced to obtain clinical PET images, acquired at 25% of the count statistics considering US practices. Our aim is to determine the limits of a deep learning based denoising algorithm (SubtlePET) applied to statistically reduced PET raw data from 3 different last generation PET scanners in comparison to the regular acquisition in phantom and patients, considering the European guidelines for radiotracer injection activities. Images of low and high contrasted (SBR = 2 and 5) spheres of the IEC phantom and high contrast (SBR = 5) of micro-spheres of Jaszczak phantom were acquired on 3 different PET devices. 110 patients with different pathologies were included. The data was acquired in list-mode and retrospectively reconstructed with the regular acquisition count statistic (PET100), 50% reduction in counts (PET50) and 66% reduction in counts (PET33). These count reduced images were post-processed with SubtlePET to obtain PET50 + SP and PET33 + SP images. Patient image quality was scored by 2 senior nuclear physicians. Peak-signal-to-Noise and Structural similarity metrics were computed to compare the low count images to regular acquisition (PET100). Results SubtlePET reliably denoised the images and maintained the SUVmax values in PET50 + SP. SubtlePET enhanced images (PET33 + SP) had slightly increased noise compared to PET100 and could lead to a potential loss of information in terms of lesion detectability. Regarding the patient datasets, the PET100 and PET50 + SP were qualitatively comparable. The SubtlePET algorithm was able to correctly recover the SUVmax values of the lesions and maintain a noise level equivalent to full-time images. Conclusion Based on our results, SubtlePET is adapted in clinical practice for half-time or half-dose acquisitions based on European recommended injected dose of 3 MBq/kg without diagnostic confidence loss. Supplementary Information The online version contains supplementary material available at 10.1186/s40658-022-00465-z.
Collapse
Affiliation(s)
- Gerald Bonardel
- Nuclear Medicine, Centre Cardiologique du Nord, Saint-Denis, France.,Nuclear Medicine, Hopital Delafontaine, Saint-Denis, France
| | | | - Pierre Decazes
- Nuclear Medicine Department, Henri Becquerel Cancer Center, Rouen, France.,QuantIF-LITIS EA4108, Rouen University Hospital, Rouen, France
| | - Mathieu Queneau
- Nuclear Medicine, Centre Cardiologique du Nord, Saint-Denis, France.,Nuclear Medicine, Hopital Delafontaine, Saint-Denis, France
| | - Romain Modzelewski
- Nuclear Medicine Department, Henri Becquerel Cancer Center, Rouen, France.,QuantIF-LITIS EA4108, Rouen University Hospital, Rouen, France
| | | | - Nicolas Le Calvez
- Nuclear Medicine, Centre Cardiologique du Nord, Saint-Denis, France.,Nuclear Medicine, Hopital Delafontaine, Saint-Denis, France
| | - Sébastien Hapdey
- Nuclear Medicine Department, Henri Becquerel Cancer Center, Rouen, France. .,QuantIF-LITIS EA4108, Rouen University Hospital, Rouen, France.
| |
Collapse
|
9
|
Xue S, Guo R, Bohn KP, Matzke J, Viscione M, Alberts I, Meng H, Sun C, Zhang M, Zhang M, Sznitman R, El Fakhri G, Rominger A, Li B, Shi K. A cross-scanner and cross-tracer deep learning method for the recovery of standard-dose imaging quality from low-dose PET. Eur J Nucl Med Mol Imaging 2022; 49:1843-1856. [PMID: 34950968 PMCID: PMC9015984 DOI: 10.1007/s00259-021-05644-1] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2021] [Accepted: 11/30/2021] [Indexed: 11/12/2022]
Abstract
PURPOSE A critical bottleneck for the credibility of artificial intelligence (AI) is replicating the results in the diversity of clinical practice. We aimed to develop an AI that can be independently applied to recover high-quality imaging from low-dose scans on different scanners and tracers. METHODS Brain [18F]FDG PET imaging of 237 patients scanned with one scanner was used for the development of AI technology. The developed algorithm was then tested on [18F]FDG PET images of 45 patients scanned with three different scanners, [18F]FET PET images of 18 patients scanned with two different scanners, as well as [18F]Florbetapir images of 10 patients. A conditional generative adversarial network (GAN) was customized for cross-scanner and cross-tracer optimization. Three nuclear medicine physicians independently assessed the utility of the results in a clinical setting. RESULTS The improvement achieved by AI recovery significantly correlated with the baseline image quality indicated by structural similarity index measurement (SSIM) (r = -0.71, p < 0.05) and normalized dose acquisition (r = -0.60, p < 0.05). Our cross-scanner and cross-tracer AI methodology showed utility based on both physical and clinical image assessment (p < 0.05). CONCLUSION The deep learning development for extensible application on unknown scanners and tracers may improve the trustworthiness and clinical acceptability of AI-based dose reduction.
Collapse
Affiliation(s)
- Song Xue
- Department of Nuclear Medicine, University of Bern, Bern, Switzerland
| | - Rui Guo
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Collaborative Innovation Center for Molecular Imaging of Precision Medicine, Ruijin Center, Shanghai, China
| | - Karl Peter Bohn
- Department of Nuclear Medicine, University of Bern, Bern, Switzerland
| | - Jared Matzke
- Department of Informatics, Technical University of Munich, Munich, Germany
| | - Marco Viscione
- Department of Nuclear Medicine, University of Bern, Bern, Switzerland
| | - Ian Alberts
- Department of Nuclear Medicine, University of Bern, Bern, Switzerland
| | - Hongping Meng
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Collaborative Innovation Center for Molecular Imaging of Precision Medicine, Ruijin Center, Shanghai, China
| | - Chenwei Sun
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Collaborative Innovation Center for Molecular Imaging of Precision Medicine, Ruijin Center, Shanghai, China
| | - Miao Zhang
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Collaborative Innovation Center for Molecular Imaging of Precision Medicine, Ruijin Center, Shanghai, China
| | - Min Zhang
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Collaborative Innovation Center for Molecular Imaging of Precision Medicine, Ruijin Center, Shanghai, China
| | | | - Georges El Fakhri
- Gordon Center for Medical Imaging, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Axel Rominger
- Department of Nuclear Medicine, University of Bern, Bern, Switzerland
| | - Biao Li
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
- Collaborative Innovation Center for Molecular Imaging of Precision Medicine, Ruijin Center, Shanghai, China.
| | - Kuangyu Shi
- Department of Nuclear Medicine, University of Bern, Bern, Switzerland
- Department of Informatics, Technical University of Munich, Munich, Germany
| |
Collapse
|
10
|
Deng F, Li X, Yang F, Sun H, Yuan J, He Q, Xu W, Yang Y, Liang D, Liu X, Mok GSP, Zheng H, Hu Z. Low-Dose 68 Ga-PSMA Prostate PET/MRI Imaging Using Deep Learning Based on MRI Priors. Front Oncol 2022; 11:818329. [PMID: 35155207 PMCID: PMC8825350 DOI: 10.3389/fonc.2021.818329] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2021] [Accepted: 12/27/2021] [Indexed: 12/02/2022] Open
Abstract
Background 68 Ga-prostate-specific membrane antigen (PSMA) PET/MRI has become an effective imaging method for prostate cancer. The purpose of this study was to use deep learning methods to perform low-dose image restoration on PSMA PET/MRI and to evaluate the effect of synthesis on the images and the medical diagnosis of patients at risk of prostate cancer. Methods We reviewed the 68 Ga-PSMA PET/MRI data of 41 patients. The low-dose PET (LDPET) images of these patients were restored to full-dose PET (FDPET) images through a deep learning method based on MRI priors. The synthesized images were evaluated according to quantitative scores from nuclear medicine doctors and multiple imaging indicators, such as peak-signal noise ratio (PSNR), structural similarity (SSIM), normalization mean square error (NMSE), and relative contrast-to-noise ratio (RCNR). Results The clinical quantitative scores of the FDPET images synthesized from 25%- and 50%-dose images based on MRI priors were 3.84±0.36 and 4.03±0.17, respectively, which were higher than the scores of the target images. Correspondingly, the PSNR, SSIM, NMSE, and RCNR values of the FDPET images synthesized from 50%-dose PET images based on MRI priors were 39.88±3.83, 0.896±0.092, 0.012±0.007, and 0.996±0.080, respectively. Conclusion According to a combination of quantitative scores from nuclear medicine doctors and evaluations with multiple image indicators, the synthesis of FDPET images based on MRI priors using and 50%-dose PET images did not affect the clinical diagnosis of prostate cancer. Prostate cancer patients can undergo 68 Ga-PSMA prostate PET/MRI scans with radiation doses reduced by up to 50% through the use of deep learning methods to synthesize FDPET images.
Collapse
Affiliation(s)
- Fuquan Deng
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Computer Department, North China Electric Power University, Baoding, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Xiaoyuan Li
- Department of Nuclear Medicine, Nanjing First Hospital, Nanjing Medical University, Nanjing, China
| | - Fengjiao Yang
- Department of Nuclear Medicine, Nanjing First Hospital, Nanjing Medical University, Nanjing, China
| | - Hongwei Sun
- United Imaging Research Institute of Intelligent Imaging, Beijing, China
| | - Jianmin Yuan
- Central Research Institute, United Imaging Healthcare Group, Shanghai, China
| | - Qiang He
- Central Research Institute, United Imaging Healthcare Group, Shanghai, China
| | - Weifeng Xu
- Computer Department, North China Electric Power University, Baoding, China
| | - Yongfeng Yang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Dong Liang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Xin Liu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Greta S P Mok
- Biomedical Imaging Laboratory (BIG), Department of Electrical and Computer Engineering, Faculty of Science and Technology, University of Macau, Avenida da Universidade, Macau SAR, China
| | - Hairong Zheng
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Zhanli Hu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| |
Collapse
|
11
|
Luo Y, Zhou L, Zhan B, Fei Y, Zhou J, Wang Y, Shen D. Adaptive rectification based adversarial network with spectrum constraint for high-quality PET image synthesis. Med Image Anal 2021; 77:102335. [PMID: 34979432 DOI: 10.1016/j.media.2021.102335] [Citation(s) in RCA: 19] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/24/2021] [Revised: 11/02/2021] [Accepted: 12/13/2021] [Indexed: 12/13/2022]
Abstract
Positron emission tomography (PET) is a typical nuclear imaging technique, which can provide crucial functional information for early brain disease diagnosis. Generally, clinically acceptable PET images are obtained by injecting a standard-dose radioactive tracer into human body, while on the other hand the cumulative radiation exposure inevitably raises concerns about potential health risks. However, reducing the tracer dose will increase the noise and artifacts of the reconstructed PET image. For the purpose of acquiring high-quality PET images while reducing radiation exposure, in this paper, we innovatively present an adaptive rectification based generative adversarial network with spectrum constraint, named AR-GAN, which uses low-dose PET (LPET) image to synthesize standard-dose PET (SPET) image of high-quality. Specifically, considering the existing differences between the synthesized SPET image by traditional GAN and the real SPET image, an adaptive rectification network (AR-Net) is devised to estimate the residual between the preliminarily predicted image and the real SPET image, based on the hypothesis that a more realistic rectified image can be obtained by incorporating both the residual and the preliminarily predicted PET image. Moreover, to address the issue of high-frequency distortions in the output image, we employ a spectral regularization term in the training optimization objective to constrain the consistency of the synthesized image and the real image in the frequency domain, which further preserves the high-frequency detailed information and improves synthesis performance. Validations on both the phantom dataset and the clinical dataset show that the proposed AR-GAN can estimate SPET images from LPET images effectively and outperform other state-of-the-art image synthesis approaches.
Collapse
Affiliation(s)
- Yanmei Luo
- School of Computer Science, Sichuan University, China
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Australia
| | - Bo Zhan
- School of Computer Science, Sichuan University, China
| | - Yuchen Fei
- School of Computer Science, Sichuan University, China
| | - Jiliu Zhou
- School of Computer Science, Sichuan University, China; School of Computer Science, Chengdu University of Information Technology, China
| | - Yan Wang
- School of Computer Science, Sichuan University, China.
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, China; Department of Research and Development, Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
| |
Collapse
|
12
|
Wang S, Cao G, Wang Y, Liao S, Wang Q, Shi J, Li C, Shen D. Review and Prospect: Artificial Intelligence in Advanced Medical Imaging. FRONTIERS IN RADIOLOGY 2021; 1:781868. [PMID: 37492170 PMCID: PMC10365109 DOI: 10.3389/fradi.2021.781868] [Citation(s) in RCA: 24] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/23/2021] [Accepted: 11/08/2021] [Indexed: 07/27/2023]
Abstract
Artificial intelligence (AI) as an emerging technology is gaining momentum in medical imaging. Recently, deep learning-based AI techniques have been actively investigated in medical imaging, and its potential applications range from data acquisition and image reconstruction to image analysis and understanding. In this review, we focus on the use of deep learning in image reconstruction for advanced medical imaging modalities including magnetic resonance imaging (MRI), computed tomography (CT), and positron emission tomography (PET). Particularly, recent deep learning-based methods for image reconstruction will be emphasized, in accordance with their methodology designs and performances in handling volumetric imaging data. It is expected that this review can help relevant researchers understand how to adapt AI for medical imaging and which advantages can be achieved with the assistance of AI.
Collapse
Affiliation(s)
- Shanshan Wang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences (CAS), Shenzhen, China
- Pengcheng Laboratrory, Shenzhen, China
| | - Guohua Cao
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Yan Wang
- School of Computer Science, Sichuan University, Chengdu, China
| | - Shu Liao
- Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
| | - Qian Wang
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
| | - Jun Shi
- School of Communication and Information Engineering, Shanghai University, Shanghai, China
| | - Cheng Li
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences (CAS), Shenzhen, China
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
- Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
| |
Collapse
|
13
|
Zhang L, Xiao Z, Zhou C, Yuan J, He Q, Yang Y, Liu X, Liang D, Zheng H, Fan W, Zhang X, Hu Z. Spatial adaptive and transformer fusion network (STFNet) for low-count PET blind denoising with MRI. Med Phys 2021; 49:343-356. [PMID: 34796526 DOI: 10.1002/mp.15368] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2021] [Revised: 10/28/2021] [Accepted: 11/08/2021] [Indexed: 01/12/2023] Open
Abstract
PURPOSE Positron emission tomography (PET) has been widely used in various clinical applications. PET is a type of emission computed tomography and operates by positron annihilation radiation. With magnetic resonance imaging (MRI) providing anatomical information, joint PET/MRI reduces the radiation exposure risk of patients. Improved hardware and imaging algorithms have been proposed to further decrease the dose from radioactive tracers or the bed duration, but few methods focus on denoising low-count PET with MRI input. The existing methods are based on fixed conventional convolution and local attention, which do not sufficiently extract and fuse contextual and complementary information from multimodal input. There is still much room for improvement. Therefore, we propose a novel deep learning method for low-count PET/MRI denoising called the spatial-adaptive and transformer fusion network (STFNet), which consists of a Siamese encoder with a spatial-adaptive block (SA-block) and the transformer fusion encoder (TFE). METHODS Our proposed STFNet consists of a Siamese encoder with an SA-block, TFE, and two branches of the decoder. First, in the encoder, we adapt the SA-block in the Siamese encoder. The SA-block comprises deformable convolution with fusion modulation (DCFM) and two convolutional operations, which can promote network extraction of more relative and long-range contextual features. Second, the pixel-to-pixel TFE helps the network establish a local and global relationship between high-level feature maps of PET and MRI. In the decoder part, we design two branches for PET denoising and MRI translation, and predictions are obtained by trainable weighted summation. This proposed algorithm is implemented to predict synthetic standard-dose neck PET images from low-count neck PET images and MRI. Additionally, this method is compared with the existing U-Net and residual U-Net methods with and without MRI input. RESULTS To demonstrate the advantages of our method, we introduce configuration studies about TFE, ablation studies, and empirical comparative studies. Quantitative analyses are based on root mean square error (RSME), peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and Pearson correlation coefficient (PCC). Additionally, qualitative results show the comparisons between our proposed method and other existing methods. All experimental results and visualizations show that our method achieves state-of-the-art performance in quantification and qualification. CONCLUSIONS Based on our experiments, STFNet performs better than existing methods in measurement and visualization. However, our proposed method may still be suboptimal because we apply only the L1 loss to train our data set, and the data set includes corrupted PET with different low counts. In the future, we may exploit a generative adversarial network (GAN)-based paradigm in our STFNet to further improve the visual quality.
Collapse
Affiliation(s)
- Lipei Zhang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Zizheng Xiao
- Department of Nuclear Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Chao Zhou
- Department of Nuclear Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Jianmin Yuan
- Central Research Institute, Shanghai United Imaging Healthcare, Shanghai, China
| | - Qiang He
- Central Research Institute, Shanghai United Imaging Healthcare, Shanghai, China
| | - Yongfeng Yang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Xin Liu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Dong Liang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Hairong Zheng
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| | - Wei Fan
- Department of Nuclear Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Xu Zhang
- Department of Nuclear Medicine, Sun Yat-sen University Cancer Center, Guangzhou, China
| | - Zhanli Hu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Chinese Academy of Sciences Key Laboratory of Health Informatics, Shenzhen, China
| |
Collapse
|
14
|
Cheng Z, Wen J, Huang G, Yan J. Applications of artificial intelligence in nuclear medicine image generation. Quant Imaging Med Surg 2021; 11:2792-2822. [PMID: 34079744 PMCID: PMC8107336 DOI: 10.21037/qims-20-1078] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2020] [Accepted: 02/14/2021] [Indexed: 12/12/2022]
Abstract
Recently, the application of artificial intelligence (AI) in medical imaging (including nuclear medicine imaging) has rapidly developed. Most AI applications in nuclear medicine imaging have focused on the diagnosis, treatment monitoring, and correlation analyses with pathology or specific gene mutation. It can also be used for image generation to shorten the time of image acquisition, reduce the dose of injected tracer, and enhance image quality. This work provides an overview of the application of AI in image generation for single-photon emission computed tomography (SPECT) and positron emission tomography (PET) either without or with anatomical information [CT or magnetic resonance imaging (MRI)]. This review focused on four aspects, including imaging physics, image reconstruction, image postprocessing, and internal dosimetry. AI application in generating attenuation map, estimating scatter events, boosting image quality, and predicting internal dose map is summarized and discussed.
Collapse
Affiliation(s)
- Zhibiao Cheng
- Department of Biomedical Engineering, School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Junhai Wen
- Department of Biomedical Engineering, School of Life Science, Beijing Institute of Technology, Beijing, China
| | - Gang Huang
- Shanghai Key Laboratory of Molecular Imaging, Shanghai University of Medicine and Health Sciences, Shanghai, China
| | - Jianhua Yan
- Shanghai Key Laboratory of Molecular Imaging, Shanghai University of Medicine and Health Sciences, Shanghai, China
| |
Collapse
|
15
|
Katsari K, Penna D, Arena V, Polverari G, Ianniello A, Italiano D, Milani R, Roncacci A, Illing RO, Pelosi E. Artificial intelligence for reduced dose 18F-FDG PET examinations: a real-world deployment through a standardized framework and business case assessment. EJNMMI Phys 2021; 8:25. [PMID: 33687602 PMCID: PMC7943690 DOI: 10.1186/s40658-021-00374-7] [Citation(s) in RCA: 30] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2020] [Accepted: 02/25/2021] [Indexed: 12/14/2022] Open
Abstract
BACKGROUND To determine whether artificial intelligence (AI) processed PET/CT images of reduced by one-third of 18-F-FDG activity compared to the standard injected dose, were non-inferior to native scans and if so to assess the potential impact of commercialization. MATERIALS AND METHODS SubtlePET™ AI was introduced in a PET/CT center in Italy. Eligible patients referred for 18F-FDG PET/CT were prospectively enrolled. Administered 18F-FDG was reduced to two-thirds of standard dose. Patients underwent one low-dose CT and two sequential PET scans; "PET-processed" with reduced dose and standard acquisition time, and "PET-native" with an elapsed time to simulate standard acquisition time and dose. PET-processed images were reconstructed using SubtlePET™. PET-native images were defined as the standard of reference. The datasets were anonymized and independently evaluated in random order by four blinded readers. The evaluation included subjective image quality (IQ) assessment, lesion detectability, and assessment of business benefits. RESULTS From February to April 2020, 61 patients were prospectively enrolled. Subjective IQ was not significantly different between datasets (4.62±0.23, p=0.237) for all scanner models, with "almost perfect" inter-reader agreement. There was no significant difference between datasets in lesions' detectability, target lesion mean SUVmax value, and liver mean SUVmean value (182.75/181.75 [SD:0.71], 9.8/11.4 [SD:1.13], 2.1/1.9 [SD:0.14] respectively). No false-positive lesions were reported in PET-processed examinations. Agreed SubtlePET™ price per examination was 15-20% of FDG savings. CONCLUSION This is the first real-world study to demonstrate the non-inferiority of AI processed 18F-FDG PET/CT examinations obtained with 66% standard dose and a methodology to define the AI solution price.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | | | - Rowland O Illing
- Affidea, Budapest, Hungary
- University College London United Kingdom, London, UK
| | | |
Collapse
|
16
|
da Costa-Luis CO, Reader AJ. Micro-Networks for Robust MR-Guided Low Count PET Imaging. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2021; 5:202-212. [PMID: 33681546 PMCID: PMC7931458 DOI: 10.1109/trpms.2020.2986414] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2019] [Revised: 12/30/2019] [Accepted: 02/08/2020] [Indexed: 01/18/2023]
Abstract
Noise suppression is particularly important in low count positron emission tomography (PET) imaging. Post-smoothing (PS) and regularization methods which aim to reduce noise also tend to reduce resolution and introduce bias. Alternatively, anatomical information from another modality such as magnetic resonance (MR) imaging can be used to improve image quality. Convolutional neural networks (CNNs) are particularly well suited to such joint image processing, but usually require large amounts of training data and have mostly been applied outside the field of medical imaging or focus on classification and segmentation, leaving PET image quality improvement relatively understudied. This article proposes the use of a relatively low-complexity CNN (micro-net) as a post-reconstruction MR-guided image processing step to reduce noise and reconstruction artefacts while also improving resolution in low count PET scans. The CNN is designed to be fully 3-D, robust to very limited amounts of training data, and to accept multiple inputs (including competitive denoising methods). Application of the proposed CNN on simulated low (30 M) count data (trained to produce standard (300 M) count reconstructions) results in a 36% lower normalized root mean squared error (NRMSE, calculated over ten realizations against the ground truth) compared to maximum-likelihood expectation maximization (MLEM) used in clinical practice. In contrast, a decrease of only 25% in NRMSE is obtained when an optimized (using knowledge of the ground truth) PS is performed. A 26% NRMSE decrease is obtained with both RM and optimized PS. Similar improvement is also observed for low count real patient datasets. Overfitting to training data is demonstrated to occur as the network size is increased. In an extreme case, a U-net (which produces better predictions for training data) is shown to completely fail on test data due to overfitting to this case of very limited training data. Meanwhile, the resultant images from the proposed CNN (which has low training data requirements) have lower noise, reduced ringing, and partial volume effects, as well as sharper edges and improved resolution compared to conventional MLEM.
Collapse
Affiliation(s)
- Casper O. da Costa-Luis
- Department of Biomedical EngineeringSchool of Biomedical Engineering and Imaging Sciences, St. Thomas’ HospitalKing’s College LondonLondonSE1 7EHU.K.
| | - Andrew J. Reader
- Department of Biomedical EngineeringSchool of Biomedical Engineering and Imaging Sciences, St. Thomas’ HospitalKing’s College LondonLondonSE1 7EHU.K.
| |
Collapse
|
17
|
Kim J, Kang S, Lee K, Ho Jung J, Kim G, Keong Lim H, Choi Y, Lee S, Yun M. Effect of Scan Time on Neuro 18F-Fluorodeoxyglucose Positron Emission Tomography Image Generated Using Deep Learning. JOURNAL OF MEDICAL IMAGING AND HEALTH INFORMATICS 2021. [DOI: 10.1166/jmihi.2021.3316] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
The purpose of this study was to generate the PET images with high signal-to-noise ratio (SNR) acquired for typical scan durations (H-PET) from short scan time PET images with low SNR (L-PET) using deep learning and to evaluate the effect of scan time on the quality of predicted PET
image. A convolutional neural network (CNN) with a concatenated connection and residual learning framework was implemented. PET data from 27 patients were acquired for 900 s, starting 60 minutes after the intravenous administration of FDG using a commercial PET/CT scanner. To investigate the
effect of scan time on the quality of the predicted H-PETs, 10 s, 30 s, 60 s, and 120 s PET data were generated by sorting the 900 s LMF data into the LMF data acquired for each scan time. Twenty-three of the 27 patient images were used for training of the proposed CNN and the remaining four
patient images were used for test of the CNN. The predicted H-PETs generated by the CNN were compared to ground-truth H-PETs, L-PETs, and filtered L-PETs processed with four commonly used denoising algorithms. The peak signal-to-noise ratios (PSNRs), normalized root mean square errors (NRMSEs),
and average regionof- interest (ROI) differences as a function of scan time were calculated. The quality of the predicted H-PETs generated by the CNN was superior to that of the L-PETs and filtered L-PETs. Lower NRMSEs and higher PSNRs were also obtained from predicted H-PETs compared to the
L-PETs and filtered L-PETs. ROI differences in the predicted H-PETs were smaller than those of the L-PETs. The quality of the predicted H-PETs gradually improved with increasing scan times. The lowest NRMSEs, highest PSNRs, and smallest ROI differences were obtained using the predicted H-PETs
for 120 s. Various performance test results for the proposed CNN indicate that it is possible to generate H-PETs from neuro FDG L-PETs using the proposed CNN method, which might allow reductions in both scan time and injection dose.
Collapse
Affiliation(s)
- Jaewon Kim
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Sungsik Kang
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Konsu Lee
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Jin Ho Jung
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Garam Kim
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Hyun Keong Lim
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Yong Choi
- Department of Electronic Engineering, Sogang University, 35 Baekbeom-ro, Mapo-gu, Seoul 04107, Korea
| | - Sangwon Lee
- Departments of Nuclear Medicine, Severance Hospital, Yonsei University, 50-1 Yonsei-ro, Seodaemun-gu, Seoul 03722, Korea
| | - Mijin Yun
- Departments of Nuclear Medicine, Severance Hospital, Yonsei University, 50-1 Yonsei-ro, Seodaemun-gu, Seoul 03722, Korea
| |
Collapse
|
18
|
Wang YRJ, Baratto L, Hawk KE, Theruvath AJ, Pribnow A, Thakor AS, Gatidis S, Lu R, Gummidipundi SE, Garcia-Diaz J, Rubin D, Daldrup-Link HE. Artificial intelligence enables whole-body positron emission tomography scans with minimal radiation exposure. Eur J Nucl Med Mol Imaging 2021; 48:2771-2781. [PMID: 33527176 DOI: 10.1007/s00259-021-05197-3] [Citation(s) in RCA: 36] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2020] [Accepted: 01/10/2021] [Indexed: 02/03/2023]
Abstract
PURPOSE To generate diagnostic 18F-FDG PET images of pediatric cancer patients from ultra-low-dose 18F-FDG PET input images, using a novel artificial intelligence (AI) algorithm. METHODS We used whole-body 18F-FDG-PET/MRI scans of 33 children and young adults with lymphoma (3-30 years) to develop a convolutional neural network (CNN), which combines inputs from simulated 6.25% ultra-low-dose 18F-FDG PET scans and simultaneously acquired MRI scans to produce a standard-dose 18F-FDG PET scan. The image quality of ultra-low-dose PET scans, AI-augmented PET scans, and clinical standard PET scans was evaluated by traditional metrics in computer vision and by expert radiologists and nuclear medicine physicians, using Wilcoxon signed-rank tests and weighted kappa statistics. RESULTS The peak signal-to-noise ratio and structural similarity index were significantly higher, and the normalized root-mean-square error was significantly lower on the AI-reconstructed PET images compared to simulated 6.25% dose images (p < 0.001). Compared to the ground-truth standard-dose PET, SUVmax values of tumors and reference tissues were significantly higher on the simulated 6.25% ultra-low-dose PET scans as a result of image noise. After the CNN augmentation, the SUVmax values were recovered to values similar to the standard-dose PET. Quantitative measures of the readers' diagnostic confidence demonstrated significantly higher agreement between standard clinical scans and AI-reconstructed PET scans (kappa = 0.942) than 6.25% dose scans (kappa = 0.650). CONCLUSIONS Our CNN model could generate simulated clinical standard 18F-FDG PET images from ultra-low-dose inputs, while maintaining clinically relevant information in terms of diagnostic accuracy and quantitative SUV measurements.
Collapse
Affiliation(s)
- Yan-Ran Joyce Wang
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA
| | - Lucia Baratto
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA
| | - K Elizabeth Hawk
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA
| | - Ashok J Theruvath
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA
| | - Allison Pribnow
- Department of Pediatrics, Pediatric Oncology, Lucile Packard Children's Hospital, Stanford University, Stanford, CA, 94304, USA
| | - Avnesh S Thakor
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA
| | - Sergios Gatidis
- Department of Diagnostic and Interventional Radiology, University Hospital Tuebingen, Tuebingen, Germany
| | - Rong Lu
- Quantitative Sciences Unit, School of Medicine, Stanford University, Stanford, CA, 94304, USA
| | - Santosh E Gummidipundi
- Quantitative Sciences Unit, School of Medicine, Stanford University, Stanford, CA, 94304, USA
| | - Jordi Garcia-Diaz
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA
| | - Daniel Rubin
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA. .,Department of Pediatrics, Pediatric Oncology, Lucile Packard Children's Hospital, Stanford University, Stanford, CA, 94304, USA.
| | - Heike E Daldrup-Link
- Department of Radiology, Molecular Imaging Program at Stanford, Stanford University, 725 Welch Road, CA, 94304, Stanford, USA. .,Department of Pediatrics, Pediatric Oncology, Lucile Packard Children's Hospital, Stanford University, Stanford, CA, 94304, USA.
| |
Collapse
|
19
|
Xie N, Gong K, Guo N, Qin Z, Wu Z, Liu H, Li Q. Penalized-likelihood PET Image Reconstruction Using 3D Structural Convolutional Sparse Coding. IEEE Trans Biomed Eng 2020; 69:4-14. [PMID: 33284746 DOI: 10.1109/tbme.2020.3042907] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Positron emission tomography (PET) is widely used for clinical diagnosis. As PET suffers from low resolution and high noise, numerous efforts try to incorporate anatomical priors into PET image reconstruction, especially with the development of hybrid PET/CT and PET/MRI systems. In this work, we proposed a cube-based 3D structural convolutional sparse coding (CSC) concept for penalized-likelihood PET image reconstruction, named 3D PET-CSC. The proposed 3D PET-CSC takes advantage of the convolutional operation and manages to incorporate anatomical priors without the need of registration or supervised training. As 3D PET-CSC codes the whole 3D PET image, instead of patches, it alleviates the staircase artifacts commonly presented in traditional patch-based sparse coding methods. Compared with traditional coding methods in Fourier domain, the proposed method extends the 3D CSC to a straightforward approach based on the pursuit of localized cubes. Moreover, we developed the residual-image and order-subset mechanisms to further reduce the computational cost and accelerate the convergence for the proposed 3D PET-CSC method. Experiments based on computer simulations and clinical datasets demonstrate the superiority of 3D PET-CSC compared with other reference methods.
Collapse
|
20
|
Spuhler K, Serrano-Sosa M, Cattell R, DeLorenzo C, Huang C. Full-count PET recovery from low-count image using a dilated convolutional neural network. Med Phys 2020; 47:4928-4938. [PMID: 32687608 DOI: 10.1002/mp.14402] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2019] [Revised: 06/11/2020] [Accepted: 07/10/2020] [Indexed: 01/18/2023] Open
Abstract
PURPOSE Positron emission tomography (PET) is an essential technique in many clinical applications that allows for quantitative imaging at the molecular level. This study aims to develop a denoising method using a novel dilated convolutional neural network (CNN) to recover full-count images from low-count images. METHODS We adopted similar hierarchical structures as the conventional U-Net and incorporated dilated kernels in each convolution to allow the network to observe larger, more robust features within the image without the requirement of downsampling and upsampling internal representations. Our dNet was trained alongside a U-Net for comparison. Both models were evaluated using a leave-one-out cross-validation procedure on a dataset of 35 subjects (~3500 slabs), which were obtained from an ongoing 18 F-Fluorodeoxyglucose (FDG) study. Low-count PET data (10% count) were generated by randomly selecting one-tenth of all events in the associated listmode file. Analysis was done on the static image from the last 10 minutes of emission data. Both low-count PET and full-count PET were reconstructed using ordered subset expectation maximization (OSEM). Objective image quality metrics, including mean absolute percent error (MAPE), peak signal-to-noise ratio (PSNR), and structural similarity index metric (SSIM), were used to analyze the deep learning methods. Both deep learning methods were further compared to a traditional Gaussian filtering method. Further, region of interest (ROI) quantitative analysis was also used to compare U-Net and dNet architectures. RESULTS Both the U-Net and our proposed network were successfully trained to synthesize full-count PET images from the generated low-count PET images. Compared to low-count PET and Gaussian filtering, both deep learning methods improved MAPE, PSNR, and SSIM. Our dNet also systematically outperformed U-Net on all three metrics (MAPE: 4.99 ± 0.68 vs 5.31 ± 0.76, P < 0.01; PSNR: 31.55 ± 1.31 dB vs 31.05 ± 1.39, P < 0.01; SSIM: 0.9513 ± 0.0154 vs 0.9447 ± 0.0178, P < 0.01). ROI quantification showed greater quantitative improvements using dNet over U-Net. CONCLUSION This study proposed a novel approach of using dilated convolutions for recovering full-count PET images from low-count PET images.
Collapse
Affiliation(s)
- Karl Spuhler
- Department of Biomedical Engineering, Stony Brook University, Stony Brook, NY, USA
| | - Mario Serrano-Sosa
- Department of Biomedical Engineering, Stony Brook University, Stony Brook, NY, USA
| | - Renee Cattell
- Department of Biomedical Engineering, Stony Brook University, Stony Brook, NY, USA
| | - Christine DeLorenzo
- Department of Biomedical Engineering, Stony Brook University, Stony Brook, NY, USA
- Department of Psychiatry, Stony Brook University, Stony Brook, NY, USA
| | - Chuan Huang
- Department of Biomedical Engineering, Stony Brook University, Stony Brook, NY, USA
- Department of Psychiatry, Stony Brook University, Stony Brook, NY, USA
- Department of Radiology, Stony Brook University, Stony Brook, NY, USA
| |
Collapse
|
21
|
Zhou L, Schaefferkoetter JD, Tham IW, Huang G, Yan J. Supervised learning with cyclegan for low-dose FDG PET image denoising. Med Image Anal 2020; 65:101770. [DOI: 10.1016/j.media.2020.101770] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2019] [Revised: 05/20/2020] [Accepted: 07/03/2020] [Indexed: 10/23/2022]
|
22
|
Wang T, Lei Y, Fu Y, Curran WJ, Liu T, Nye JA, Yang X. Machine learning in quantitative PET: A review of attenuation correction and low-count image reconstruction methods. Phys Med 2020; 76:294-306. [PMID: 32738777 PMCID: PMC7484241 DOI: 10.1016/j.ejmp.2020.07.028] [Citation(s) in RCA: 55] [Impact Index Per Article: 13.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/23/2020] [Revised: 07/13/2020] [Accepted: 07/21/2020] [Indexed: 02/08/2023] Open
Abstract
The rapid expansion of machine learning is offering a new wave of opportunities for nuclear medicine. This paper reviews applications of machine learning for the study of attenuation correction (AC) and low-count image reconstruction in quantitative positron emission tomography (PET). Specifically, we present the developments of machine learning methodology, ranging from random forest and dictionary learning to the latest convolutional neural network-based architectures. For application in PET attenuation correction, two general strategies are reviewed: 1) generating synthetic CT from MR or non-AC PET for the purposes of PET AC, and 2) direct conversion from non-AC PET to AC PET. For low-count PET reconstruction, recent deep learning-based studies and the potential advantages over conventional machine learning-based methods are presented and discussed. In each application, the proposed methods, study designs and performance of published studies are listed and compared with a brief discussion. Finally, the overall contributions and remaining challenges are summarized.
Collapse
Affiliation(s)
- Tonghe Wang
- Department of Radiation Oncology, Emory University, Atlanta, GA, USA; Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Yang Lei
- Department of Radiation Oncology, Emory University, Atlanta, GA, USA
| | - Yabo Fu
- Department of Radiation Oncology, Emory University, Atlanta, GA, USA
| | - Walter J Curran
- Department of Radiation Oncology, Emory University, Atlanta, GA, USA; Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Tian Liu
- Department of Radiation Oncology, Emory University, Atlanta, GA, USA; Winship Cancer Institute, Emory University, Atlanta, GA, USA
| | - Jonathon A Nye
- Department of Radiology and Imaging Sciences, Emory University, Atlanta, GA, USA
| | - Xiaofeng Yang
- Department of Radiation Oncology, Emory University, Atlanta, GA, USA; Winship Cancer Institute, Emory University, Atlanta, GA, USA.
| |
Collapse
|
23
|
Sanaat A, Arabi H, Mainta I, Garibotto V, Zaidi H. Projection Space Implementation of Deep Learning-Guided Low-Dose Brain PET Imaging Improves Performance over Implementation in Image Space. J Nucl Med 2020; 61:1388-1396. [PMID: 31924718 DOI: 10.2967/jnumed.119.239327] [Citation(s) in RCA: 48] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2019] [Accepted: 01/09/2020] [Indexed: 01/08/2023] Open
Abstract
Our purpose was to assess the performance of full-dose (FD) PET image synthesis in both image and sinogram space from low-dose (LD) PET images and sinograms without sacrificing diagnostic quality using deep learning techniques. Methods: Clinical brain PET/CT studies of 140 patients were retrospectively used for LD-to-FD PET conversion. Five percent of the events were randomly selected from the FD list-mode PET data to simulate a realistic LD acquisition. A modified 3-dimensional U-Net model was implemented to predict FD sinograms in the projection space (PSS) and FD images in image space (PIS) from their corresponding LD sinograms and images, respectively. The quality of the predicted PET images was assessed by 2 nuclear medicine specialists using a 5-point grading scheme. Quantitative analysis using established metrics including the peak signal-to-noise ratio (PSNR), structural similarity index metric (SSIM), regionwise SUV bias, and first-, second- and high-order texture radiomic features in 83 brain regions for the test and evaluation datasets was also performed. Results: All PSS images were scored 4 or higher (good to excellent) by the nuclear medicine specialists. PSNR and SSIM values of 0.96 ± 0.03 and 0.97 ± 0.02, respectively, were obtained for PIS, and values of 31.70 ± 0.75 and 37.30 ± 0.71, respectively, were obtained for PSS. The average SUV bias calculated over all brain regions was 0.24% ± 0.96% and 1.05% ± 1.44% for PSS and PIS, respectively. The Bland-Altman plots reported the lowest SUV bias (0.02) and variance (95% confidence interval, -0.92 to +0.84) for PSS, compared with the reference FD images. The relative error of the homogeneity radiomic feature belonging to the gray-level cooccurrence matrix category was -1.07 ± 1.77 and 0.28 ± 1.4 for PIS and PSS, respectively. Conclusion: The qualitative assessment and quantitative analysis demonstrated that the FD PET PSS led to superior performance, resulting in higher image quality and lower SUV bias and variance than for FD PET PIS.
Collapse
Affiliation(s)
- Amirhossein Sanaat
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland
| | - Hossein Arabi
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland
| | - Ismini Mainta
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland
| | - Valentina Garibotto
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland.,Geneva University Neurocenter, Geneva University, Geneva, Switzerland
| | - Habib Zaidi
- Division of Nuclear Medicine and Molecular Imaging, Geneva University Hospital, Geneva, Switzerland .,Geneva University Neurocenter, Geneva University, Geneva, Switzerland.,Department of Nuclear Medicine and Molecular Imaging, University Medical Center Groningen, University of Groningen, Groningen, Netherlands; and.,Department of Nuclear Medicine, University of Southern Denmark, Odense, Denmark
| |
Collapse
|
24
|
Yu B, Wang Y, Wang L, Shen D, Zhou L. Medical Image Synthesis via Deep Learning. ADVANCES IN EXPERIMENTAL MEDICINE AND BIOLOGY 2020; 1213:23-44. [PMID: 32030661 DOI: 10.1007/978-3-030-33128-3_2] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/30/2023]
Abstract
Medical images have been widely used in clinics, providing visual representations of under-skin tissues in human body. By applying different imaging protocols, diverse modalities of medical images with unique characteristics of visualization can be produced. Considering the cost of scanning high-quality single modality images or homogeneous multiple modalities of images, medical image synthesis methods have been extensively explored for clinical applications. Among them, deep learning approaches, especially convolutional neural networks (CNNs) and generative adversarial networks (GANs), have rapidly become dominating for medical image synthesis in recent years. In this chapter, based on a general review of the medical image synthesis methods, we will focus on introducing typical CNNs and GANs models for medical image synthesis. Especially, we will elaborate our recent work about low-dose to high-dose PET image synthesis, and cross-modality MR image synthesis, using these models.
Collapse
Affiliation(s)
- Biting Yu
- School of Computing and Information Technology, University of Wollongong, Wollongong, NSW, Australia
| | - Yan Wang
- School of Computer Science, Sichuan University, Chengdu, China
| | - Lei Wang
- School of Computing and Information Technology, University of Wollongong, Wollongong, NSW, Australia
| | - Dinggang Shen
- IDEA Lab, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Camperdown, NSW, Australia.
| |
Collapse
|
25
|
Lei Y, Dong X, Wang T, Higgins K, Liu T, Curran WJ, Mao H, Nye JA, Yang X. Whole-body PET estimation from low count statistics using cycle-consistent generative adversarial networks. Phys Med Biol 2019; 64:215017. [PMID: 31561244 DOI: 10.1088/1361-6560/ab4891] [Citation(s) in RCA: 50] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/05/2023]
Abstract
Lowering either the administered activity or scan time is desirable in PET imaging as it decreases the patient's radiation burden or improves patient comfort and reduces motion artifacts. But reducing these parameters lowers overall photon counts and increases noise, adversely impacting image contrast and quantification. To address this low count statistics problem, we propose a cycle-consistent generative adversarial network (Cycle GAN) model to estimate diagnostic quality PET images using low count data. Cycle GAN learns a transformation to synthesize diagnostic PET images using low count data that would be indistinguishable from our standard clinical protocol. The algorithm also learns an inverse transformation such that cycle low count PET data (inverse of synthetic estimate) generated from synthetic full count PET is close to the true low count PET. We introduced residual blocks into the generator to catch the differences between low count and full count PET in the training dataset and better handle noise. The average mean error and normalized mean square error in whole body were -0.14% ± 1.43% and 0.52% ± 0.19% with Cycle GAN model, compared to 5.59% ± 2.11% and 3.51% ± 4.14% on the original low count PET images. Normalized cross-correlation is improved from 0.970 to 0.996, and peak signal-to-noise ratio is increased from 39.4 dB to 46.0 dB with proposed method. We developed a deep learning-based approach to accurately estimate diagnostic quality PET datasets from one eighth of photon counts, and has great potential to improve low count PET image quality to the level of diagnostic PET used in clinical settings.
Collapse
Affiliation(s)
- Yang Lei
- Department of Radiation Oncology, Emory University, Atlanta, GA, United States of America. Co-first author
| | | | | | | | | | | | | | | | | |
Collapse
|
26
|
Lu W, Onofrey JA, Lu Y, Shi L, Ma T, Liu Y, Liu C. An investigation of quantitative accuracy for deep learning based denoising in oncological PET. ACTA ACUST UNITED AC 2019; 64:165019. [DOI: 10.1088/1361-6560/ab3242] [Citation(s) in RCA: 60] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022]
|
27
|
Wang Y, Zhou L, Yu B, Wang L, Zu C, Lalush DS, Lin W, Wu X, Zhou J, Shen D. 3D Auto-Context-Based Locality Adaptive Multi-Modality GANs for PET Synthesis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2019; 38:1328-1339. [PMID: 30507527 PMCID: PMC6541547 DOI: 10.1109/tmi.2018.2884053] [Citation(s) in RCA: 81] [Impact Index Per Article: 16.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/08/2023]
Abstract
Positron emission tomography (PET) has been substantially used recently. To minimize the potential health risk caused by the tracer radiation inherent to PET scans, it is of great interest to synthesize the high-quality PET image from the low-dose one to reduce the radiation exposure. In this paper, we propose a 3D auto-context-based locality adaptive multi-modality generative adversarial networks model (LA-GANs) to synthesize the high-quality FDG PET image from the low-dose one with the accompanying MRI images that provide anatomical information. Our work has four contributions. First, different from the traditional methods that treat each image modality as an input channel and apply the same kernel to convolve the whole image, we argue that the contributions of different modalities could vary at different image locations, and therefore a unified kernel for a whole image is not optimal. To address this issue, we propose a locality adaptive strategy for multi-modality fusion. Second, we utilize 1 ×1 ×1 kernel to learn this locality adaptive fusion so that the number of additional parameters incurred by our method is kept minimum. Third, the proposed locality adaptive fusion mechanism is learned jointly with the PET image synthesis in a 3D conditional GANs model, which generates high-quality PET images by employing large-sized image patches and hierarchical features. Fourth, we apply the auto-context strategy to our scheme and propose an auto-context LA-GANs model to further refine the quality of synthesized images. Experimental results show that our method outperforms the traditional multi-modality fusion methods used in deep networks, as well as the state-of-the-art PET estimation approaches.
Collapse
Affiliation(s)
- Yan Wang
- School of Computer Science, Sichuan University, China
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Australia
| | - Biting Yu
- School of Computing and Information Technology, University of Wollongong, Australia
| | - Lei Wang
- School of Computing and Information Technology, University of Wollongong, Australia
| | - Chen Zu
- School of Computing and Information Technology, University of Wollongong, Australia
| | - David S. Lalush
- Department of Biomedical Engineering, University of North Carolina at Chapel Hill and North Carolina State University, USA
| | - Weili Lin
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, USA
| | - Xi Wu
- School of Computer Science, Chengdu University of Information Technology, China
| | - Jiliu Zhou
- School of Computer Science, Chengdu University of Information Technology, China
| | - Dinggang Shen
- IDEA Lab, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, USA
- Department of Brain and Cognitive Engineering, Korea University, Republic of Korea
| |
Collapse
|
28
|
Abstract
PET images often suffer poor signal-to-noise ratio (SNR). Our objective is to improve the SNR of PET images using a deep neural network (DNN) model and MRI images without requiring any higher SNR PET images in training. Our proposed DNN model consists of three modified U-Nets (3U-net). The PET training input data and targets were reconstructed using filtered-backprojection (FBP) and maximum likelihood expectation maximization (MLEM), respectively. FBP reconstruction was used because of its computational efficiency so that the trained network not only removes noise, but also accelerates image reconstruction. Digital brain phantoms downloaded from BrainWeb were used to evaluate the proposed method. Poisson noise was added into sinogram data to simulate a 6 min brain PET scan. Attenuation effect was included and corrected before the image reconstruction. Extra Poisson noise was introduced to the training inputs to improve the network denoising capability. Three independent experiments were conducted to examine the reproducibility. A lesion was inserted into testing data to evaluate the impact of mismatched MRI information using the contrast-to-noise ratio (CNR). The negative impact on noise reduction was also studied when miscoregistration between PET and MRI images occurs. Compared with 1U-net trained with only PET images, training with PET/MRI decreased the mean squared error (MSE) by 31.3% and 34.0% for 1U-net and 3U-net, respectively. The MSE reduction is equivalent to an increase in the count level by 2.5 folds and 2.9 folds for 1U-net and 3U-net, respectively. Compared with the MLEM images, the lesion CNR was improved 2.7 folds and 1.4 folds for 1U-net and 3U-net, respectively. The results show that the proposed method could improve the PET SNR without having higher SNR PET images.
Collapse
Affiliation(s)
- Chih-Chieh Liu
- Department of Biomedical Engineering, University of California, Davis, CA, United States of America
| | | |
Collapse
|
29
|
Zhang Z, Sejdić E. Radiological images and machine learning: Trends, perspectives, and prospects. Comput Biol Med 2019; 108:354-370. [PMID: 31054502 PMCID: PMC6531364 DOI: 10.1016/j.compbiomed.2019.02.017] [Citation(s) in RCA: 63] [Impact Index Per Article: 12.6] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2018] [Revised: 02/19/2019] [Accepted: 02/19/2019] [Indexed: 01/18/2023]
Abstract
The application of machine learning to radiological images is an increasingly active research area that is expected to grow in the next five to ten years. Recent advances in machine learning have the potential to recognize and classify complex patterns from different radiological imaging modalities such as x-rays, computed tomography, magnetic resonance imaging and positron emission tomography imaging. In many applications, machine learning based systems have shown comparable performance to human decision-making. The applications of machine learning are the key ingredients of future clinical decision making and monitoring systems. This review covers the fundamental concepts behind various machine learning techniques and their applications in several radiological imaging areas, such as medical image segmentation, brain function studies and neurological disease diagnosis, as well as computer-aided systems, image registration, and content-based image retrieval systems. Synchronistically, we will briefly discuss current challenges and future directions regarding the application of machine learning in radiological imaging. By giving insight on how take advantage of machine learning powered applications, we expect that clinicians can prevent and diagnose diseases more accurately and efficiently.
Collapse
Affiliation(s)
- Zhenwei Zhang
- Department of Electrical and Computer Engineering, Swanson School of Engineering, University of Pittsburgh, Pittsburgh, PA, 15261, USA
| | - Ervin Sejdić
- Department of Electrical and Computer Engineering, Swanson School of Engineering, University of Pittsburgh, Pittsburgh, PA, 15261, USA.
| |
Collapse
|
30
|
Multi-Modality Cascaded Convolutional Neural Networks for Alzheimer's Disease Diagnosis. Neuroinformatics 2019; 16:295-308. [PMID: 29572601 DOI: 10.1007/s12021-018-9370-4] [Citation(s) in RCA: 137] [Impact Index Per Article: 27.4] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/18/2022]
Abstract
Accurate and early diagnosis of Alzheimer's disease (AD) plays important role for patient care and development of future treatment. Structural and functional neuroimages, such as magnetic resonance images (MRI) and positron emission tomography (PET), are providing powerful imaging modalities to help understand the anatomical and functional neural changes related to AD. In recent years, machine learning methods have been widely studied on analysis of multi-modality neuroimages for quantitative evaluation and computer-aided-diagnosis (CAD) of AD. Most existing methods extract the hand-craft imaging features after image preprocessing such as registration and segmentation, and then train a classifier to distinguish AD subjects from other groups. This paper proposes to construct cascaded convolutional neural networks (CNNs) to learn the multi-level and multimodal features of MRI and PET brain images for AD classification. First, multiple deep 3D-CNNs are constructed on different local image patches to transform the local brain image into more compact high-level features. Then, an upper high-level 2D-CNN followed by softmax layer is cascaded to ensemble the high-level features learned from the multi-modality and generate the latent multimodal correlation features of the corresponding image patches for classification task. Finally, these learned features are combined by a fully connected layer followed by softmax layer for AD classification. The proposed method can automatically learn the generic multi-level and multimodal features from multiple imaging modalities for classification, which are robust to the scale and rotation variations to some extent. No image segmentation and rigid registration are required in pre-processing the brain images. Our method is evaluated on the baseline MRI and PET images of 397 subjects including 93 AD patients, 204 mild cognitive impairment (MCI, 76 pMCI +128 sMCI) and 100 normal controls (NC) from Alzheimer's Disease Neuroimaging Initiative (ADNI) database. Experimental results show that the proposed method achieves an accuracy of 93.26% for classification of AD vs. NC and 82.95% for classification pMCI vs. NC, demonstrating the promising classification performance.
Collapse
|
31
|
Affiliation(s)
- Ciprian Catana
- From the Department of Radiology, Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital and Harvard Medical School, A.A. Martinos Center, 149 13th St, Room 2.301, Charlestown, MA 02129
| |
Collapse
|
32
|
Wang Y, Zhou L, Wang L, Yu B, Zu C, Lalush DS, Lin W, Wu X, Zhou J, Shen D. Locality Adaptive Multi-modality GANs for High-Quality PET Image Synthesis. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2018; 11070:329-337. [PMID: 31058275 PMCID: PMC6494468 DOI: 10.1007/978-3-030-00928-1_38] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/09/2023]
Abstract
Positron emission topography (PET) has been substantially used in recent years. To minimize the potential health risks caused by the tracer radiation inherent to PET scans, it is of great interest to synthesize the high-quality full-dose PET image from the low-dose one to reduce the radiation exposure while maintaining the image quality. In this paper, we propose a locality adaptive multi-modality generative adversarial networks model (LA-GANs) to synthesize the full-dose PET image from both the low-dose one and the accompanying T1-weighted MRI to incorporate anatomical information for better PET image synthesis. This paper has the following contributions. First, we propose a new mechanism to fuse multi-modality information in deep neural networks. Different from the traditional methods that treat each image modality as an input channel and apply the same kernel to convolute the whole image, we argue that the contributions of different modalities could vary at different image locations, and therefore a unified kernel for a whole image is not appropriate. To address this issue, we propose a method that is locality adaptive for multimodality fusion. Second, to learn this locality adaptive fusion, we utilize 1 × 1 × 1 kernel so that the number of additional parameters incurred by our method is kept minimum. This also naturally produces a fused image which acts as a pseudo input for the subsequent learning stages. Third, the proposed locality adaptive fusion mechanism is learned jointly with the PET image synthesis in an end-to-end trained 3D conditional GANs model developed by us. Our 3D GANs model generates high quality PET images by employing large-sized image patches and hierarchical features. Experimental results show that our method outperforms the traditional multi-modality fusion methods used in deep networks, as well as the state-of-the-art PET estimation approaches.
Collapse
Affiliation(s)
- Yan Wang
- School of Computer Science, Sichuan University, Chengdu, China
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Sydney, Australia
| | - Lei Wang
- School of Computing and Information Technology, University of Wollongong, Wollongong, Australia
| | - Biting Yu
- School of Computing and Information Technology, University of Wollongong, Wollongong, Australia
| | - Chen Zu
- School of Computing and Information Technology, University of Wollongong, Wollongong, Australia
| | - David S Lalush
- Joint Department of Biomedical Engineering, University of North Carolina at Chapel Hill and North Carolina State University, Raleigh, NC, USA
| | - Weili Lin
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, USA
| | - Xi Wu
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
| | - Jiliu Zhou
- School of Computer Science, Sichuan University, Chengdu, China
- School of Computer Science, Chengdu University of Information Technology, Chengdu, China
| | - Dinggang Shen
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, USA
| |
Collapse
|
33
|
Wang Y, Yu B, Wang L, Zu C, Lalush DS, Lin W, Wu X, Zhou J, Shen D, Zhou L. 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. Neuroimage 2018; 174:550-562. [PMID: 29571715 DOI: 10.1016/j.neuroimage.2018.03.045] [Citation(s) in RCA: 219] [Impact Index Per Article: 36.5] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2017] [Revised: 03/11/2018] [Accepted: 03/17/2018] [Indexed: 11/29/2022] Open
Abstract
Positron emission tomography (PET) is a widely used imaging modality, providing insight into both the biochemical and physiological processes of human body. Usually, a full dose radioactive tracer is required to obtain high-quality PET images for clinical needs. This inevitably raises concerns about potential health hazards. On the other hand, dose reduction may cause the increased noise in the reconstructed PET images, which impacts the image quality to a certain extent. In this paper, in order to reduce the radiation exposure while maintaining the high quality of PET images, we propose a novel method based on 3D conditional generative adversarial networks (3D c-GANs) to estimate the high-quality full-dose PET images from low-dose ones. Generative adversarial networks (GANs) include a generator network and a discriminator network which are trained simultaneously with the goal of one beating the other. Similar to GANs, in the proposed 3D c-GANs, we condition the model on an input low-dose PET image and generate a corresponding output full-dose PET image. Specifically, to render the same underlying information between the low-dose and full-dose PET images, a 3D U-net-like deep architecture which can combine hierarchical features by using skip connection is designed as the generator network to synthesize the full-dose image. In order to guarantee the synthesized PET image to be close to the real one, we take into account of the estimation error loss in addition to the discriminator feedback to train the generator network. Furthermore, a concatenated 3D c-GANs based progressive refinement scheme is also proposed to further improve the quality of estimated images. Validation was done on a real human brain dataset including both the normal subjects and the subjects diagnosed as mild cognitive impairment (MCI). Experimental results show that our proposed 3D c-GANs method outperforms the benchmark methods and achieves much better performance than the state-of-the-art methods in both qualitative and quantitative measures.
Collapse
Affiliation(s)
- Yan Wang
- School of Computer Science, Sichuan University, China
| | - Biting Yu
- School of Computing and Information Technology, University of Wollongong, Australia
| | - Lei Wang
- School of Computing and Information Technology, University of Wollongong, Australia
| | - Chen Zu
- School of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, China
| | - David S Lalush
- Joint Department of Biomedical Engineering, University of North Carolina at Chapel Hill and North Carolina State University, NC, USA
| | - Weili Lin
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, USA
| | - Xi Wu
- School of Computer Science, Chengdu University of Information Technology, China
| | - Jiliu Zhou
- School of Computer Science, Sichuan University, China; School of Computer Science, Chengdu University of Information Technology, China
| | - Dinggang Shen
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, USA; Department of Brain and Cognitive Engineering, Korea University, Seoul, 02841, South Korea.
| | - Luping Zhou
- School of Electrical and Information Engineering, University of Sydney, Australia; School of Computing and Information Technology, University of Wollongong, Australia.
| |
Collapse
|
34
|
Wang Y, Ma G, Wu X, Zhou J. Patch-Based Label Fusion with Structured Discriminant Embedding for Hippocampus Segmentation. Neuroinformatics 2018; 16:411-423. [PMID: 29512026 DOI: 10.1007/s12021-018-9364-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/17/2022]
Abstract
Automatic and accurate segmentation of hippocampal structures in medical images is of great importance in neuroscience studies. In multi-atlas based segmentation methods, to alleviate the misalignment when registering atlases to the target image, patch-based methods have been widely studied to improve the performance of label fusion. However, weights assigned to the fused labels are usually computed based on predefined features (e.g. image intensities), thus being not necessarily optimal. Due to the lack of discriminating features, the original feature space defined by image intensities may limit the description accuracy. To solve this problem, we propose a patch-based label fusion with structured discriminant embedding method to automatically segment the hippocampal structure from the target image in a voxel-wise manner. Specifically, multi-scale intensity features and texture features are first extracted from the image patch for feature representation. Margin fisher analysis (MFA) is then applied to the neighboring samples in the atlases for the target voxel, in order to learn a subspace in which the distance between intra-class samples is minimized and the distance between inter-class samples is simultaneously maximized. Finally, the k-nearest neighbor (kNN) classifier is employed in the learned subspace to determine the final label for the target voxel. In the experiments, we evaluate our proposed method by conducting hippocampus segmentation using the ADNI dataset. Both the qualitative and quantitative results show that our method outperforms the conventional multi-atlas based segmentation methods.
Collapse
Affiliation(s)
- Yan Wang
- College of Computer Science, Sichuan University, Chengdu, China. .,Fujian Provincial Key Laboratory of Information Processing and Intelligent Control (Minjiang University), Fuzhou, 350121, China.
| | - Guangkai Ma
- Space Control and Inertial Technology Research Center, Harbin Institute of Technology, Harbin, China
| | - Xi Wu
- Department of Computer Science, Chengdu University of Information Technology, Chengdu, China
| | - Jiliu Zhou
- College of Computer Science, Sichuan University, Chengdu, China.,Department of Computer Science, Chengdu University of Information Technology, Chengdu, China
| |
Collapse
|
35
|
Wang Y, Zu C, Hu G, Luo Y, Ma Z, He K, Wu X, Zhou J. Automatic Tumor Segmentation with Deep Convolutional Neural Networks for Radiotherapy Applications. Neural Process Lett 2018. [DOI: 10.1007/s11063-017-9759-3] [Citation(s) in RCA: 25] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
36
|
Yang X, Liu ZY, Qiao H, Song YB, Ren SN, Ji DX, Zheng SW. Underwater image matching by incorporating structural constraints. INT J ADV ROBOT SYST 2017. [DOI: 10.1177/1729881417738100] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Affiliation(s)
- Xu Yang
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, People’s Republic of China
| | - Zhi-Yong Liu
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, People’s Republic of China
- Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, People’s Republic of China
- University of Chinese Academy of Sciences, Beijing, People’s Republic of China
| | - Hong Qiao
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, People’s Republic of China
- Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai, People’s Republic of China
- University of Chinese Academy of Sciences, Beijing, People’s Republic of China
| | - Yong-Bo Song
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, People’s Republic of China
| | - Shu-Nan Ren
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, People’s Republic of China
| | - Da-Xiong Ji
- Ocean College, Zhejiang University, Zhoushan, Peoples’s Republic of China
| | - Sui-Wu Zheng
- State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, People’s Republic of China
- Huizhou Advanced Manufacturing Technology Research Center Co., Ltd, Huizhou, People’s Republic of China
| |
Collapse
|
37
|
Xiang L, Qiao Y, Nie D, An L, Wang Q, Shen D. Deep Auto-context Convolutional Neural Networks for Standard-Dose PET Image Estimation from Low-Dose PET/MRI. Neurocomputing 2017; 267:406-416. [PMID: 29217875 PMCID: PMC5714510 DOI: 10.1016/j.neucom.2017.06.048] [Citation(s) in RCA: 151] [Impact Index Per Article: 21.6] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
Abstract
Positron emission tomography (PET) is an essential technique in many clinical applications such as tumor detection and brain disorder diagnosis. In order to obtain high-quality PET images, a standard-dose radioactive tracer is needed, which inevitably causes the risk of radiation exposure damage. For reducing the patient's exposure to radiation and maintaining the high quality of PET images, in this paper, we propose a deep learning architecture to estimate the high-quality standard-dose PET (SPET) image from the combination of the low-quality low-dose PET (LPET) image and the accompanying T1-weighted acquisition from magnetic resonance imaging (MRI). Specifically, we adapt the convolutional neural network (CNN) to account for the two channel inputs of LPET and T1, and directly learn the end-to-end mapping between the inputs and the SPET output. Then, we integrate multiple CNN modules following the auto-context strategy, such that the tentatively estimated SPET of an early CNN can be iteratively refined by subsequent CNNs. Validations on real human brain PET/MRI data show that our proposed method can provide competitive estimation quality of the PET images, compared to the state-of-the-art methods. Meanwhile, our method is highly efficient to test on a new subject, e.g., spending ~2 seconds for estimating an entire SPET image in contrast to ~16 minutes by the state-of-the-art method. The results above demonstrate the potential of our method in real clinical applications.
Collapse
Affiliation(s)
- Lei Xiang
- Med-X Research Institute, School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Yu Qiao
- Shenzhen key lab of Comp. Vis. & Pat. Rec., Shenzhen Institutes of Advanced Technology, CAS, Shenzhen, China
| | - Dong Nie
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, USA
| | - Le An
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, USA
| | - Qian Wang
- Med-X Research Institute, School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Dinggang Shen
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, USA
- Department of Brain and Cognitive Engineering, Korea University, Seoul 02841, Republic of Korea
| |
Collapse
|
38
|
Abstract
Simultaneous PET-MR imaging improves deficiencies in PET images. The primary areas in which magnetic resonance (MR) has been applied to guide PET results are in correction for patient motion and in improving the effects of PET resolution and partial volume averaging. MR-guided motion correction of PET has been applied to respiratory, cardiac, and gross body movements and shown to improve lesion detectability and contrast. Partial volume correction or resolution improvement of PET governed by MR imaging anatomic information improves visualization of structures and quantitative accuracy. Evaluation in clinical applications is needed to determine their true impacts.
Collapse
Affiliation(s)
- David S Lalush
- Joint Department of Biomedical Engineering, The University of North Carolina, Campus Box 7575, 152 MacNider Hall, Chapel Hill, NC 27599-7575, USA; Joint Department of Biomedical Engineering, North Carolina State University, Campus Box 7115, 911 Oval Drive, Raleigh, NC 27695-7115, USA.
| |
Collapse
|
39
|
Zu C, Zhu L, Zhang D. Iterative sparsity score for feature selection and its extension for multimodal data. Neurocomputing 2017. [DOI: 10.1016/j.neucom.2016.08.124] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
40
|
|
41
|
Wang Y, Ma G, An L, Shi F, Zhang P, Lalush DS, Wu X, Pu Y, Zhou J, Shen D. Semisupervised Tripled Dictionary Learning for Standard-Dose PET Image Prediction Using Low-Dose PET and Multimodal MRI. IEEE Trans Biomed Eng 2016; 64:569-579. [PMID: 27187939 DOI: 10.1109/tbme.2016.2564440] [Citation(s) in RCA: 39] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
OBJECTIVE To obtain high-quality positron emission tomography (PET) image with low-dose tracer injection, this study attempts to predict the standard-dose PET (S-PET) image from both its low-dose PET (L-PET) counterpart and corresponding magnetic resonance imaging (MRI). METHODS It was achieved by patch-based sparse representation (SR), using the training samples with a complete set of MRI, L-PET and S-PET modalities for dictionary construction. However, the number of training samples with complete modalities is often limited. In practice, many samples generally have incomplete modalities (i.e., with one or two missing modalities) that thus cannot be used in the prediction process. In light of this, we develop a semisupervised tripled dictionary learning (SSTDL) method for S-PET image prediction, which can utilize not only the samples with complete modalities (called complete samples) but also the samples with incomplete modalities (called incomplete samples), to take advantage of the large number of available training samples and thus further improve the prediction performance. RESULTS Validation was done on a real human brain dataset consisting of 18 subjects, and the results show that our method is superior to the SR and other baseline methods. CONCLUSION This paper proposed a new S-PET prediction method, which can significantly improve the PET image quality with low-dose injection. SIGNIFICANCE The proposed method is favorable in clinical application since it can decrease the potential radiation risk for patients.
Collapse
|