51
|
Fu Y, Dong S, Niu M, Xue L, Guo H, Huang Y, Xu Y, Yu T, Shi K, Yang Q, Shi Y, Zhang H, Tian M, Zhuo C. AIGAN: Attention-encoding Integrated Generative Adversarial Network for the reconstruction of low-dose CT and low-dose PET images. Med Image Anal 2023; 86:102787. [PMID: 36933386 DOI: 10.1016/j.media.2023.102787] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2022] [Revised: 11/05/2022] [Accepted: 02/22/2023] [Indexed: 03/04/2023]
Abstract
X-ray computed tomography (CT) and positron emission tomography (PET) are two of the most commonly used medical imaging technologies for the evaluation of many diseases. Full-dose imaging for CT and PET ensures the image quality but usually raises concerns about the potential health risks of radiation exposure. The contradiction between reducing the radiation exposure and remaining diagnostic performance can be addressed effectively by reconstructing the low-dose CT (L-CT) and low-dose PET (L-PET) images to the same high-quality ones as full-dose (F-CT and F-PET). In this paper, we propose an Attention-encoding Integrated Generative Adversarial Network (AIGAN) to achieve efficient and universal full-dose reconstruction for L-CT and L-PET images. AIGAN consists of three modules: the cascade generator, the dual-scale discriminator and the multi-scale spatial fusion module (MSFM). A sequence of consecutive L-CT (L-PET) slices is first fed into the cascade generator that integrates with a generation-encoding-generation pipeline. The generator plays the zero-sum game with the dual-scale discriminator for two stages: the coarse and fine stages. In both stages, the generator generates the estimated F-CT (F-PET) images as like the original F-CT (F-PET) images as possible. After the fine stage, the estimated fine full-dose images are then fed into the MSFM, which fully explores the inter- and intra-slice structural information, to output the final generated full-dose images. Experimental results show that the proposed AIGAN achieves the state-of-the-art performances on commonly used metrics and satisfies the reconstruction needs for clinical standards.
Collapse
Affiliation(s)
- Yu Fu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China; Binjiang Institute, Zhejiang University, Hangzhou, China
| | - Shunjie Dong
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Meng Niu
- Department of Radiology, The First Hospital of Lanzhou University, Lanzhou, China
| | - Le Xue
- Department of Nuclear Medicine and Medical PET Center The Second Hospital of Zhejiang University School of Medicine, Hangzhou, China
| | - Hanning Guo
- Institute of Neuroscience and Medicine, Medical Imaging Physics (INM-4), Forschungszentrum Jülich, Jülich, Germany
| | - Yanyan Huang
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Yuanfan Xu
- Hangzhou Universal Medical Imaging Diagnostic Center, Hangzhou, China
| | - Tianbai Yu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Kuangyu Shi
- Department of Nuclear Medicine, University Hospital Bern, Bern, Switzerland
| | - Qianqian Yang
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Yiyu Shi
- Department of Computer Science and Engineering, University of Notre Dame, Notre Dame, IN, USA
| | - Hong Zhang
- Binjiang Institute, Zhejiang University, Hangzhou, China; Department of Nuclear Medicine and Medical PET Center The Second Hospital of Zhejiang University School of Medicine, Hangzhou, China
| | - Mei Tian
- Human Phenome Institute, Fudan University, Shanghai, China.
| | - Cheng Zhuo
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China; Key Laboratory of Collaborative Sensing and Autonomous Unmanned Systems of Zhejiang Province, Hangzhou, China.
| |
Collapse
|
52
|
Chen J, Chen S, Wee L, Dekker A, Bermejo I. Deep learning based unpaired image-to-image translation applications for medical physics: a systematic review. Phys Med Biol 2023; 68. [PMID: 36753766 DOI: 10.1088/1361-6560/acba74] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Accepted: 02/08/2023] [Indexed: 02/10/2023]
Abstract
Purpose. There is a growing number of publications on the application of unpaired image-to-image (I2I) translation in medical imaging. However, a systematic review covering the current state of this topic for medical physicists is lacking. The aim of this article is to provide a comprehensive review of current challenges and opportunities for medical physicists and engineers to apply I2I translation in practice.Methods and materials. The PubMed electronic database was searched using terms referring to unpaired (unsupervised), I2I translation, and medical imaging. This review has been reported in compliance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement. From each full-text article, we extracted information extracted regarding technical and clinical applications of methods, Transparent Reporting for Individual Prognosis Or Diagnosis (TRIPOD) study type, performance of algorithm and accessibility of source code and pre-trained models.Results. Among 461 unique records, 55 full-text articles were included in the review. The major technical applications described in the selected literature are segmentation (26 studies), unpaired domain adaptation (18 studies), and denoising (8 studies). In terms of clinical applications, unpaired I2I translation has been used for automatic contouring of regions of interest in MRI, CT, x-ray and ultrasound images, fast MRI or low dose CT imaging, CT or MRI only based radiotherapy planning, etc Only 5 studies validated their models using an independent test set and none were externally validated by independent researchers. Finally, 12 articles published their source code and only one study published their pre-trained models.Conclusion. I2I translation of medical images offers a range of valuable applications for medical physicists. However, the scarcity of external validation studies of I2I models and the shortage of publicly available pre-trained models limits the immediate applicability of the proposed methods in practice.
Collapse
Affiliation(s)
- Junhua Chen
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, 6229 ET, The Netherlands
| | - Shenlun Chen
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, 6229 ET, The Netherlands
| | - Leonard Wee
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, 6229 ET, The Netherlands
| | - Andre Dekker
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, 6229 ET, The Netherlands
| | - Inigo Bermejo
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, 6229 ET, The Netherlands
| |
Collapse
|
53
|
Chen C, Wang J, Pan J, Bian C, Zhang Z. GraphSKT: Graph-Guided Structured Knowledge Transfer for Domain Adaptive Lesion Detection. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:507-518. [PMID: 36201413 DOI: 10.1109/tmi.2022.3212784] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Adversarial-based adaptation has dominated the area of domain adaptive detection over the past few years. Despite their general efficacy for various tasks, the learned representations may not capture the intrinsic topological structures of the whole images and thus are vulnerable to distributional shifts especially in real-world applications, such as geometric distortions across imaging devices in medical images. In this case, forcefully matching data distributions across domains cannot ensure precise knowledge transfer and are prone to result in the negative transfer. In this paper, we explore the problem of domain adaptive lesion detection from the perspective of relational reasoning, and propose a Graph-Structured Knowledge Transfer (GraphSKT) framework to perform hierarchical reasoning by modeling both the intra- and inter-domain topological structures. To be specific, we utilize cross-domain correspondence to mine meaningful foreground regions for representing graph nodes and explicitly endow each node with contextual information. Then, the intra- and inter-domain graphs are built on the top of instance-level features to achieve a high-level understanding of the lesion and whole medical image, and transfer the structured knowledge from source to target domains. The contextual and semantic information is propagated through graph nodes methodically, enhancing the expressive power of learned features for the lesion detection tasks. Extensive experiments on two types of challenging datasets demonstrate that the proposed GraphSKT significantly outperforms the state-of-the-art approaches for detection of polyps in colonoscopy images and of mass in mammographic images.
Collapse
|
54
|
Lin YC, Lin Y, Huang YL, Ho CY, Chiang HJ, Lu HY, Wang CC, Wang JJ, Ng SH, Lai CH, Lin G. Generalizable transfer learning of automated tumor segmentation from cervical cancers toward a universal model for uterine malignancies in diffusion-weighted MRI. Insights Imaging 2023; 14:14. [PMID: 36690870 PMCID: PMC9871146 DOI: 10.1186/s13244-022-01356-8] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2022] [Accepted: 12/04/2022] [Indexed: 01/25/2023] Open
Abstract
PURPOSE To investigate the generalizability of transfer learning (TL) of automated tumor segmentation from cervical cancers toward a universal model for cervical and uterine malignancies in diffusion-weighted magnetic resonance imaging (DWI). METHODS In this retrospective multicenter study, we analyzed pelvic DWI data from 169 and 320 patients with cervical and uterine malignancies and divided them into the training (144 and 256) and testing (25 and 64) datasets, respectively. A pretrained model was established using DeepLab V3 + from the cervical cancer dataset, followed by TL experiments adjusting the training data sizes and fine-tuning layers. The model performance was evaluated using the dice similarity coefficient (DSC). RESULTS In predicting tumor segmentation for all cervical and uterine malignancies, TL models improved the DSCs from the pretrained cervical model (DSC 0.43) when adding 5, 13, 26, and 51 uterine cases for training (DSC improved from 0.57, 0.62, 0.68, 0.70, p < 0.001). Following the crossover at adding 128 cases (DSC 0.71), the model trained by combining data from adding all the 256 patients exhibited the highest DSCs for the combined cervical and uterine datasets (DSC 0.81) and cervical only dataset (DSC 0.91). CONCLUSIONS TL may improve the generalizability of automated tumor segmentation of DWI from a specific cancer type toward multiple types of uterine malignancies especially in limited case numbers.
Collapse
Affiliation(s)
- Yu-Chun Lin
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan ,grid.145695.a0000 0004 1798 0922Department of Medical Imaging and Radiological Sciences, Chang Gung University, Taoyuan, 33302 Taiwan ,grid.454210.60000 0004 1756 1461Clinical Metabolomics Core Laboratory, Chang Gung Memorial Hospital at Linkou, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Yenpo Lin
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Yen-Ling Huang
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Chih-Yi Ho
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Hsin-Ju Chiang
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan ,grid.454210.60000 0004 1756 1461Clinical Metabolomics Core Laboratory, Chang Gung Memorial Hospital at Linkou, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Hsin-Ying Lu
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan ,grid.454210.60000 0004 1756 1461Clinical Metabolomics Core Laboratory, Chang Gung Memorial Hospital at Linkou, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Chun-Chieh Wang
- grid.145695.a0000 0004 1798 0922Department of Medical Imaging and Radiological Sciences, Chang Gung University, Taoyuan, 33302 Taiwan ,grid.145695.a0000 0004 1798 0922Department of Radiation Oncology, Chang Gung Memorial Hospital at Linkou and Chang Gung University, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Jiun-Jie Wang
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan ,grid.145695.a0000 0004 1798 0922Department of Medical Imaging and Radiological Sciences, Chang Gung University, Taoyuan, 33302 Taiwan
| | - Shu-Hang Ng
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Chyong-Huey Lai
- grid.145695.a0000 0004 1798 0922Gynecologic Cancer Research Center, Department of Obstetrics and Gynecology, Chang Gung Memorial Hospital at Linkou and Chang Gung University, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| | - Gigin Lin
- grid.413801.f0000 0001 0711 0593Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital at Linkou and Keelung, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan ,grid.454210.60000 0004 1756 1461Clinical Metabolomics Core Laboratory, Chang Gung Memorial Hospital at Linkou, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan ,grid.145695.a0000 0004 1798 0922Gynecologic Cancer Research Center, Department of Obstetrics and Gynecology, Chang Gung Memorial Hospital at Linkou and Chang Gung University, 5 Fuhsing St., Guishan, Taoyuan, 33382 Taiwan
| |
Collapse
|
55
|
Xie H, Thorn S, Liu YH, Lee S, Liu Z, Wang G, Sinusas AJ, Liu C. Deep-Learning-Based Few-Angle Cardiac SPECT Reconstruction Using Transformer. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2023; 7:33-40. [PMID: 37397179 PMCID: PMC10312390 DOI: 10.1109/trpms.2022.3187595] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2023]
Abstract
Convolutional neural networks (CNNs) have been extremely successful in various medical imaging tasks. However, because the size of the convolutional kernel used in a CNN is much smaller than the image size, CNN has a strong spatial inductive bias and lacks a global understanding of the input images. Vision Transformer, a recently emerged network structure in computer vision, can potentially overcome the limitations of CNNs for image-reconstruction tasks. In this work, we proposed a slice-by-slice Transformer network (SSTrans-3D) to reconstruct cardiac SPECT images from 3D few-angle data. To be specific, the network reconstructs the whole 3D volume using a slice-by-slice scheme. By doing so, SSTrans-3D alleviates the memory burden required by 3D reconstructions using Transformer. The network can still obtain a global understanding of the image volume with the Transformer attention blocks. Lastly, already reconstructed slices are used as the input to the network so that SSTrans-3D can potentially obtain more informative features from these slices. Validated on porcine, phantom, and human studies acquired using a GE dedicated cardiac SPECT scanner, the proposed method produced images with clearer heart cavity, higher cardiac defect contrast, and more accurate quantitative measurements on the testing data as compared with a deep U-net.
Collapse
Affiliation(s)
| | - Stephanie Thorn
- Department of Internal Medicine (Cardiology) at Yale University
| | - Yi-Hwa Liu
- Department of Internal Medicine (Cardiology) at Yale University
| | - Supum Lee
- Department of Internal Medicine (Cardiology) at Yale University
| | - Zhao Liu
- Department of Radiology and Biomedical Imaging at Yale University
| | - Ge Wang
- Department of Biomedical Engineering at Rensselaer Polytechnic Institute
| | - Albert J Sinusas
- Department of Biomedical Engineering
- Department of Internal Medicine (Cardiology) at Yale University
- Department of Radiology and Biomedical Imaging at Yale University
| | - Chi Liu
- Department of Biomedical Engineering
- Department of Radiology and Biomedical Imaging at Yale University
| |
Collapse
|
56
|
Wang J, Tang Y, Wu Z, Tsui BMW, Chen W, Yang X, Zheng J, Li M. Domain-adaptive denoising network for low-dose CT via noise estimation and transfer learning. Med Phys 2023; 50:74-88. [PMID: 36018732 DOI: 10.1002/mp.15952] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2022] [Revised: 08/19/2022] [Accepted: 08/19/2022] [Indexed: 01/25/2023] Open
Abstract
BACKGROUND In recent years, low-dose computed tomography (LDCT) has played an important role in the diagnosis CT to reduce the potential adverse effects of X-ray radiation on patients, while maintaining the same diagnostic image quality. PURPOSE Deep learning (DL)-based methods have played an increasingly important role in the field of LDCT imaging. However, its performance is highly dependent on the consistency of feature distributions between training data and test data. Due to patient's breathing movements during data acquisition, the paired LDCT and normal dose CT images are difficult to obtain from realistic imaging scenarios. Moreover, LDCT images from simulation or clinical CT examination often have different feature distributions due to the pollution by different amounts and types of image noises. If a network model trained with a simulated dataset is used to directly test clinical patients' LDCT data, its denoising performance may be degraded. Based on this, we propose a novel domain-adaptive denoising network (DADN) via noise estimation and transfer learning to resolve the out-of-distribution problem in LDCT imaging. METHODS To overcome the previous adaptation issue, a novel network model consisting of a reconstruction network and a noise estimation network was designed. The noise estimation network based on a double branch structure is used for image noise extraction and adaptation. Meanwhile, the U-Net-based reconstruction network uses several spatially adaptive normalization modules to fuse multi-scale noise input. Moreover, to facilitate the adaptation of the proposed DADN network to new imaging scenarios, we set a two-stage network training plan. In the first stage, the public simulated dataset is used for training. In the second transfer training stage, we will continue to fine-tune the network model with a torso phantom dataset, while some parameters are frozen. The main reason using the two-stage training scheme is based on the fact that the feature distribution of image content from the public dataset is complex and diverse, whereas the feature distribution of noise pattern from the torso phantom dataset is closer to realistic imaging scenarios. RESULTS In an evaluation study, the trained DADN model is applied to both the public and clinical patient LDCT datasets. Through the comparison of visual inspection and quantitative results, it is shown that the proposed DADN network model can perform well in terms of noise and artifact suppression, while effectively preserving image contrast and details. CONCLUSIONS In this paper, we have proposed a new DL network to overcome the domain adaptation problem in LDCT image denoising. Moreover, the results demonstrate the feasibility and effectiveness of the application of our proposed DADN network model as a new DL-based LDCT image denoising method.
Collapse
Affiliation(s)
- Jiping Wang
- Institute of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, China.,Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, China
| | - Yufei Tang
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, China.,School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, China
| | - Zhongyi Wu
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, China.,School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, China
| | - Benjamin M W Tsui
- Department of Radiology, School of Medicine, Johns Hopkins University, Baltimore, Maryland, USA
| | - Wei Chen
- Minfound Medical Systems Co. Ltd., Shaoxing, Zhejiang, China
| | - Xiaodong Yang
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, China
| | - Jian Zheng
- Institute of Electronic Information Engineering, Changchun University of Science and Technology, Changchun, China.,Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, China.,School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, China
| | - Ming Li
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, China.,School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, China
| |
Collapse
|
57
|
Wang S, Liu Y, Zhang P, Chen P, Li Z, Yan R, Li S, Hou R, Gui Z. Compound feature attention network with edge enhancement for low-dose CT denoising. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2023; 31:915-933. [PMID: 37355934 DOI: 10.3233/xst-230064] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/26/2023]
Abstract
BACKGROUND Low-dose CT (LDCT) images usually contain serious noise and artifacts, which weaken the readability of the image. OBJECTIVE To solve this problem, we propose a compound feature attention network with edge enhancement for LDCT denoising (CFAN-Net), which consists of an edge-enhanced module and a proposed compound feature attention block (CFAB). METHODS The edge enhancement module extracts edge details with the trainable Sobel convolution. CFAB consists of an interactive feature learning module (IFLM), a multi-scale feature fusion module (MFFM), and a joint attention module (JAB), which removes noise from LDCT images in a coarse-to-fine manner. First, in IFLM, the noise is initially removed by cross-latitude interactive judgment learning. Second, in MFFM, multi-scale and pixel attention are integrated to explore fine noise removal. Finally, in JAB, we focus on key information, extract useful features, and improve the efficiency of network learning. To construct a high-quality image, we repeat the above operation by cascading CFAB. RESULTS By applying CFAN-Net to process the 2016 NIH AAPM-Mayo LDCT challenge test dataset, experiments show that the peak signal-to-noise ratio value is 33.9692 and the structural similarity value is 0.9198. CONCLUSIONS Compared with several existing LDCT denoising algorithms, CFAN-Net effectively preserves the texture of CT images while removing noise and artifacts.
Collapse
Affiliation(s)
- Shubin Wang
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Yi Liu
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Pengcheng Zhang
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Ping Chen
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Zhiyuan Li
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Rongbiao Yan
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Shu Li
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Ruifeng Hou
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| | - Zhiguo Gui
- State Key Laboratory of Dynamic Testing Technology, School of Information and Communication Engineering, North University of China, Taiyuan Shanxi Province, China
| |
Collapse
|
58
|
Zhou Z, Huber NR, Inoue A, McCollough CH, Yu L. Multislice input for 2D and 3D residual convolutional neural network noise reduction in CT. J Med Imaging (Bellingham) 2023; 10:014003. [PMID: 36743869 PMCID: PMC9888548 DOI: 10.1117/1.jmi.10.1.014003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2022] [Accepted: 01/09/2023] [Indexed: 02/03/2023] Open
Abstract
Purpose Deep convolutional neural network (CNN)-based methods are increasingly used for reducing image noise in computed tomography (CT). Current attempts at CNN denoising are based on 2D or 3D CNN models with a single- or multiple-slice input. Our work aims to investigate if the multiple-slice input improves the denoising performance compared with the single-slice input and if a 3D network architecture is better than a 2D version at utilizing the multislice input. Approach Two categories of network architectures can be used for the multislice input. First, multislice images can be stacked channel-wise as the multichannel input to a 2D CNN model. Second, multislice images can be employed as the 3D volumetric input to a 3D CNN model, in which the 3D convolution layers are adopted. We make performance comparisons among 2D CNN models with one, three, and seven input slices and two versions of 3D CNN models with seven input slices and one or three output slices. Evaluation was performed on liver CT images using three quantitative metrics with full-dose images as reference. Visual assessment was made by an experienced radiologist. Results When the input channels of the 2D CNN model increases from one to three to seven, a trend of improved performance was observed. Comparing the three models with the seven-slice input, the 3D CNN model with a one-slice output outperforms the other models in terms of noise texture and homogeneity in liver parenchyma as well as subjective visualization of vessels. Conclusions We conclude the that multislice input is an effective strategy for improving performance for 2D deep CNN denoising models. The pure 3D CNN model tends to have a better performance than the other models in terms of continuity across axial slices, but the difference was not significant compared with the 2D CNN model with the same number of slices as the input.
Collapse
Affiliation(s)
- Zhongxing Zhou
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | - Nathan R. Huber
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | - Akitoshi Inoue
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | | | - Lifeng Yu
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| |
Collapse
|
59
|
Zhang Z, Yang M, Li H, Chen S, Wang J, Xu L. An Innovative Low-dose CT Inpainting Algorithm based on Limited-angle Imaging Inpainting Model. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2023; 31:131-152. [PMID: 36373341 DOI: 10.3233/xst-221260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
BACKGROUND With the popularity of computed tomography (CT) technique, an increasing number of patients are receiving CT scans. Simultaneously, the public's attention to CT radiation dose is also increasing. How to obtain CT images suitable for clinical diagnosis while reducing the radiation dose has become the focus of researchers. OBJECTIVE To demonstrate that limited-angle CT imaging technique can be used to acquire lower dose CT images, we propose a generative adversarial network-based image inpainting model-Low-dose imaging and Limited-angle imaging inpainting Model (LDLAIM), this method can effectively restore low-dose CT images with limited-angle imaging, which verifies that limited-angle CT imaging technique can be used to acquire low-dose CT images. METHODS In this work, we used three datasets, including chest and abdomen dataset, head dataset and phantom dataset. They are used to synthesize low-dose and limited-angle CT images for network training. During training stage, we divide each dataset into training set, validation set and testing set according to the ratio of 8:1:1, and use the validation set to validate after finishing an epoch training, and use the testing set to test after finishing all the training. The proposed method is based on generative adversarial networks(GANs), which consists of a generator and a discriminator. The generator consists of residual blocks and encoder-decoder, and uses skip connection. RESULTS We use SSIM, PSNR and RMSE to evaluate the performance of the proposed method. In the chest and abdomen dataset, the mean SSIM, PSNR and RMSE of the testing set are 0.984, 35.385 and 0.017, respectively. In the head dataset, the mean SSIM, PSNR and RMSE of the testing set are 0.981, 38.664 and 0.011, respectively. In the phantom dataset, the mean SSIM, PSNR and RMSE of the testing set are 0.977, 33.468 and 0.022, respectively. By comparing the experimental results of other algorithms in these three datasets, it can be found that the proposed method is superior to other algorithms in these indicators. Meanwhile, the proposed method also achieved the highest score in the subjective quality score. CONCLUSIONS Experimental results show that the proposed method can effectively restore CT images when both low-dose CT imaging techniques and limited-angle CT imaging techniques are used simultaneously. This work proves that the limited-angle CT imaging technique can be used to reduce the CT radiation dose, and also provides a new idea for the research of low-dose CT imaging.
Collapse
Affiliation(s)
- Ziheng Zhang
- Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China
- University of Science and Technology of China, Hefei, Anhui, China
| | - Minghan Yang
- Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China
| | - Huijuan Li
- Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China
- University of Science and Technology of China, Hefei, Anhui, China
| | - Shuai Chen
- Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China
| | - Jianye Wang
- Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei, Anhui, China
| | - Lei Xu
- The First Affiliated Hospitalof University of Science and Technology of China, Hefei, Anhui, China
| |
Collapse
|
60
|
X-ray CT image denoising with MINF: A modularized iterative network framework for data from multiple dose levels. Comput Biol Med 2023; 152:106419. [PMID: 36527781 DOI: 10.1016/j.compbiomed.2022.106419] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2022] [Revised: 11/30/2022] [Accepted: 12/10/2022] [Indexed: 12/14/2022]
Abstract
In clinical applications, multi-dose scan protocols will cause the noise levels of computed tomography (CT) images to fluctuate widely. The popular low-dose CT (LDCT) denoising network outputs denoised images through an end-to-end mapping between an LDCT image and its corresponding ground truth. The limitation of this method is that the reduced noise level of the image may not meet the diagnostic needs of doctors. To establish a denoising model adapted to the multi-noise levels robustness, we proposed a novel and efficient modularized iterative network framework (MINF) to learn the feature of the original LDCT and the outputs of the previous modules, which can be reused in each following module. The proposed network can achieve the goal of gradual denoising, outputting clinical images with different denoising levels, and providing the reviewing physicians with increased confidence in their diagnosis. Moreover, a multi-scale convolutional neural network (MCNN) module is designed to extract as much feature information as possible during the network's training. Extensive experiments on public and private clinical datasets were carried out, and comparisons with several state-of-the-art methods show that the proposed method can achieve satisfactory results for noise suppression of LDCT images. In further comparisons with modularized adaptive processing neural network (MAP-NN), the proposed network shows superior step-by-step or gradual denoising performance. Considering the high quality of gradual denoising results, the proposed method can obtain satisfactory performance in terms of image contrast and detail protection as the level of denoising increases, which shows its potential to be suitable for a multi-dose levels denoising task.
Collapse
|
61
|
Advancing 3D Medical Image Analysis with Variable Dimension Transform based Supervised 3D Pre-training. Neurocomputing 2023. [DOI: 10.1016/j.neucom.2023.01.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/22/2023]
|
62
|
Kobayashi T, Nishii T, Umehara K, Ota J, Ohta Y, Fukuda T, Ishida T. Deep learning-based noise reduction for coronary CT angiography: using four-dimensional noise-reduction images as the ground truth. Acta Radiol 2022; 64:1831-1840. [PMID: 36475893 DOI: 10.1177/02841851221141656] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
Background To assess low-contrast areas such as plaque and coronary artery stenosis, coronary computed tomography angiography (CCTA) needs to provide images with lower noise without increasing radiation doses. Purpose To develop a deep learning-based noise-reduction method for CCTA using four-dimensional noise reduction (4DNR) as the ground truth for supervised learning. Material and Methods \We retrospectively collected 100 retrospective ECG-gated CCTAs. We created 4DNR images using non-rigid registration and weighted averaging three timeline CCTA volumetric data with intervals of 50 ms in the mid-diastolic phase. Our method set the original reconstructed image as the input and the 4DNR as the target image and obtained the noise-reduced image via residual learning. We evaluated the objective image quality of the original and deep learning-based noise-reduction (DLNR) images based on the image noise of the aorta and the contrast-to-noise ratio (CNR) of the coronary arteries. Further, a board-certified radiologist evaluated the blurring of several heart structures using a 5-point Likert scale subjectively and assigned a coronary artery disease reporting and data system (CAD-RADS) category independently. Results DLNR CCTAs showed 64.5% lower image noise ( P < 0.001) and achieved a 2.9 times higher CNR of coronary arteries than that in original images, without significant blurring in subjective comparison ( P > 0.1). The intra-observer agreement of CAD-RADS in the DLNR image was excellent (0.87, 95% confidence interval = 0.77–0.99) with original CCTAs. Conclusion Our DLNR method supervised by 4DNR significantly reduced the image noise of CCTAs without affecting the assessment of coronary stenosis.
Collapse
Affiliation(s)
- Takuma Kobayashi
- Department of Medical Physics and Engineering, Graduate School of Medicine, Osaka University, Suita, Japan
- Department of Radiology, National Cerebral and Cardiovascular Center, Suita, Japan
| | - Tatsuya Nishii
- Department of Radiology, National Cerebral and Cardiovascular Center, Suita, Japan
| | - Kensuke Umehara
- Department of Medical Physics and Engineering, Graduate School of Medicine, Osaka University, Suita, Japan
- Medical Informatics Section, Department of Medical Technology, QST Hospital, National Institutes for Quantum Science and Technology (QST), Chiba, Japan
- Applied MRI Research, Department of Molecular Imaging and Theranostics, Institute for Quantum Medical Science, National Institutes for Quantum Science and Technology, Chiba, Japan
| | - Junko Ota
- Department of Medical Physics and Engineering, Graduate School of Medicine, Osaka University, Suita, Japan
- Medical Informatics Section, Department of Medical Technology, QST Hospital, National Institutes for Quantum Science and Technology (QST), Chiba, Japan
- Applied MRI Research, Department of Molecular Imaging and Theranostics, Institute for Quantum Medical Science, National Institutes for Quantum Science and Technology, Chiba, Japan
| | - Yasutoshi Ohta
- Department of Radiology, National Cerebral and Cardiovascular Center, Suita, Japan
| | - Tetsuya Fukuda
- Department of Radiology, National Cerebral and Cardiovascular Center, Suita, Japan
| | - Takayuki Ishida
- Department of Medical Physics and Engineering, Graduate School of Medicine, Osaka University, Suita, Japan
| |
Collapse
|
63
|
Li D, Bian Z, Li S, He J, Zeng D, Ma J. Noise Characteristics Modeled Unsupervised Network for Robust CT Image Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:3849-3861. [PMID: 35939459 DOI: 10.1109/tmi.2022.3197400] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Deep learning (DL)-based methods show great potential in computed tomography (CT) imaging field. The DL-based reconstruction methods are usually evaluated on the training and testing datasets which are obtained from the same distribution, i.e., the same CT scan protocol (i.e., the region setting, kVp, mAs, etc.). In this work, we focus on analyzing the robustness of the DL-based methods against protocol-specific distribution shifts (i.e., the training and testing datasets are from different region settings, different kVp settings, or different mAs settings, respectively). The results show that the DL-based reconstruction methods are sensitive to the protocol-specific perturbations which can be attributed to the noise distribution shift between the training and testing datasets. Based on these findings, we presented a low-dose CT reconstruction method using an unsupervised strategy with the consideration of noise distribution to address the issue of protocol-specific perturbations. Specifically, unpaired sinogram data is enrolled into the network training, which represents unique information for specific imaging protocol, and a Gaussian mixture model (GMM) is introduced to characterize the noise distribution in CT images. It can be termed as GMM based unsupervised CT reconstruction network (GMM-unNet) method. Moreover, an expectation-maximization algorithm is designed to optimize the presented GMM-unNet method. Extensive experiments are performed on three datasets from different scan protocols, which demonstrate that the presented GMM-unNet method outperforms the competing methods both qualitatively and quantitatively.
Collapse
|
64
|
Image denoising in the deep learning era. Artif Intell Rev 2022. [DOI: 10.1007/s10462-022-10305-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
65
|
Zhang Y, He W, Chen F, Wu J, He Y, Xu Z. Denoise ultra-low-field 3D magnetic resonance images using a joint signal-image domain filter. JOURNAL OF MAGNETIC RESONANCE (SAN DIEGO, CALIF. : 1997) 2022; 344:107319. [PMID: 36332511 DOI: 10.1016/j.jmr.2022.107319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2022] [Revised: 09/17/2022] [Accepted: 10/18/2022] [Indexed: 06/16/2023]
Abstract
Ultra-low-field magnetic resonance imaging (MRI) could suffer from heavy uncorrelated noise, and its removal could be a critical post-processing task. As a primary source of interference, Gaussian noise could corrupt the sampled MR signal (k-space data), especially at lower B0 field strength. For this reason, we consider both signal and image domains by proposing a new joint filter characterized by a Kalman filter with linear prediction and a nonlocal mean filter with higher-order singular value decomposition (HOSVD) for denoising 3D MR data. The Kalman filter first attenuates the noise in k-space, and then its reconstruction images are used to guide HOSVD denoising process with exploring self-similarity among 3D structures. The clearer prefiltered images could also generate improved HOSVD learned bases used to transform the noise corrupted patch groups in the original MR data. The flexibility of proposed method is also demonstrated by integrating other k-space filters into the algorithm scheme. Experimental data includes simulated MR images with the varying noise level and real MR images obtained from our 50 mT MRI scanner. The results reveal that our method has a better noise-removal ability and introduces lesser unexpected artifacts than other related MRI denoising approaches.
Collapse
Affiliation(s)
- Yuxiang Zhang
- State Key Laboratory of Power Transmission Equipment and System Security and New Technology, Chongqing University, Chongqing, China
| | - Wei He
- State Key Laboratory of Power Transmission Equipment and System Security and New Technology, Chongqing University, Chongqing, China
| | - Fangge Chen
- State Key Laboratory of Power Transmission Equipment and System Security and New Technology, Chongqing University, Chongqing, China
| | - Jiamin Wu
- Shenzhen Academy of Aerospace Technology, Shenzhen, China; Harbin Institute of Technology, Harbin, China
| | - Yucheng He
- Shenzhen Academy of Aerospace Technology, Shenzhen, China
| | - Zheng Xu
- State Key Laboratory of Power Transmission Equipment and System Security and New Technology, Chongqing University, Chongqing, China.
| |
Collapse
|
66
|
Liu H, Liao P, Chen H, Zhang Y. ERA-WGAT: Edge-enhanced residual autoencoder with a window-based graph attention convolutional network for low-dose CT denoising. BIOMEDICAL OPTICS EXPRESS 2022; 13:5775-5793. [PMID: 36733738 PMCID: PMC9872905 DOI: 10.1364/boe.471340] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/25/2022] [Revised: 09/03/2022] [Accepted: 09/19/2022] [Indexed: 06/18/2023]
Abstract
Computed tomography (CT) has become a powerful tool for medical diagnosis. However, minimizing X-ray radiation risk for the patient poses significant challenges to obtain suitable low dose CT images. Although various low-dose CT methods using deep learning techniques have produced impressive results, convolutional neural network based methods focus more on local information and hence are very limited for non-local information extraction. This paper proposes ERA-WGAT, a residual autoencoder incorporating an edge enhancement module that performs convolution with eight types of learnable operators providing rich edge information and a window-based graph attention convolutional network that combines static and dynamic attention modules to explore non-local self-similarity. We use the compound loss function that combines MSE loss and multi-scale perceptual loss to mitigate the over-smoothing problem. Compared with current low-dose CT denoising methods, ERA-WGAT confirmed superior noise suppression and perceived image quality.
Collapse
Affiliation(s)
- Han Liu
- College of Computer Science, Sichuan University, Chengdu 610065, China
| | - Peixi Liao
- Department of Scientific Research and Education, The Sixth People’s Hospital of Chengdu, Chengdu 610051, China
| | - Hu Chen
- College of Computer Science, Sichuan University, Chengdu 610065, China
| | - Yi Zhang
- College of Computer Science, Sichuan University, Chengdu 610065, China
| |
Collapse
|
67
|
A plan verification platform for online adaptive proton therapy using deep learning-based Monte–Carlo denoising. Phys Med 2022; 103:18-25. [DOI: 10.1016/j.ejmp.2022.09.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/12/2022] [Revised: 09/27/2022] [Accepted: 09/28/2022] [Indexed: 11/21/2022] Open
|
68
|
Xiao D, Pitman WMK, Yiu BYS, Chee AJY, Yu ACH. Minimizing Image Quality Loss After Channel Count Reduction for Plane Wave Ultrasound via Deep Learning Inference. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2022; 69:2849-2861. [PMID: 35862334 DOI: 10.1109/tuffc.2022.3192854] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
High-frame-rate ultrasound imaging uses unfocused transmissions to insonify an entire imaging view for each transmit event, thereby enabling frame rates over 1000 frames per second (fps). At these high frame rates, it is naturally challenging to realize real-time transfer of channel-domain raw data from the transducer to the system back end. Our work seeks to halve the total data transfer rate by uniformly decimating the receive channel count by 50% and, in turn, doubling the array pitch. We show that despite the reduced channel count and the inevitable use of a sparse array aperture, the resulting beamformed image quality can be maintained by designing a custom convolutional encoder-decoder neural network to infer the radio frequency (RF) data of the nullified channels. This deep learning framework was trained with in vivo human carotid data (5-MHz plane wave imaging, 128 channels, 31 steering angles over a 30° span, and 62 799 frames in total). After training, the network was tested on an in vitro point target scenario that was dissimilar to the training data, in addition to in vivo carotid validation datasets. In the point target phantom image beamformed from inferred channel data, spatial aliasing artifacts attributed to array pitch doubling were found to be reduced by up to 10 dB. For carotid imaging, our proposed approach yielded a lumen-to-tissue contrast that was on average within 3 dB compared to the full-aperture image, whereas without channel data inferencing, the carotid lumen was obscured. When implemented on an RTX-2080 GPU, the inference time to apply the trained network was 4 ms, which favors real-time imaging. Overall, our technique shows that with the help of deep learning, channel data transfer rates can be effectively halved with limited impact on the resulting image quality.
Collapse
|
69
|
Kim W, Lee J, Kang M, Kim JS, Choi JH. Wavelet subband-specific learning for low-dose computed tomography denoising. PLoS One 2022; 17:e0274308. [PMID: 36084002 PMCID: PMC9462582 DOI: 10.1371/journal.pone.0274308] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2022] [Accepted: 08/25/2022] [Indexed: 11/19/2022] Open
Abstract
Deep neural networks have shown great improvements in low-dose computed tomography (CT) denoising. Early algorithms were primarily optimized to obtain an accurate image with low distortion between the denoised image and reference full-dose image at the cost of yielding an overly smoothed unrealistic CT image. Recent research has sought to preserve the fine details of denoised images with high perceptual quality, which has been accompanied by a decrease in objective quality due to a trade-off between perceptual quality and distortion. We pursue a network that can generate accurate and realistic CT images with high objective and perceptual quality within one network, achieving a better perception-distortion trade-off. To achieve this goal, we propose a stationary wavelet transform-assisted network employing the characteristics of high- and low-frequency domains of the wavelet transform and frequency subband-specific losses defined in the wavelet domain. We first introduce a stationary wavelet transform for the network training procedure. Then, we train the network using objective loss functions defined for high- and low-frequency domains to enhance the objective quality of the denoised CT image. With this network design, we train the network again after replacing the objective loss functions with perceptual loss functions in high- and low-frequency domains. As a result, we acquired denoised CT images with high perceptual quality using this strategy while minimizing the objective quality loss. We evaluated our algorithms on the phantom and clinical images, and the quantitative and qualitative results indicate that ours outperform the existing state-of-the-art algorithms in terms of objective and perceptual quality.
Collapse
Affiliation(s)
- Wonjin Kim
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| | - Jaayeon Lee
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| | - Mihyun Kang
- Department of Cyber Security, Ewha Womans University, Seoul, Republic of Korea
| | - Jin Sung Kim
- Department of Radiation Oncology, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Jang-Hwan Choi
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| |
Collapse
|
70
|
Liu H, Jin X, Liu L, Jin X. Low-Dose CT Image Denoising Based on Improved DD-Net and Local Filtered Mechanism. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:2692301. [PMID: 35965772 PMCID: PMC9365583 DOI: 10.1155/2022/2692301] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/20/2022] [Revised: 06/02/2022] [Accepted: 06/20/2022] [Indexed: 11/18/2022]
Abstract
Low-dose CT (LDCT) images can reduce the radiation damage to the patients; however, the unavoidable information loss will influence the clinical diagnosis under low-dose conditions, such as noise, streak artifacts, and smooth details. LDCT image denoising is a significant topic in medical image processing to overcome the above deficits. This work proposes an improved DD-Net (DenseNet and deconvolution-based network) joint local filtered mechanism, the DD-Net is enhanced by introducing improved residual dense block to strengthen the feature representation ability, and the local filtered mechanism and gradient loss are also employed to effectively restore the subtle structures. First, the LDCT image is inputted into the network to obtain the denoised image. The original loss between the denoised image and normal-dose CT (NDCT) image is calculated, and the difference image between the NDCT image and the denoised image is obtained. Second, a mask image is generated by taking a threshold operation to the difference image, and the filtered LDCT and NDCT images are obtained by conducting an elementwise multiplication operation with LDCT and NDCT images using the mask image. Third, the filtered image is inputted into the network to obtain the filtered denoised image, and the correction loss is calculated. At last, the sum of original loss and correction loss of the improved DD-Net is used to optimize the network. Considering that it is insufficient to generate the edge information using the combination of mean square error (MSE) and multiscale structural similarity (MS-SSIM), we introduce the gradient loss that can calculate the loss of the high-frequency portion. The experimental results show that the proposed method can achieve better performance than conventional schemes and most neural networks. Our source code is made available at https://github.com/LHE-IT/Low-dose-CT-Image-Denoising/tree/main/Local Filtered Mechanism.
Collapse
Affiliation(s)
- Hongen Liu
- School of Software, Yunnan University, Kunming 650091, Yunnan, China
| | - Xin Jin
- School of Software, Yunnan University, Kunming 650091, Yunnan, China
| | - Ling Liu
- School of Software, Yunnan University, Kunming 650091, Yunnan, China
| | - Xin Jin
- School of Software, Yunnan University, Kunming 650091, Yunnan, China
- Engineering Research Center of Cyberspace, Yunnan University, Kunming 650000, Yunnan, China
| |
Collapse
|
71
|
Liu J, Jiang H, Ning F, Li M, Pang W. DFSNE-Net: Deviant feature sensitive noise estimate network for low-dose CT denoising. Comput Biol Med 2022; 149:106061. [DOI: 10.1016/j.compbiomed.2022.106061] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Revised: 08/10/2022] [Accepted: 08/27/2022] [Indexed: 11/26/2022]
|
72
|
Zhang JZ, Ganesh H, Raslau FD, Nair R, Escott E, Wang C, Wang G, Zhang J. Deep learning versus iterative reconstruction on image quality and dose reduction in abdominal CT: a live animal study. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac7999] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2022] [Accepted: 06/16/2022] [Indexed: 11/12/2022]
Abstract
Abstract
Objective. While simulated low-dose CT images and phantom studies cannot fully approximate subjective and objective effects of deep learning (DL) denoising on image quality, live animal models may afford this assessment. This study is to investigate the potential of DL in CT dose reduction on image quality compared to iterative reconstruction (IR). Approach. The upper abdomen of a live 4 year old sheep was scanned on a CT scanner at different exposure levels. Images were reconstructed using FBP and ADMIRE with 5 strengths. A modularized DL network with 5 modules was used for image reconstruction via progressive denoising. Radiomic features were extracted from a region over the liver. Concordance correlation coefficient (CCC) was applied to quantify agreement between any two sets of radiomic features. Coefficient of variation was calculated to measure variation in a radiomic feature series. Structural similarity index (SSIM) was used to measure the similarity between any two images. Diagnostic quality, low-contrast detectability, and image texture were qualitatively evaluated by two radiologists. Pearson correlation coefficient was computed across all dose-reconstruction/denoising combinations. Results. A total of 66 image sets, with 405 radiomic features extracted from each, are analyzed. IR and DL can improve diagnostic quality and low-contrast detectability and similarly modulate image texture features. In terms of SSIM, DL has higher potential in preserving image structure. There is strong correlation between SSIM and radiologists’ evaluations for diagnostic quality (0.559) and low-contrast detectability (0.635) but moderate correlation for texture (0.313). There is moderate correlation between CCC of radiomic features and radiologists’ evaluation for diagnostic quality (0.397), low-contrast detectability (0.417), and texture (0.326), implying that improvement of image features may not relate to improvement of diagnostic quality. Conclusion. DL shows potential to further reduce radiation dose while preserving structural similarity, while IR is favored by radiologists and more predictably alters radiomic features.
Collapse
|
73
|
Zhu J, Su T, Zhang X, Yang J, Mi D, Zhang Y, Gao X, Zheng H, Liang D, Ge Y. Feasibility study of three-material decomposition in dual-energy cone-beam CT imaging with deep learning. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac7b09] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2021] [Accepted: 06/21/2022] [Indexed: 11/11/2022]
Abstract
Abstract
Objective. In this work, a dedicated end-to-end deep convolutional neural network, named as Triple-CBCT, is proposed to demonstrate the feasibility of reconstructing three different material distribution volumes from the dual-energy CBCT projection data. Approach. In Triple-CBCT, the features of the sinogram and the CT image are independently extracted and cascaded via a customized domain transform network module. This Triple-CBCT network was trained by numerically synthesized dual-energy CBCT data, and was tested with experimental dual-energy CBCT data of the Iodine-CaCl2 solution and pig leg specimen scanned on an in-house benchtop system. Main results. Results show that the information stored in both the sinogram and CT image domains can be used together to improve the decomposition quality of multiple materials (water, iodine, CaCl2 or bone) from the dual-energy projections. In addition, both the numerical and experimental results demonstrate that the Triple-CBCT is able to generate high-fidelity dual-energy CBCT basis images. Significance. An innovative end-to-end network that joints the sinogram and CT image domain information is developed to facilitate high quality automatic decomposition from the dual-energy CBCT scans.
Collapse
|
74
|
Ma R, Hu J, Sari H, Xue S, Mingels C, Viscione M, Kandarpa VSS, Li WB, Visvikis D, Qiu R, Rominger A, Li J, Shi K. An encoder-decoder network for direct image reconstruction on sinograms of a long axial field of view PET. Eur J Nucl Med Mol Imaging 2022; 49:4464-4477. [PMID: 35819497 DOI: 10.1007/s00259-022-05861-2] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2021] [Accepted: 06/02/2022] [Indexed: 11/04/2022]
Abstract
PURPOSE Deep learning is an emerging reconstruction method for positron emission tomography (PET), which can tackle complex PET corrections in an integrated procedure. This paper optimizes the direct PET reconstruction from sinogram on a long axial field of view (LAFOV) PET. METHODS This paper proposes a novel deep learning architecture to reduce the biases during direct reconstruction from sinograms to images. This architecture is based on an encoder-decoder network, where the perceptual loss is used with pre-trained convolutional layers. It is trained and tested on data of 80 patients acquired from recent Siemens Biograph Vision Quadra long axial FOV (LAFOV) PET/CT. The patients are randomly split into a training dataset of 60 patients, a validation dataset of 10 patients, and a test dataset of 10 patients. The 3D sinograms are converted into 2D sinogram slices and used as input to the network. In addition, the vendor reconstructed images are considered as ground truths. Finally, the proposed method is compared with DeepPET, a benchmark deep learning method for PET reconstruction. RESULTS Compared with DeepPET, the proposed network significantly reduces the root-mean-squared error (NRMSE) from 0.63 to 0.6 (p < 0.01) and increases the structural similarity index (SSIM) and peak signal-to-noise ratio (PSNR) from 0.93 to 0.95 (p < 0.01) and from 82.02 to 82.36 (p < 0.01), respectively. The reconstruction time is approximately 10 s per patient, which is shortened by 23 times compared with the conventional method. The errors of mean standardized uptake values (SUVmean) for lesions between ground truth and the predicted result are reduced from 33.5 to 18.7% (p = 0.03). In addition, the error of max SUV is reduced from 32.7 to 21.8% (p = 0.02). CONCLUSION The results demonstrate the feasibility of using deep learning to reconstruct images with acceptable image quality and short reconstruction time. It is shown that the proposed method can improve the quality of deep learning-based reconstructed images without additional CT images for attenuation and scattering corrections. This study demonstrated the feasibility of deep learning to rapidly reconstruct images without additional CT images for complex corrections from actual clinical measurements on LAFOV PET. Despite improving the current development, AI-based reconstruction does not work appropriately for untrained scenarios due to limited extrapolation capability and cannot completely replace conventional reconstruction currently.
Collapse
Affiliation(s)
- Ruiyao Ma
- Department of Engineering Physics, Tsinghua University, and Key Laboratory of Particle & Radiation Imaging (Tsinghua University), Ministry of Education, Beijing, 100084, China.,Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland.,Institute of Radiation Medicine, Helmholtz Zentrum München German Research Center for Environmental Health (GmbH), Bavaria, Neuherberg, Germany
| | - Jiaxi Hu
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Hasan Sari
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland.,Advanced Clinical Imaging Technology, Siemens Healthcare AG, Lausanne, Switzerland
| | - Song Xue
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Clemens Mingels
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Marco Viscione
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | | | - Wei Bo Li
- Institute of Radiation Medicine, Helmholtz Zentrum München German Research Center for Environmental Health (GmbH), Bavaria, Neuherberg, Germany
| | | | - Rui Qiu
- Department of Engineering Physics, Tsinghua University, and Key Laboratory of Particle & Radiation Imaging (Tsinghua University), Ministry of Education, Beijing, 100084, China.
| | - Axel Rominger
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| | - Junli Li
- Department of Engineering Physics, Tsinghua University, and Key Laboratory of Particle & Radiation Imaging (Tsinghua University), Ministry of Education, Beijing, 100084, China.
| | - Kuangyu Shi
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
| |
Collapse
|
75
|
Huang Z, Zhao R, Leung FHF, Banerjee S, Lee TTY, Yang D, Lun DPK, Lam KM, Zheng YP, Ling SH. Joint Spine Segmentation and Noise Removal From Ultrasound Volume Projection Images With Selective Feature Sharing. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1610-1624. [PMID: 35041596 DOI: 10.1109/tmi.2022.3143953] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Volume Projection Imaging from ultrasound data is a promising technique to visualize spine features and diagnose Adolescent Idiopathic Scoliosis. In this paper, we present a novel multi-task framework to reduce the scan noise in volume projection images and to segment different spine features simultaneously, which provides an appealing alternative for intelligent scoliosis assessment in clinical applications. Our proposed framework consists of two streams: i) A noise removal stream based on generative adversarial networks, which aims to achieve effective scan noise removal in a weakly-supervised manner, i.e., without paired noisy-clean samples for learning; ii) A spine segmentation stream, which aims to predict accurate bone masks. To establish the interaction between these two tasks, we propose a selective feature-sharing strategy to transfer only the beneficial features, while filtering out the useless or harmful information. We evaluate our proposed framework on both scan noise removal and spine segmentation tasks. The experimental results demonstrate that our proposed method achieves promising performance on both tasks, which provides an appealing approach to facilitating clinical diagnosis.
Collapse
|
76
|
Choi K. A Comparative Study between Image- and Projection-Domain Self-Supervised Learning for Ultra Low-Dose CBCT. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:2076-2079. [PMID: 36085987 DOI: 10.1109/embc48229.2022.9871947] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
We compare image domain and projection domain denoising approaches with self-supervised learning for ultra low-dose cone-beam CT (CBCT), where number of detected x-ray photons is significantly low. For image-domain self-supervised denoising, we first reconstruct CBCT images with the standard filtered backprojection. For model training, we use blind-spot filtering to partially blind images and recover the blind spots. For projection-domain self-supervised denoising, we regard the post-log projections as training examples of convolutional neural network. From experimental results with various low-dose CBCT settings, the projection-domain denoiser outperforms the image-domain denoiser both in image quality and accuracy for ultra low-dose CBCT.
Collapse
|
77
|
Guangzhe Z, Yimeng Z, Ge M, Min Y. Bilateral U‐Net semantic segmentation with spatial attention mechanism. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY 2022. [DOI: 10.1049/cit2.12118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022] Open
Affiliation(s)
- Zhao Guangzhe
- Beijing University of Civil Engineering and Architecture College of Electrical and Information Engineering Beijing China
| | - Zhang Yimeng
- Beijing University of Civil Engineering and Architecture College of Electrical and Information Engineering Beijing China
| | - Maoning Ge
- Graduate School of Informatics Nagoya University Nagoya Japan
| | - Yu Min
- Department of General Surgery Guangdong Provincial People's Hospital Guangdong Academy of Medical Sciences Guangzhou China
| |
Collapse
|
78
|
CCN-CL: A content-noise complementary network with contrastive learning for low-dose computed tomography denoising. Comput Biol Med 2022; 147:105759. [PMID: 35752116 DOI: 10.1016/j.compbiomed.2022.105759] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2022] [Revised: 05/26/2022] [Accepted: 06/18/2022] [Indexed: 11/20/2022]
Abstract
In recent years, low-dose computed tomography (LDCT) has played an increasingly important role in the diagnosis CT to reduce the potential adverse effects of x-ray radiation on patients while maintaining the same diagnostic image quality. Current deep learning-based denoising methods applied to LDCT imaging only use normal dose CT (NDCT) images as positive examples to guide the denoising process. Recent studies on contrastive learning have proved that the original images as negative examples can also be helpful for network learning. Therefore, this paper proposes a novel content-noise complementary network with contrastive learning for an LDCT denoising task. First, to better train our proposed network, a contrastive learning loss, taking the NDCT image as a positive example and the original LDCT image as a negative example to guide the network learning is added. Furthermore, we also design a network structure that combines content-noise complementary learning strategy, attention mechanism, and deformable convolution for better network performance. In an evaluation study, we compare the performance of our designed network with some of the state-of-the-art methods in the 2016 NIH-AAPM-Mayo Clinic Low Dose CT Grand Challenge dataset. The quantitative and qualitative evaluation results demonstrate the feasibility and effectiveness of applying our proposed CCN-CL network model as a new deep learning-based LDCT denoising method.
Collapse
|
79
|
Wu Q, Tang H, Liu H, Chen YC. Masked Joint Bilateral Filtering via Deep Image Prior for Digital X-ray Image Denoising. IEEE J Biomed Health Inform 2022; 26:4008-4019. [PMID: 35653453 DOI: 10.1109/jbhi.2022.3179652] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
Medical image denoising faces great challenges. Although deep learning methods have shown great potential, their efficiency is severely affected by millions of trainable parameters. The non-linearity of neural networks also makes them difficult to be understood. Therefore, existing deep learning methods have been sparingly applied to clinical tasks. To this end, we integrate known filtering operators into deep learning and propose a novel Masked Joint Bilateral Filtering (MJBF) via deep image prior for digital X-ray image denoising. Specifically, MJBF consists of a deep image prior generator and an iterative filtering block. The deep image prior generator produces plentiful image priors by a multi-scale fusion network. The generated image priors serve as the guidance for the iterative filtering block, which is utilized for the actual edge-preserving denoising. The iterative filtering block contains three trainable Joint Bilateral Filters (JBFs), each with only 18 trainable parameters. Moreover, a masking strategy is introduced to reduce redundancy and improve the understanding of the proposed network. Experimental results on the ChestX-ray14 dataset and real data show that the proposed MJBF has achieved superior performance in terms of noise suppression and edge preservation. Tests on the portability of the proposed method demonstrate that this denoising modality is simple yet effective, and could have a clinical impact on medical imaging in the future.
Collapse
|
80
|
Liu J, Kang Y, Xia Z, Qiang J, Zhang J, Zhang Y, Chen Y. MRCON-Net: Multiscale reweighted convolutional coding neural network for low-dose CT imaging. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 221:106851. [PMID: 35576686 DOI: 10.1016/j.cmpb.2022.106851] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/30/2021] [Revised: 03/28/2022] [Accepted: 04/30/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Low-dose computed tomography (LDCT) has become increasingly important for alleviating X-ray radiation damage. However, reducing the administered radiation dose may lead to degraded CT images with amplified mottle noise and nonstationary streak artifacts. Previous studies have confirmed that deep learning (DL) is promising for improving LDCT imaging. However, most DL-based frameworks are built intuitively, lack interpretability, and suffer from image detail information loss, which has become a general challenging issue. METHODS A multiscale reweighted convolutional coding neural network (MRCON-Net) is developed to address the above problems. MRCON-Net is compact and more explainable than other networks. First, inspired by the learning-based reweighted iterative soft thresholding algorithm (ISTA), we extend traditional convolutional sparse coding (CSC) to its reweighted convolutional learning form. Second, we use dilated convolution to extract multiscale image features, allowing our single model to capture the correlations between features of different scales. Finally, to automatically adjust the elements in the feature code to correct the obtained solution, a channel attention (CA) mechanism is utilized to learn appropriate weights. RESULTS The visual results obtained based on the American Association of Physicians in Medicine (AAPM) Challenge and United Image Healthcare (UIH) clinical datasets confirm that the proposed model significantly reduces serious artifact noise while retaining the desired structures. Quantitative results show that the average structural similarity index measurement (SSIM) and peak signal-to-noise ratio (PSNR) achieved on the AAPM Challenge dataset are 0.9491 and 40.66, respectively, and the SSIM and PSNR achieved on the UIH clinical dataset are 0.915 and 42.44, respectively; these are promising quantitative results. CONCLUSION Compared with recent state-of-the-art methods, the proposed model achieves subtle structure-enhanced LDCT imaging. In addition, through ablation studies, the components of the proposed model are validated to achieve performance improvements.
Collapse
Affiliation(s)
- Jin Liu
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China; Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China.
| | - Yanqin Kang
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China; Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China
| | - Zhenyu Xia
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China
| | - Jun Qiang
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China
| | - JunFeng Zhang
- School of Computer and Information Engineering, Henan University of Economics and Law, Zhengzhou, China
| | - Yikun Zhang
- Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China; School of Cyber Science and Engineering, Southeast University, Nanjing, China; School of Computer Science and Engineering, Southeast University, Nanjing, China
| | - Yang Chen
- Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China; School of Cyber Science and Engineering, Southeast University, Nanjing, China; School of Computer Science and Engineering, Southeast University, Nanjing, China
| |
Collapse
|
81
|
The use of deep learning methods in low-dose computed tomography image reconstruction: a systematic review. COMPLEX INTELL SYST 2022. [DOI: 10.1007/s40747-022-00724-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
Abstract
AbstractConventional reconstruction techniques, such as filtered back projection (FBP) and iterative reconstruction (IR), which have been utilised widely in the image reconstruction process of computed tomography (CT) are not suitable in the case of low-dose CT applications, because of the unsatisfying quality of the reconstructed image and inefficient reconstruction time. Therefore, as the demand for CT radiation dose reduction continues to increase, the use of artificial intelligence (AI) in image reconstruction has become a trend that attracts more and more attention. This systematic review examined various deep learning methods to determine their characteristics, availability, intended use and expected outputs concerning low-dose CT image reconstruction. Utilising the methodology of Kitchenham and Charter, we performed a systematic search of the literature from 2016 to 2021 in Springer, Science Direct, arXiv, PubMed, ACM, IEEE, and Scopus. This review showed that algorithms using deep learning technology are superior to traditional IR methods in noise suppression, artifact reduction and structure preservation, in terms of improving the image quality of low-dose reconstructed images. In conclusion, we provided an overview of the use of deep learning approaches in low-dose CT image reconstruction together with their benefits, limitations, and opportunities for improvement.
Collapse
|
82
|
Okamoto T, Kumakiri T, Haneishi H. Patch-based artifact reduction for three-dimensional volume projection data of sparse-view micro-computed tomography. Radiol Phys Technol 2022; 15:206-223. [PMID: 35622229 DOI: 10.1007/s12194-022-00661-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2022] [Revised: 04/27/2022] [Accepted: 04/28/2022] [Indexed: 11/27/2022]
Abstract
Micro-computed tomography (micro-CT) enables the non-destructive acquisition of three-dimensional (3D) morphological structures at the micrometer scale. Although it is expected to be used in pathology and histology to analyze the 3D microstructure of tissues, micro-CT imaging of tissue specimens requires a long scan time. A high-speed imaging method, sparse-view CT, can reduce the total scan time and radiation dose; however, it causes severe streak artifacts on tomographic images reconstructed with analytical algorithms due to insufficient sampling. In this paper, we propose an artifact reduction method for 3D volume projection data from sparse-view micro-CT. Specifically, we developed a patch-based lightweight fully convolutional network to estimate full-view 3D volume projection data from sparse-view 3D volume projection data. We evaluated the effectiveness of the proposed method using physically acquired datasets. The qualitative and quantitative results showed that the proposed method achieved high estimation accuracy and suppressed streak artifacts in the reconstructed images. In addition, we confirmed that the proposed method requires both short training and prediction times. Our study demonstrates that the proposed method has great potential for artifact reduction for 3D volume projection data under sparse-view conditions.
Collapse
Affiliation(s)
- Takayuki Okamoto
- Graduate School of Science and Engineering, Chiba University, Chiba, 263-8522, Japan.
| | - Toshio Kumakiri
- Graduate School of Science and Engineering, Chiba University, Chiba, 263-8522, Japan
| | - Hideaki Haneishi
- Center for Frontier Medical Engineering, Chiba University, Chiba, 263-8522, Japan
| |
Collapse
|
83
|
Scalco E, Rizzo G, Mastropietro A. The stability of oncologic MRI radiomic features and the potential role of deep learning: a review. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac60b9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2021] [Accepted: 03/24/2022] [Indexed: 11/11/2022]
Abstract
Abstract
The use of MRI radiomic models for the diagnosis, prognosis and treatment response prediction of tumors has been increasingly reported in literature. However, its widespread adoption in clinics is hampered by issues related to features stability. In the MRI radiomic workflow, the main factors that affect radiomic features computation can be found in the image acquisition and reconstruction phase, in the image pre-processing steps, and in the segmentation of the region of interest on which radiomic indices are extracted. Deep Neural Networks (DNNs), having shown their potentiality in the medical image processing and analysis field, can be seen as an attractive strategy to partially overcome the issues related to radiomic stability and mitigate their impact. In fact, DNN approaches can be prospectively integrated in the MRI radiomic workflow to improve image quality, obtain accurate and reproducible segmentations and generate standardized images. In this review, DNN methods that can be included in the image processing steps of the radiomic workflow are described and discussed, in the light of a detailed analysis of the literature in the context of MRI radiomic reliability.
Collapse
|
84
|
Sun L, Chen J, Xu Y, Gong M, Yu K, Batmanghelich K. Hierarchical Amortized GAN for 3D High Resolution Medical Image Synthesis. IEEE J Biomed Health Inform 2022; 26:3966-3975. [PMID: 35522642 PMCID: PMC9413516 DOI: 10.1109/jbhi.2022.3172976] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Generative Adversarial Networks (GAN) have many potential medical imaging applications, including data augmentation, domain adaptation, and model explanation. Due to the limited memory of Graphical Processing Units (GPUs), most current 3D GAN models are trained on low-resolution medical images, these models either cannot scale to high-resolution or are prone to patchy artifacts. In this work, we propose a novel end-to-end GAN architecture that can generate high-resolution 3D images. We achieve this goal by using different configurations between training and inference. During training, we adopt a hierarchical structure that simultaneously generates a low-resolution version of the image and a randomly selected sub-volume of the high-resolution image. The hierarchical design has two advantages: First, the memory demand for training on high-resolution images is amortized among sub-volumes. Furthermore, anchoring the high-resolution sub-volumes to a single low-resolution image ensures anatomical consistency between sub-volumes. During inference, our model can directly generate full high-resolution images. We also incorporate an encoder with a similar hierarchical structure into the model to extract features from the images. Experiments on 3D thorax CT and brain MRI demonstrate that our approach outperforms state of the art in image generation. We also demonstrate clinical applications of the proposed model in data augmentation and clinical-relevant feature extraction.
Collapse
|
85
|
Chen J, Bermejo I, Dekker A, Wee L. Generative models improve radiomics performance in different tasks and different datasets: An experimental study. Phys Med 2022; 98:11-17. [PMID: 35468494 DOI: 10.1016/j.ejmp.2022.04.008] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Revised: 03/11/2022] [Accepted: 04/17/2022] [Indexed: 02/02/2023] Open
Abstract
PURPOSE Radiomics is an active area of research focusing on high throughput feature extraction from medical images with a wide array of applications in clinical practice, such as clinical decision support in oncology. However, noise in low dose computed tomography (CT) scans can impair the accurate extraction of radiomic features. In this article, we investigate the possibility of using deep learning generative models to improve the performance of radiomics from low dose CTs. METHODS We used two datasets of low dose CT scans - NSCLC Radiogenomics and LIDC-IDRI - as test datasets for two tasks - pre-treatment survival prediction and lung cancer diagnosis. We used encoder-decoder networks and conditional generative adversarial networks (CGANs) trained in a previous study as generative models to transform low dose CT images into full dose CT images. Radiomic features extracted from the original and improved CT scans were used to build two classifiers - a support vector machine (SVM) and a deep attention based multiple instance learning model - for survival prediction and lung cancer diagnosis respectively. Finally, we compared the performance of the models derived from the original and improved CT scans. RESULTS Denoising with the encoder-decoder network and the CGAN improved the area under the curve (AUC) of survival prediction from 0.52 to 0.57 (p-value < 0.01). On the other hand, the encoder-decoder network and the CGAN improved the AUC of lung cancer diagnosis from 0.84 to 0.88 and 0.89 respectively (p-value < 0.01). Finally, there are no statistically significant improvements in AUC using encoder-decoder networks and CGAN (p-value = 0.34) when networks trained at 75 and 100 epochs. CONCLUSION Generative models can improve the performance of low dose CT-based radiomics in different tasks. Hence, denoising using generative models seems to be a necessary pre-processing step for calculating radiomic features from low dose CTs.
Collapse
Affiliation(s)
- Junhua Chen
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht 6229 ET, Netherlands.
| | - Inigo Bermejo
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht 6229 ET, Netherlands
| | - Andre Dekker
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht 6229 ET, Netherlands
| | - Leonard Wee
- Department of Radiation Oncology (MAASTRO), GROW School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht 6229 ET, Netherlands
| |
Collapse
|
86
|
Kalare K, Bajpai M, Sarkar S, Munshi P. Deep neural network for beam hardening artifacts removal in image reconstruction. APPL INTELL 2022. [DOI: 10.1007/s10489-021-02604-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
87
|
Patwari M, Gutjahr R, Raupach R, Maier A. Limited parameter denoising for low-dose X-ray computed tomography using deep reinforcement learning. Med Phys 2022; 49:4540-4553. [PMID: 35362172 DOI: 10.1002/mp.15643] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2021] [Revised: 10/07/2021] [Accepted: 03/22/2022] [Indexed: 11/09/2022] Open
Abstract
BACKGROUND The use of deep learning has successfully solved several problems in the field of medical imaging. Deep learning has been applied to the CT denoising problem successfully. However, the use of deep learning requires large amounts of data to train deep convolutional networks (CNNs). Moreover, due to large parameter count, such deep CNNs may cause unexpected results. PURPOSE In this study, we introduce a novel CT denoising framework, which has interpretable behaviour, and provides useful results with limited data. METHODS We employ bilateral filtering in both the projection and volume domains to remove noise. To account for non-stationary noise, we tune the σ parameters of the volume for every projection view, and for every volume pixel. The tuning is carried out by two deep CNNs. Due to impracticality of labelling, the two deep CNNs are trained via a Deep-Q reinforcement learning task. The reward for the task is generated by using a custom reward function represented by a neural network. Our experiments were carried out on abdominal scans for the Mayo Clinic TCIA dataset, and the AAPM Low Dose CT Grand Challenge. RESULTS Our denoising framework has excellent denoising performance increasing the PSNR from 28.53 to 28.93, and increasing the SSIM from 0.8952 to 0.9204. We outperform several state-of-the-art deep CNNs, which have several orders of magnitude higher number of parameters (p-value (PSNR) = 0.000, p-value (SSIM) = 0.000). Our method does not introduce any blurring, which is introduced by MSE loss based methods, or any deep learning artifacts, which are introduced by WGAN based models. Our ablation studies show that parameter tuning and using our reward network results in the best possible results. CONCLUSIONS We present a novel CT denoising framework, which focuses on interpretability to deliver good denoising performance, especially with limited data. Our method outperforms state-of-the-art deep neural networks. Future work will be focused on accelerating our method, and generalizing to different geometries and body parts. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Mayank Patwari
- Pattern Recognition Lab, Friedrich-Alexander Universität Erlangen-Nürnberg, Erlangen, 91058, Germany.,CT Concepts, Siemens Healthineers AG, Forchheim, 91301, Germany
| | - Ralf Gutjahr
- CT Concepts, Siemens Healthineers AG, Forchheim, 91301, Germany
| | - Rainer Raupach
- CT Concepts, Siemens Healthineers AG, Forchheim, 91301, Germany
| | - Andreas Maier
- Pattern Recognition Lab, Friedrich-Alexander Universität Erlangen-Nürnberg, Erlangen, 91058, Germany
| |
Collapse
|
88
|
Segmentation of ultrasound image sequences by combing a novel deep siamese network with a deformable contour model. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07054-2] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
89
|
Yurt M, Özbey M, UH Dar S, Tinaz B, Oguz KK, Çukur T. Progressively Volumetrized Deep Generative Models for Data-Efficient Contextual Learning of MR Image Recovery. Med Image Anal 2022; 78:102429. [DOI: 10.1016/j.media.2022.102429] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2021] [Revised: 03/14/2022] [Accepted: 03/18/2022] [Indexed: 10/18/2022]
|
90
|
Leuliet T, Maxim V, Peyrin F, Sixou B. Impact of the training loss in deep learning based CT reconstruction of bone microarchitecture. Med Phys 2022; 49:2952-2964. [PMID: 35218039 DOI: 10.1002/mp.15577] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2021] [Revised: 12/23/2021] [Accepted: 02/13/2022] [Indexed: 11/07/2022] Open
Abstract
PURPOSE Computed tomography (CT) is a technique of choice to image bone structure at different scales. Methods to enhance the quality of degraded reconstructions obtained from low-dose CT data have shown impressive results recently, especially in the realm of supervised deep learning. As the choice of the loss function affects the reconstruction quality, it is necessary to focus on the way neural networks evaluate the correspondence between predicted and target images during the training stage. This is even more true in the case of bone microarchitecture imaging at high spatial resolution where both the quantitative analysis of Bone Mineral Density (BMD) and bone microstructure are essential for assessing diseases such as osteoporosis. Our aim is thus to evaluate the quality of reconstruction on key metrics for diagnosis depending on the loss function that has been used for training the neural network. METHODS We compare and analyze volumes that are reconstructed with neural networks trained with pixelwise, structural and adversarial loss functions or with a combination of them. We perform realistic simulations of various low-dose acquisitions of bone microarchitecture. Our comparative study is performed with metrics that have an interest regarding the diagnosis of bone diseases. We therefore focus on bone-specific metrics such as BV/TV, resolution, connectivity assessed with the Euler number and quantitative analysis of BMD to evaluate the quality of reconstruction obtained with networks trained with the different loss functions. RESULTS We find that using L1 norm as the pixelwise loss is the best choice compared to L2 or no pixelwise loss since it improves resolution without deteriorating other metrics. VGG perceptual loss, especially when combined with an adversarial loss, allows to better retrieve topological and morphological parameters of bone microarchitecture compared to SSIM. This however leads to a decreased resolution performance. The adversarial loss enchances the reconstruction performance in terms of BMD distribution accuracy. CONCLUSIONS In order to retrieve the quantitative and structural characteristics of bone microarchitecture that are essential for post-reconstruction diagnosis, our results suggest to use L1 norm as part of the loss function. Then, trade-offs should be made depending on the application: VGG perceptual loss improves accuracy in terms of connectivity at the cost of a deteriorated resolution, and adversarial losses help better retrieve BMD distribution while significantly increasing the training time. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Théo Leuliet
- Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Etienne, CNRS, Inserm, CREATIS UMR 5220, U1206, LYON, F-69621, France
| | - Voichiţa Maxim
- Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Etienne, CNRS, Inserm, CREATIS UMR 5220, U1206, LYON, F-69621, France
| | - Françoise Peyrin
- Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Etienne, CNRS, Inserm, CREATIS UMR 5220, U1206, LYON, F-69621, France
| | - Bruno Sixou
- Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Etienne, CNRS, Inserm, CREATIS UMR 5220, U1206, LYON, F-69621, France
| |
Collapse
|
91
|
Geng M, Meng X, Yu J, Zhu L, Jin L, Jiang Z, Qiu B, Li H, Kong H, Yuan J, Yang K, Shan H, Han H, Yang Z, Ren Q, Lu Y. Content-Noise Complementary Learning for Medical Image Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:407-419. [PMID: 34529565 DOI: 10.1109/tmi.2021.3113365] [Citation(s) in RCA: 29] [Impact Index Per Article: 14.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Medical imaging denoising faces great challenges, yet is in great demand. With its distinctive characteristics, medical imaging denoising in the image domain requires innovative deep learning strategies. In this study, we propose a simple yet effective strategy, the content-noise complementary learning (CNCL) strategy, in which two deep learning predictors are used to learn the respective content and noise of the image dataset complementarily. A medical image denoising pipeline based on the CNCL strategy is presented, and is implemented as a generative adversarial network, where various representative networks (including U-Net, DnCNN, and SRDenseNet) are investigated as the predictors. The performance of these implemented models has been validated on medical imaging datasets including CT, MR, and PET. The results show that this strategy outperforms state-of-the-art denoising algorithms in terms of visual quality and quantitative metrics, and the strategy demonstrates a robust generalization capability. These findings validate that this simple yet effective strategy demonstrates promising potential for medical image denoising tasks, which could exert a clinical impact in the future. Code is available at: https://github.com/gengmufeng/CNCL-denoising.
Collapse
|
92
|
Montoya JC, Zhang C, Li Y, Li K, Chen GH. Reconstruction of three-dimensional tomographic patient models for radiation dose modulation in CT from two scout views using deep learning. Med Phys 2022; 49:901-916. [PMID: 34908175 PMCID: PMC9080958 DOI: 10.1002/mp.15414] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2021] [Revised: 11/11/2021] [Accepted: 11/16/2021] [Indexed: 02/03/2023] Open
Abstract
BACKGROUND A tomographic patient model is essential for radiation dose modulation in x-ray computed tomography (CT). Currently, two-view scout images (also known as topograms) are used to estimate patient models with relatively uniform attenuation coefficients. These patient models do not account for the detailed anatomical variations of human subjects, and thus, may limit the accuracy of intraview or organ-specific dose modulations in emerging CT technologies. PURPOSE The purpose of this work was to show that 3D tomographic patient models can be generated from two-view scout images using deep learning strategies, and the reconstructed 3D patient models indeed enable accurate prescriptions of fluence-field modulated or organ-specific dose delivery in the subsequent CT scans. METHODS CT images and the corresponding two-view scout images were retrospectively collected from 4214 individual CT exams. The collected data were curated for the training of a deep neural network architecture termed ScoutCT-NET to generate 3D tomographic attenuation models from two-view scout images. The trained network was validated using a cohort of 55 136 images from 212 individual patients. To evaluate the accuracy of the reconstructed 3D patient models, radiation delivery plans were generated using ScoutCT-NET 3D patient models and compared with plans prescribed based on true CT images (gold standard) for both fluence-field-modulated CT and organ-specific CT. Radiation dose distributions were estimated using Monte Carlo simulations and were quantitatively evaluated using the Gamma analysis method. Modulated dose profiles were compared against state-of-the-art tube current modulation schemes. Impacts of ScoutCT-NET patient model-based dose modulation schemes on universal-purpose CT acquisitions and organ-specific acquisitions were also compared in terms of overall image appearance, noise magnitude, and noise uniformity. RESULTS The results demonstrate that (1) The end-to-end trained ScoutCT-NET can be used to generate 3D patient attenuation models and demonstrate empirical generalizability. (2) The 3D patient models can be used to accurately estimate the spatial distribution of radiation dose delivered by standard helical CTs prior to the actual CT acquisition; compared to the gold-standard dose distribution, 95.0% of the voxels in the ScoutCT-NET based dose maps have acceptable gamma values for 5 mm distance-to-agreement and 10% dose difference. (3) The 3D patient models also enabled accurate prescription of fluence-field modulated CT to generate a more uniform noise distribution across the patient body compared to tube current-modulated CT. (4) ScoutCT-NET 3D patient models enabled accurate prescription of organ-specific CT to boost image quality for a given body region-of-interest under a given radiation dose constraint. CONCLUSION 3D tomographic attenuation models generated by ScoutCT-NET from two-view scout images can be used to prescribe fluence-field-modulated or organ-specific CT scans with high accuracy for the overall objective of radiation dose reduction or image quality improvement for a given imaging task.
Collapse
Affiliation(s)
- Juan C Montoya
- Department of Medical Physics, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
| | - Chengzhu Zhang
- Department of Medical Physics, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
| | - Yinsheng Li
- Department of Medical Physics, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
| | - Ke Li
- Department of Medical Physics, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
- Department of Radiology, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
| | - Guang-Hong Chen
- Department of Medical Physics, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
- Department of Radiology, University of Wisconsin School of Medicine and Public Health, Madison, Wisconsin, USA
| |
Collapse
|
93
|
Cui X, Guo Y, Zhang X, Shangguan H, Liu B, Wang A. Artifact-Assisted multi-level and multi-scale feature fusion attention network for low-dose CT denoising. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2022; 30:875-889. [PMID: 35694948 DOI: 10.3233/xst-221149] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Since low-dose computed tomography (LDCT) images typically have higher noise that may affect accuracy of disease diagnosis, the objective of this study is to develop and evaluate a new artifact-assisted feature fusion attention (AAFFA) network to extract and reduce image artifact and noise in LDCT images. METHODS In AAFFA network, a feature fusion attention block is constructed for local multi-scale artifact feature extraction and progressive fusion from coarse to fine. A multi-level fusion architecture based on skip connection and attention modules is also introduced for artifact feature extraction. Specifically, long-range skip connections are used to enhance and fuse artifact features with different depth levels. Then, the fused shallower features enter channel attention for better extraction of artifact features, and the fused deeper features are sent into pixel attention for focusing on the artifact pixel information. Besides, an artifact channel is designed to provide rich artifact features and guide the extraction of noise and artifact features. The AAPM LDCT Challenge dataset is used to train and test the network. The performance is evaluated by using both visual observation and quantitative metrics including peak signal-noise-ratio (PSNR), structural similarity index (SSIM) and visual information fidelity (VIF). RESULTS Using AAFFA network improves the averaged PSNR/SSIM/VIF values of AAPM LDCT images from 43.4961, 0.9595, 0.3926 to 48.2513, 0.9859, 0.4589, respectively. CONCLUSIONS The proposed AAFFA network is able to effectively reduce noise and artifacts while preserving object edges. Assessment of visual quality and quantitative index demonstrates the significant improvement compared with other image denoising methods.
Collapse
Affiliation(s)
- Xueying Cui
- School of Applied Science, Taiyuan University of Science and Technology, Taiyuan, China
| | - Yingting Guo
- School of Applied Science, Taiyuan University of Science and Technology, Taiyuan, China
| | - Xiong Zhang
- School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan, China
| | - Hong Shangguan
- School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan, China
| | - Bin Liu
- School of Applied Science, Taiyuan University of Science and Technology, Taiyuan, China
| | - Anhong Wang
- School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan, China
| |
Collapse
|
94
|
Iterative Reconstruction for Low-Dose CT using Deep Gradient Priors of Generative Model. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2022. [DOI: 10.1109/trpms.2022.3148373] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
95
|
|
96
|
Treitl KM, Ricke J, Baur-Melnyk A. Whole-body magnetic resonance imaging (WBMRI) versus whole-body computed tomography (WBCT) for myeloma imaging and staging. Skeletal Radiol 2022; 51:43-58. [PMID: 34031705 PMCID: PMC8626374 DOI: 10.1007/s00256-021-03799-4] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/17/2021] [Revised: 04/19/2021] [Accepted: 04/25/2021] [Indexed: 02/02/2023]
Abstract
Myeloma-associated bone disease (MBD) develops in about 80-90% of patients and severely affects their quality of life, as it accounts for the majority of mortality and morbidity. Imaging in multiple myeloma (MM) and MBD is of utmost importance in order to detect bone and bone marrow lesions as well as extraosseous soft-tissue masses and complications before the initiation of treatment. It is required for determination of the stage of disease and aids in the assessment of treatment response. Whole-body low-dose computed tomography (WBLDCT) is the key modality to establish the initial diagnosis of MM and is now recommended as reference standard procedure for the detection of lytic destruction in MBD. In contrast, whole-body magnetic resonance imaging (WBMRI) has higher sensitivity for the detection of focal and diffuse plasma cell infiltration patterns of the bone marrow and identifies them prior to osteolytic destruction. It is recommended for the evaluation of spinal and vertebral lesions, while functional, diffusion-weighted MRI (DWI-MRI) is a promising tool for the assessment of treatment response. This review addresses the current improvements and limitations of WBCT and WBMRI for diagnosis and staging in MM, underlining the fact that both modalities offer complementary information. It further summarizes the corresponding radiological findings and novel technological aspects of both modalities.
Collapse
Affiliation(s)
- Karla M. Treitl
- grid.5252.00000 0004 1936 973XDepartment of Radiology, University Hospital, LMU Munich, Marchioninistr. 15, 81377 Munich, Germany
| | - Jens Ricke
- grid.5252.00000 0004 1936 973XDepartment of Radiology, University Hospital, LMU Munich, Marchioninistr. 15, 81377 Munich, Germany
| | - Andrea Baur-Melnyk
- grid.5252.00000 0004 1936 973XDepartment of Radiology, University Hospital, LMU Munich, Marchioninistr. 15, 81377 Munich, Germany
| |
Collapse
|
97
|
Su T, Sun X, Yang J, Mi D, Zhang Y, Wu H, Fang S, Chen Y, Zheng H, Liang D, Ge Y. DIRECT-Net: A unified mutual-domain material decomposition network for quantitative dual-energy CT imaging. Med Phys 2021; 49:917-934. [PMID: 34935146 DOI: 10.1002/mp.15413] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2021] [Revised: 11/23/2021] [Accepted: 12/08/2021] [Indexed: 01/13/2023] Open
Abstract
PURPOSE The purpose of this paper is to present an end-to-end deep convolutional neural network to improve the dual-energy CT (DECT) material decomposition performance. METHODS In this study, we proposes a unified mutual-domain (sinogram domain and CT domain) material decomposition network (DIRECT-Net) for DECT imaging. By design, the DIRECT-Net has immediate access to mutual-domain data, and utilizes stacked convolution neural network layers for noise reduction and material decomposition. The training data are numerically generated following the fundamental DECT imaging physics. Numerical simulation of the XCAT digital phantom, experiments of a biological specimen, a calcium chloride phantom and an iodine solution phantom are carried out to evaluate the performance of DIRECT-Net. Comparisons are performed with different DECT decomposition algorithms. RESULTS Results demonstrate that the proposed DIRECT-Net can generate water and bone basis images with less artifacts compared to the other decomposition methods. Additionally, the quantification errors of the calcium chloride (75-375 mg/cm3 ) and the iodine (2-20 mg/cm3 ) are less than 4%. CONCLUSIONS An end-to-end material decomposition network is proposed for quantitative DECT imaging. The qualitative and quantitative results demonstrate that this new DIRECT-Net has promising benefits in improving the DECT image quality.
Collapse
Affiliation(s)
- Ting Su
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Xindong Sun
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Jiecheng Yang
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Donghua Mi
- Department of Vascular Neurology, Beijing Tiantan Hospital, Capital Medical University, Beijing, China
| | - Yikun Zhang
- School of Computer Science and Engineering, Southeast University, Nanjing, China
| | - Haodi Wu
- Wuhan National Laboratory for Optoelectronics (WNLO), Huazhong University of Science and Technology (HUST), Wuhan, China
| | - Shibo Fang
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Yang Chen
- School of Computer Science and Engineering, Southeast University, Nanjing, China
| | - Hairong Zheng
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Dong Liang
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Yongshuai Ge
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.,Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| |
Collapse
|
98
|
Aslan S, Liu Z, Nikitin V, Bicer T, Leyffer S, Gürsoy D. Joint ptycho-tomography with deep generative priors. MACHINE LEARNING: SCIENCE AND TECHNOLOGY 2021. [DOI: 10.1088/2632-2153/ac1d35] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/11/2022] Open
Abstract
Abstract
Joint ptycho-tomography is a powerful computational imaging framework to recover the refractive properties of a 3D object while relaxing the requirements for probe overlap that is common in conventional phase retrieval. We use an augmented Lagrangian scheme for formulating the constrained optimization problem and employ an alternating direction method of multipliers (ADMM) for the joint solution. ADMM allows the problem to be split into smaller and computationally more efficient subproblems: ptychographic phase retrieval, tomographic reconstruction, and regularization of the solution. We extend our ADMM framework with plug-and-play (PnP) denoisers by replacing the regularization subproblem with a general denoising operator based on machine learning. While the PnP framework enables integrating such learned priors as denoising operators, tuning of the denoiser prior remains challenging. To overcome this challenge, we propose a denoiser parameter to control the effect of the denoiser and to accelerate the solution. In our simulations, we demonstrate that our proposed framework with parameter tuning and learned priors generates high-quality reconstructions under limited and noisy measurement data.
Collapse
|
99
|
Zhang X, Han Z, Shangguan H, Han X, Cui X, Wang A. Artifact and Detail Attention Generative Adversarial Networks for Low-Dose CT Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3901-3918. [PMID: 34329159 DOI: 10.1109/tmi.2021.3101616] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Generative adversarial networks are being extensively studied for low-dose computed tomography denoising. However, due to the similar distribution of noise, artifacts, and high-frequency components of useful tissue images, it is difficult for existing generative adversarial network-based denoising networks to effectively separate the artifacts and noise in the low-dose computed tomography images. In addition, aggressive denoising may damage the edge and structural information of the computed tomography image and make the denoised image too smooth. To solve these problems, we propose a novel denoising network called artifact and detail attention generative adversarial network. First, a multi-channel generator is proposed. Based on the main feature extraction channel, an artifacts and noise attention channel and an edge feature attention channel are added to improve the denoising network's ability to pay attention to the noise and artifacts features and edge features of the image. Additionally, a new structure called multi-scale Res2Net discriminator is proposed, and the receptive field in the module is expanded by extracting the multi-scale features in the same scale of the image to improve the discriminative ability of discriminator. The loss functions are specially designed for each sub-channel of the denoising network corresponding to its function. Through the cooperation of multiple loss functions, the convergence speed, stability, and denoising effect of the network are accelerated, improved, and guaranteed, respectively. Experimental results show that the proposed denoising network can preserve the important information of the low-dose computed tomography image and achieve better denoising effect when compared to the state-of-the-art algorithms.
Collapse
|
100
|
Xia W, Lu Z, Huang Y, Shi Z, Liu Y, Chen H, Chen Y, Zhou J, Zhang Y. MAGIC: Manifold and Graph Integrative Convolutional Network for Low-Dose CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3459-3472. [PMID: 34110990 DOI: 10.1109/tmi.2021.3088344] [Citation(s) in RCA: 27] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Low-dose computed tomography (LDCT) scans, which can effectively alleviate the radiation problem, will degrade the imaging quality. In this paper, we propose a novel LDCT reconstruction network that unrolls the iterative scheme and performs in both image and manifold spaces. Because patch manifolds of medical images have low-dimensional structures, we can build graphs from the manifolds. Then, we simultaneously leverage the spatial convolution to extract the local pixel-level features from the images and incorporate the graph convolution to analyze the nonlocal topological features in manifold space. The experiments show that our proposed method outperforms both the quantitative and qualitative aspects of state-of-the-art methods. In addition, aided by a projection loss component, our proposed method also demonstrates superior performance for semi-supervised learning. The network can remove most noise while maintaining the details of only 10% (40 slices) of the training data labeled.
Collapse
|