1
|
Kim W, Jeon SY, Byun G, Yoo H, Choi JH. A systematic review of deep learning-based denoising for low-dose computed tomography from a perceptual quality perspective. Biomed Eng Lett 2024; 14:1153-1173. [PMID: 39465112 PMCID: PMC11502640 DOI: 10.1007/s13534-024-00419-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2024] [Revised: 08/03/2024] [Accepted: 08/18/2024] [Indexed: 10/29/2024] Open
Abstract
Low-dose computed tomography (LDCT) scans are essential in reducing radiation exposure but often suffer from significant image noise that can impair diagnostic accuracy. While deep learning approaches have enhanced LDCT denoising capabilities, the predominant reliance on objective metrics like PSNR and SSIM has resulted in over-smoothed images that lack critical detail. This paper explores advanced deep learning methods tailored specifically to improve perceptual quality in LDCT images, focusing on generating diagnostic-quality images preferred in clinical practice. We review and compare current methodologies, including perceptual loss functions and generative adversarial networks, addressing the significant limitations of current benchmarks and the subjective nature of perceptual quality evaluation. Through a systematic analysis, this study underscores the urgent need for developing methods that balance both perceptual and diagnostic quality, proposing new directions for future research in the field.
Collapse
Affiliation(s)
- Wonjin Kim
- Department of Mechanical Engineering, Korean Advanced Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon, 34141 Korea
- AI Analysis Team, Dotter Inc., 225 Gasan Digital 1-ro, Geumchoen-gu, Seoul, 08501 Korea
| | - Sun-Young Jeon
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| | - Gyuri Byun
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| | - Hongki Yoo
- Department of Mechanical Engineering, Korean Advanced Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon, 34141 Korea
| | - Jang-Hwan Choi
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
- Computational Medicine, Graduate Program in System Health Science and Engineering, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| |
Collapse
|
2
|
Sun C, Salimi Y, Angeliki N, Boudabbous S, Zaidi H. An efficient dual-domain deep learning network for sparse-view CT reconstruction. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 256:108376. [PMID: 39173481 DOI: 10.1016/j.cmpb.2024.108376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/23/2024] [Revised: 08/02/2024] [Accepted: 08/15/2024] [Indexed: 08/24/2024]
Abstract
BACKGROUND AND OBJECTIVE We develop an efficient deep-learning based dual-domain reconstruction method for sparse-view CT reconstruction with small training parameters and comparable running time. We aim to investigate the model's capability and its clinical value by performing objective and subjective quality assessments using clinical CT projection data acquired on commercial scanners. METHODS We designed two lightweight networks, namely Sino-Net and Img-Net, to restore the projection and image signal from the DD-Net reconstructed images in the projection and image domains, respectively. The proposed network has small training parameters and comparable running time among dual-domain based reconstruction networks and is easy to train (end-to-end). We prospectively collected clinical thoraco-abdominal CT projection data acquired on a Siemens Biograph 128 Edge CT scanner to train and validate the proposed network. Further, we quantitatively evaluated the CT Hounsfield unit (HU) values on 21 organs and anatomic structures, such as the liver, aorta, and ribcage. We also analyzed the noise properties and compared the signal-to-noise ratio (SNR) and the contrast-to-noise ratio (CNR) of the reconstructed images. Besides, two radiologists conducted the subjective qualitative evaluation including the confidence and conspicuity of anatomic structures, and the overall image quality using a 1-5 likert scoring system. RESULTS Objective and subjective evaluation showed that the proposed algorithm achieves competitive results in eliminating noise and artifacts, restoring fine structure details, and recovering edges and contours of anatomic structures using 384 views (1/6 sparse rate). The proposed method exhibited good computational cost performance on clinical projection data. CONCLUSION This work presents an efficient dual-domain learning network for sparse-view CT reconstruction on raw projection data from a commercial scanner. The study also provides insights for designing an organ-based image quality assessment pipeline for sparse-view reconstruction tasks, potentially benefiting organ-specific dose reduction by sparse-view imaging.
Collapse
Affiliation(s)
- Chang Sun
- Beijing University of Posts and Telecommunications, School of Information and Communication Engineering, 100876 Beijing, China; Geneva University Hospital, Division of Nuclear Medicine and Molecular Imaging, CH-1211 Geneva, Switzerland
| | - Yazdan Salimi
- Geneva University Hospital, Division of Nuclear Medicine and Molecular Imaging, CH-1211 Geneva, Switzerland
| | - Neroladaki Angeliki
- Geneva University Hospital, Division of Radiology, CH-1211, Geneva, Switzerland
| | - Sana Boudabbous
- Geneva University Hospital, Division of Radiology, CH-1211, Geneva, Switzerland
| | - Habib Zaidi
- Geneva University Hospital, Division of Nuclear Medicine and Molecular Imaging, CH-1211 Geneva, Switzerland; Department of Nuclear Medicine and Molecular Imaging, University of Groningen, University Medical Center Groningen, Groningen, Netherlands; Department of Nuclear Medicine, University of Southern Denmark, Odense, Denmark; University Research and Innovation Center, Óbuda University, Budapest, Hungary.
| |
Collapse
|
3
|
Xue H, Yao Y, Teng Y. Noise-assisted hybrid attention networks for low-dose PET and CT denoising. Med Phys 2024. [PMID: 39431968 DOI: 10.1002/mp.17430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 07/25/2024] [Accepted: 09/04/2024] [Indexed: 10/22/2024] Open
Abstract
BACKGROUND Positron emission tomography (PET) and computed tomography (CT) play a vital role in tumor-related medical diagnosis, assessment, and treatment planning. However, full-dose PET and CT pose the risk of excessive radiation exposure to patients, whereas low-dose images compromise image quality, impacting subsequent tumor recognition and disease diagnosis. PURPOSE To solve such problems, we propose a Noise-Assisted Hybrid Attention Network (NAHANet) to reconstruct full-dose PET and CT images from low-dose PET (LDPET) and CT (LDCT) images to reduce patient radiation risks while ensuring the performance of subsequent tumor recognition. METHODS NAHANet contains two branches: the noise feature prediction branch (NFPB) and the cascaded reconstruction branch. Among them, NFPB providing noise features for the cascade reconstruction branch. The cascaded reconstruction branch comprises a shallow feature extraction module and a reconstruction module which contains a series of cascaded noise feature fusion blocks (NFFBs). Among these, the NFFB fuses the features extracted from low-dose images with the noise features obtained by NFPB to improve the feature extraction capability. To validate the effectiveness of the NAHANet method, we performed experiments using two public available datasets: the Ultra-low Dose PET Imaging Challenge dataset and Low Dose CT Grand Challenge dataset. RESULTS As a result, the proposed NAHANet achieved higher performance on common indicators. For example, on the CT dataset, the PSNR and SSIM indicators were improved by 4.1 dB and 0.06 respectively, and the rMSE indicator was reduced by 5.46 compared with the LDCT; on the PET dataset, the PSNR and SSIM was improved by 3.37 dB and 0.02, and the rMSE was reduced by 9.04 compared with the LDPET. CONCLUSIONS This paper proposes a transformer-based denoising algorithm, which utilizes hybrid attention to extract high-level features of low dose images and fuses noise features to optimize the denoising performance of the network, achieving good performance improvements on low-dose CT and PET datasets.
Collapse
Affiliation(s)
- Hengzhi Xue
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yudong Yao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Electrical and Computer Engineering, Steven Institute of Technology, Hoboken, New Jersey, USA
| | - Yueyang Teng
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Shenyang, China
| |
Collapse
|
4
|
Zhang J, Sha J, Liu W, Zhou Y, Liu H, Zuo Z. Quantification of Intratumoral Heterogeneity: Distinguishing Histological Subtypes in Clinical T1 Stage Lung Adenocarcinoma Presenting as Pure Ground-Glass Nodules on Computed Tomography. Acad Radiol 2024; 31:4244-4255. [PMID: 38627129 DOI: 10.1016/j.acra.2024.04.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2024] [Revised: 04/02/2024] [Accepted: 04/06/2024] [Indexed: 10/21/2024]
Abstract
RATIONALE AND OBJECTIVES To quantify intratumor heterogeneity (ITH) in clinical T1 stage lung adenocarcinoma presenting as pure ground-glass nodules (pGGN) on computed tomography, assessing its value in distinguishing histological subtypes. MATERIALS AND METHODS An ITH score was developed for quantitative measurement by integrating local radiomics features and global pixel distribution patterns. Diagnostic efficacy in distinguishing histological subtypes was evaluated using receiver operating characteristic curve analysis and area under the curve (AUC) values. The ITH score's performance was compared to those of conventional radiomics (C-radiomics), and radiological assessments conducted by experienced radiologists. RESULTS The ITH score demonstrated excellent performance in distinguishing lepidic-predominant adenocarcinoma (LPA) from other histological subtypes of clinical T1 stage lung adenocarcinoma presenting as pGGN. It outperformed both C-radiomics and radiological findings, exhibiting higher AUCs of 0.784 (95% confidence interval [CI]: 0.742-0.826) and 0.801 (95% CI: 0.739-0.863) in the training and validation cohorts, respectively. The AUCs of C-radiomics were 0.764 (95% CI: 0.718-0.810, DeLong test, p = 0.025) and 0.760 (95% CI: 0.692-0.829, p = 0.023) and those of radiological findings were 0.722 (95% CI: 0.673-0.771, p = 0.003) and 0.754 (95% CI: 0.684-0.823, p = 0.016) in the training and validation cohorts, respectively. Subgroup analysis revealed varying diagnostic efficacy across clinical T1 stages, with the highest efficacy in the T1a stage, followed by the T1b stage, and lowest in the T1c stage. CONCLUSION The ITH score presents a superior method for evaluating histological subtypes and distinguishing LPA from other subtypes in clinical T1 stage lung adenocarcinoma presenting as pGGN.
Collapse
Affiliation(s)
- Jian Zhang
- Department of Radiology, Wuhan Pulmonary Hospital, Wuhan 430000, Hubei, PR China
| | - Jinlu Sha
- Department of Radiology, Wuhan Pulmonary Hospital, Wuhan 430000, Hubei, PR China
| | - Wen Liu
- Department of Radiology, The Third Xiangya Hospital of Central South University, Changsha 410013, Hunan, China
| | - Yinjun Zhou
- Department of Radiology, Xiangtan Central Hospital, Xiangtan 411100, Hunan, PR China
| | - Haibo Liu
- Department of Radiology, Xiangtan Central Hospital, Xiangtan 411100, Hunan, PR China
| | - Zhichao Zuo
- School of Mathematics and Computational Science, Xiangtan University, Xiangtan 411105, Hunan, China.
| |
Collapse
|
5
|
Lee W, Wagner F, Galdran A, Shi Y, Xia W, Wang G, Mou X, Ahamed MA, Imran AAZ, Oh JE, Kim K, Baek JT, Lee D, Hong B, Tempelman P, Lyu D, Kuiper A, van Blokland L, Calisto MB, Hsieh S, Han M, Baek J, Maier A, Wang A, Gold GE, Choi JH. Low-dose computed tomography perceptual image quality assessment. Med Image Anal 2024; 99:103343. [PMID: 39265362 DOI: 10.1016/j.media.2024.103343] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2024] [Revised: 08/30/2024] [Accepted: 09/03/2024] [Indexed: 09/14/2024]
Abstract
In computed tomography (CT) imaging, optimizing the balance between radiation dose and image quality is crucial due to the potentially harmful effects of radiation on patients. Although subjective assessments by radiologists are considered the gold standard in medical imaging, these evaluations can be time-consuming and costly. Thus, objective methods, such as the peak signal-to-noise ratio and structural similarity index measure, are often employed as alternatives. However, these metrics, initially developed for natural images, may not fully encapsulate the radiologists' assessment process. Consequently, interest in developing deep learning-based image quality assessment (IQA) methods that more closely align with radiologists' perceptions is growing. A significant barrier to this development has been the absence of open-source datasets and benchmark models specific to CT IQA. Addressing these challenges, we organized the Low-dose Computed Tomography Perceptual Image Quality Assessment Challenge in conjunction with the Medical Image Computing and Computer Assisted Intervention 2023. This event introduced the first open-source CT IQA dataset, consisting of 1,000 CT images of various quality, annotated with radiologists' assessment scores. As a benchmark, this challenge offers a comprehensive analysis of six submitted methods, providing valuable insight into their performance. This paper presents a summary of these methods and insights. This challenge underscores the potential for developing no-reference IQA methods that could exceed the capabilities of full-reference IQA methods, making a significant contribution to the research community with this novel dataset. The dataset is accessible at https://zenodo.org/records/7833096.
Collapse
Affiliation(s)
- Wonkyeong Lee
- Ewha Womans University, 52 Ewhayeodae-gil, Seodaemun-gu, Seoul 03760, Republic of Korea.
| | - Fabian Wagner
- Friedrich-Alexander-Universität Erlangen-Nürnberg, Schloßplatz 4, Erlangen 91054, Germany
| | - Adrian Galdran
- Universitat Pompeu Fabra, Plaça de la Mercè, 12, Ciutat Vella, Barcelona 08002, Spain
| | - Yongyi Shi
- Rensselaer Polytechnic Institute, 110 8th St, Troy, NY 12180, USA
| | - Wenjun Xia
- Rensselaer Polytechnic Institute, 110 8th St, Troy, NY 12180, USA
| | - Ge Wang
- Rensselaer Polytechnic Institute, 110 8th St, Troy, NY 12180, USA
| | - Xuanqin Mou
- Xi'an Jiaotong University, 28, Xianning West Road, Xi'an City, Shaanxi Province 710049, People's Republic of China
| | - Md Atik Ahamed
- Department of Computer Science, University of Kentucky, Lexington, KY 40506, USA
| | | | - Ji Eun Oh
- Chungnam National University College of Medicine, 266 Munghwa-ro, Daejeon 35015, Republic of Korea
| | - Kyungsang Kim
- MGH and Harvard Medical School, 25 Shattuck Street, Boston, MA 02115, USA
| | - Jong Tak Baek
- Chungnam National University College of Medicine, 266 Munghwa-ro, Daejeon 35015, Republic of Korea
| | - Dongheon Lee
- Chungnam National University College of Medicine, 266 Munghwa-ro, Daejeon 35015, Republic of Korea
| | - Boohwi Hong
- Chungnam National University College of Medicine, 266 Munghwa-ro, Daejeon 35015, Republic of Korea
| | - Philip Tempelman
- Delft University of Technology, Mekelweg 5, CD Delft 2628, Netherlands
| | - Donghang Lyu
- Leiden University, Rapenburg 70, EZ Leiden 2311, Netherlands
| | - Adrian Kuiper
- Delft University of Technology, Mekelweg 5, CD Delft 2628, Netherlands
| | - Lars van Blokland
- Delft University of Technology, Mekelweg 5, CD Delft 2628, Netherlands
| | - Maria Baldeon Calisto
- Universidad San Francisco de Quito, Campus Cumbayá, Diego de Robles s/n, Quito 170901, Ecuador
| | - Scott Hsieh
- Mayo Clinic, 200 First St., SW Rochester, MN 55905, USA
| | - Minah Han
- Yonsei University, A50 Yonsei-ro, Seodaemun-gu, Seoul 03722, Republic of Korea
| | - Jongduk Baek
- Yonsei University, A50 Yonsei-ro, Seodaemun-gu, Seoul 03722, Republic of Korea
| | - Andreas Maier
- Friedrich-Alexander-Universität Erlangen-Nürnberg, Schloßplatz 4, Erlangen 91054, Germany
| | - Adam Wang
- Stanford University, 450 Jane Stanford Way, Stanford, CA, 94305, USA
| | - Garry Evan Gold
- Stanford University, 450 Jane Stanford Way, Stanford, CA, 94305, USA
| | - Jang-Hwan Choi
- Ewha Womans University, 52 Ewhayeodae-gil, Seodaemun-gu, Seoul 03760, Republic of Korea; Computational Medicine, Graduate Program in System Health Science and Engineering, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760, Republic of Korea.
| |
Collapse
|
6
|
Zhao F, Liu M, Xiang M, Li D, Jiang X, Jin X, Lin C, Wang R. Unsupervised and Self-supervised Learning in Low-Dose Computed Tomography Denoising: Insights from Training Strategies. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01213-8. [PMID: 39231886 DOI: 10.1007/s10278-024-01213-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 06/30/2024] [Accepted: 07/01/2024] [Indexed: 09/06/2024]
Abstract
In recent years, X-ray low-dose computed tomography (LDCT) has garnered widespread attention due to its significant reduction in the risk of patient radiation exposure. However, LDCT images often contain a substantial amount of noises, adversely affecting diagnostic quality. To mitigate this, a plethora of LDCT denoising methods have been proposed. Among them, deep learning (DL) approaches have emerged as the most effective, due to their robust feature extraction capabilities. Yet, the prevalent use of supervised training paradigms is often impractical due to the challenges in acquiring low-dose and normal-dose CT pairs in clinical settings. Consequently, unsupervised and self-supervised deep learning methods have been introduced for LDCT denoising, showing considerable potential for clinical applications. These methods' efficacy hinges on training strategies. Notably, there appears to be no comprehensive reviews of these strategies. Our review aims to address this gap, offering insights and guidance for researchers and practitioners. Based on training strategies, we categorize the LDCT methods into six groups: (i) cycle consistency-based, (ii) score matching-based, (iii) statistical characteristics of noise-based, (iv) similarity-based, (v) LDCT synthesis model-based, and (vi) hybrid methods. For each category, we delve into the theoretical underpinnings, training strategies, strengths, and limitations. In addition, we also summarize the open source codes of the reviewed methods. Finally, the review concludes with a discussion on open issues and future research directions.
Collapse
Affiliation(s)
- Feixiang Zhao
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, 1 East Third Road, Chengdu, 610059, Sichuan, China
| | - Mingzhe Liu
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
- College of Computer Science and Cyber Security, Chengdu University of Technology, 1 East Third Road, Chengdu, 610059, Sichuan, China
| | - Mingrong Xiang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China.
- School of Information Technology, Deakin University, Melbourne Burwood Campus, 221 Burwood Hwy, Melbourne, 3125, Victoria, Australia.
| | - Dongfen Li
- College of Computer Science and Cyber Security, Chengdu University of Technology, 1 East Third Road, Chengdu, 610059, Sichuan, China
| | - Xin Jiang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
| | - Xiance Jin
- Department of Radiotherapy Center, The first Affiliated Hospital of Wenzhou Medical University, Ouhai District, Wenzhou, 325000, Zhejiang, China
| | - Cai Lin
- Department of Burn, Wound Repair and Regenerative Medicine Center, The first Affiliated Hospital of Wenzhou Medical University, Ouhai District, Wenzhou, 325000, Zhejiang, China
| | - Ruili Wang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China
- School of Mathematical and Computational Science, Massey University, SH17, Albany, 0632, Auckland, New Zealand
| |
Collapse
|
7
|
Huang J, Yang L, Wang F, Wu Y, Nan Y, Wu W, Wang C, Shi K, Aviles-Rivero AI, Schönlieb CB, Zhang D, Yang G. Enhancing global sensitivity and uncertainty quantification in medical image reconstruction with Monte Carlo arbitrary-masked mamba. Med Image Anal 2024; 99:103334. [PMID: 39255733 DOI: 10.1016/j.media.2024.103334] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Revised: 08/05/2024] [Accepted: 09/01/2024] [Indexed: 09/12/2024]
Abstract
Deep learning has been extensively applied in medical image reconstruction, where Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) represent the predominant paradigms, each possessing distinct advantages and inherent limitations: CNNs exhibit linear complexity with local sensitivity, whereas ViTs demonstrate quadratic complexity with global sensitivity. The emerging Mamba has shown superiority in learning visual representation, which combines the advantages of linear scalability and global sensitivity. In this study, we introduce MambaMIR, an Arbitrary-Masked Mamba-based model with wavelet decomposition for joint medical image reconstruction and uncertainty estimation. A novel Arbitrary Scan Masking (ASM) mechanism "masks out" redundant information to introduce randomness for further uncertainty estimation. Compared to the commonly used Monte Carlo (MC) dropout, our proposed MC-ASM provides an uncertainty map without the need for hyperparameter tuning and mitigates the performance drop typically observed when applying dropout to low-level tasks. For further texture preservation and better perceptual quality, we employ the wavelet transformation into MambaMIR and explore its variant based on the Generative Adversarial Network, namely MambaMIR-GAN. Comprehensive experiments have been conducted for multiple representative medical image reconstruction tasks, demonstrating that the proposed MambaMIR and MambaMIR-GAN outperform other baseline and state-of-the-art methods in different reconstruction tasks, where MambaMIR achieves the best reconstruction fidelity and MambaMIR-GAN has the best perceptual quality. In addition, our MC-ASM provides uncertainty maps as an additional tool for clinicians, while mitigating the typical performance drop caused by the commonly used dropout.
Collapse
Affiliation(s)
- Jiahao Huang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom.
| | - Liutao Yang
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Fanwen Wang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom
| | - Yinzhe Wu
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom.
| | - Yang Nan
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom
| | - Weiwen Wu
- School of Biomedical Engineering, Shenzhen Campus of Sun Yat-sen University, Guangdong, China
| | - Chengyan Wang
- Human Phenome Institute, Fudan University, Shanghai, China
| | - Kuangyu Shi
- Department of Nuclear Medicine, Inselspital, University of Bern, Bern, Switzerland; Department of Informatics, Technical University of Munich, Munich, Germany
| | - Angelica I Aviles-Rivero
- Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, United Kingdom
| | - Carola-Bibiane Schönlieb
- Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, United Kingdom
| | - Daoqiang Zhang
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Guang Yang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom; School of Biomedical Engineering & Imaging Sciences, King's College London, London WC2R 2LS, United Kingdom.
| |
Collapse
|
8
|
Winfree T, McCollough C, Yu L. Development and validation of a noise insertion algorithm for photon-counting-detector CT. Med Phys 2024; 51:5943-5953. [PMID: 38923526 PMCID: PMC11489017 DOI: 10.1002/mp.17263] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Revised: 05/30/2024] [Accepted: 05/31/2024] [Indexed: 06/28/2024] Open
Abstract
BACKGROUND Inserting noise into existing patient projection data to simulate lower-radiation-dose exams has been frequently used in traditional energy-integrating-detector (EID)-CT to optimize radiation dose in clinical protocols and to generate paired images for training deep-learning-based reconstruction and noise reduction methods. Recent introduction of photon counting detector CT (PCD-CT) also requires such a method to accomplish these tasks. However, clinical PCD-CT scanners often restrict the users access to the raw count data, exporting only the preprocessed, log-normalized sinogram. Therefore, it remains a challenge to employ projection domain noise insertion algorithms on PCD-CT. PURPOSE To develop and validate a projection domain noise insertion algorithm for PCD-CT that does not require access to the raw count data. MATERIALS AND METHODS A projection-domain noise model developed originally for EID-CT was adapted for PCD-CT. This model requires, as input, a map of the incident number of photons at each detector pixel when no object is in the beam. To obtain the map of incident number of photons, air scans were acquired on a PCD-CT scanner, then the noise equivalent photon number (NEPN) was calculated from the variance in the log normalized projection data of each scan. Additional air scans were acquired at various mA settings to investigate the impact of pulse pileup on the linearity of NEPN measurement. To validate the noise insertion algorithm, Noise Power Spectra (NPS) were generated from a 30 cm water tank scan and used to compare the noise texture and noise level of measured and simulated half dose and quarter dose images. An anthropomorphic thorax phantom was scanned with automatic exposure control, and noise levels at different slice locations were compared between simulated and measured half dose and quarter dose images. Spectral correlation between energy thresholds T1 and T2, and energy bins, B1 and B2, was compared between simulated and measured data across a wide range of tube current. Additionally, noise insertion was performed on a clinical patient case for qualitative assessment. RESULTS The NPS generated from simulated low dose water tank images showed similar shape and amplitude to that generated from the measured low dose images, differing by a maximum of 5.0% for half dose (HD) T1 images, 6.3% for HD T2 images, 4.1% for quarter dose (QD) T1 images, and 6.1% for QD T2 images. Noise versus slice measurements of the lung phantom showed comparable results between measured and simulated low dose images, with root mean square percent errors of 5.9%, 5.4%, 5.0%, and 4.6% for QD T1, HD T1, QD T2, and HD T2, respectively. NEPN measurements in air were linear up until 112 mA, after which pulse pileup effects significantly distort the air scan NEPN profile. Spectral correlation between T1 and T2 in simulation agreed well with that in the measured data in typical dose ranges. CONCLUSIONS A projection-domain noise insertion algorithm was developed and validated for PCD-CT to synthesize low-dose images from existing scans. It can be used for optimizing scanning protocols and generating paired images for training deep-learning-based methods.
Collapse
Affiliation(s)
| | | | - Lifeng Yu
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| |
Collapse
|
9
|
Chen Z, Hu B, Niu C, Chen T, Li Y, Shan H, Wang G. IQAGPT: computed tomography image quality assessment with vision-language and ChatGPT models. Vis Comput Ind Biomed Art 2024; 7:20. [PMID: 39101954 DOI: 10.1186/s42492-024-00171-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2024] [Accepted: 07/24/2024] [Indexed: 08/06/2024] Open
Abstract
Large language models (LLMs), such as ChatGPT, have demonstrated impressive capabilities in various tasks and attracted increasing interest as a natural language interface across many domains. Recently, large vision-language models (VLMs) that learn rich vision-language correlation from image-text pairs, like BLIP-2 and GPT-4, have been intensively investigated. However, despite these developments, the application of LLMs and VLMs in image quality assessment (IQA), particularly in medical imaging, remains unexplored. This is valuable for objective performance evaluation and potential supplement or even replacement of radiologists' opinions. To this end, this study introduces IQAGPT, an innovative computed tomography (CT) IQA system that integrates image-quality captioning VLM with ChatGPT to generate quality scores and textual reports. First, a CT-IQA dataset comprising 1,000 CT slices with diverse quality levels is professionally annotated and compiled for training and evaluation. To better leverage the capabilities of LLMs, the annotated quality scores are converted into semantically rich text descriptions using a prompt template. Second, the image-quality captioning VLM is fine-tuned on the CT-IQA dataset to generate quality descriptions. The captioning model fuses image and text features through cross-modal attention. Third, based on the quality descriptions, users verbally request ChatGPT to rate image-quality scores or produce radiological quality reports. Results demonstrate the feasibility of assessing image quality using LLMs. The proposed IQAGPT outperformed GPT-4 and CLIP-IQA, as well as multitask classification and regression models that solely rely on images.
Collapse
Affiliation(s)
- Zhihao Chen
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, 200433, China
| | - Bin Hu
- Department of Radiology, Huashan Hospital, Fudan University, Shanghai, 200040, China
| | - Chuang Niu
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, 12180, US
| | - Tao Chen
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, 200433, China
| | - Yuxin Li
- Department of Radiology, Huashan Hospital, Fudan University, Shanghai, 200040, China.
| | - Hongming Shan
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, 200433, China.
- MOE Frontiers Center for Brain Science, Fudan University, Shanghai, 200032, China.
- Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence (Ministry of Education), Fudan University, Shanghai, 200433, China.
| | - Ge Wang
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, 12180, US.
| |
Collapse
|
10
|
Liao P, Zhang X, Wu Y, Chen H, Du W, Liu H, Yang H, Zhang Y. Weakly supervised low-dose computed tomography denoising based on generative adversarial networks. Quant Imaging Med Surg 2024; 14:5571-5590. [PMID: 39144020 PMCID: PMC11320552 DOI: 10.21037/qims-24-68] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2024] [Accepted: 06/17/2024] [Indexed: 08/16/2024]
Abstract
Background Low-dose computed tomography (LDCT) is a diagnostic imaging technique designed to minimize radiation exposure to the patient. However, this reduction in radiation may compromise computed tomography (CT) image quality, adversely impacting clinical diagnoses. Various advanced LDCT methods have emerged to mitigate this challenge, relying on well-matched LDCT and normal-dose CT (NDCT) image pairs for training. Nevertheless, these methods often face difficulties in distinguishing image details from nonuniformly distributed noise, limiting their denoising efficacy. Additionally, acquiring suitably paired datasets in the medical domain poses challenges, further constraining their applicability. Hence, the objective of this study was to develop an innovative denoising framework for LDCT images employing unpaired data. Methods In this paper, we propose a LDCT denoising network (DNCNN) that alleviates the need for aligning LDCT and NDCT images. Our approach employs generative adversarial networks (GANs) to learn and model the noise present in LDCT images, establishing a mapping from the pseudo-LDCT to the actual NDCT domain without the need for paired CT images. Results Within the domain of weakly supervised methods, our proposed model exhibited superior objective metrics on the simulated dataset when compared to CycleGAN and selective kernel-based cycle-consistent GAN (SKFCycleGAN): the peak signal-to-noise ratio (PSNR) was 43.9441, the structural similarity index measure (SSIM) was 0.9660, and the visual information fidelity (VIF) was 0.7707. In the clinical dataset, we conducted a visual effect analysis by observing various tissues through different observation windows. Our proposed method achieved a no-reference structural sharpness (NRSS) value of 0.6171, which was closest to that of the NDCT images (NRSS =0.6049), demonstrating its superiority over other denoising techniques in preserving details, maintaining structural integrity, and enhancing edge contrast. Conclusions Through extensive experiments on both simulated and clinical datasets, we demonstrated the superior efficacy of our proposed method in terms of denoising quality and quantity. Our method exhibits superiority over both supervised techniques, including block-matching and 3D filtering (BM3D), residual encoder-decoder convolutional neural network (RED-CNN), and Wasserstein generative adversarial network-VGG (WGAN-VGG), and over weakly supervised approaches, including CycleGAN and SKFCycleGAN.
Collapse
Affiliation(s)
- Peixi Liao
- Department of Stomatology, The Sixth People’s Hospital of Chengdu, Chengdu, China
| | - Xucan Zhang
- The National Key Laboratory of Fundamental Science on Synthetic Vision, Sichuan University, Chengdu, China
| | - Yaoyao Wu
- The School of Computer Science, Sichuan University, Chengdu, China
| | - Hu Chen
- The College of Computer Science, Sichuan University, Chengdu, China
| | - Wenchao Du
- The College of Computer Science, Sichuan University, Chengdu, China
| | - Hong Liu
- The College of Computer Science, Sichuan University, Chengdu, China
| | - Hongyu Yang
- The College of Computer Science, Sichuan University, Chengdu, China
| | - Yi Zhang
- The School of Cyber Science and Engineering, Sichuan University, Chengdu, China
| |
Collapse
|
11
|
Du W, Cui H, He L, Chen H, Zhang Y, Yang H. Structure-aware diffusion for low-dose CT imaging. Phys Med Biol 2024; 69:155008. [PMID: 38942004 DOI: 10.1088/1361-6560/ad5d47] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2024] [Accepted: 06/28/2024] [Indexed: 06/30/2024]
Abstract
Reducing the radiation dose leads to the x-ray computed tomography (CT) images suffering from heavy noise and artifacts, which inevitably interferes with the subsequent clinic diagnostic and analysis. Leading works have explored diffusion models for low-dose CT imaging to avoid the structure degeneration and blurring effects of previous deep denoising models. However, most of them always begin their generative processes with Gaussian noise, which has little or no structure priors of the clean data distribution, thereby leading to long-time inference and unpleasant reconstruction quality. To alleviate these problems, this paper presents a Structure-Aware Diffusion model (SAD), an end-to-end self-guided learning framework for high-fidelity CT image reconstruction. First, SAD builds a nonlinear diffusion bridge between clean and degraded data distributions, which could directly learn the implicit physical degradation prior from observed measurements. Second, SAD integrates the prompt learning mechanism and implicit neural representation into the diffusion process, where rich and diverse structure representations extracted by degraded inputs are exploited as prompts, which provides global and local structure priors, to guide CT image reconstruction. Finally, we devise an efficient self-guided diffusion architecture using an iterative updated strategy, which further refines structural prompts during each generative step to drive finer image reconstruction. Extensive experiments on AAPM-Mayo and LoDoPaB-CT datasets demonstrate that our SAD could achieve superior performance in terms of noise removal, structure preservation, and blind-dose generalization, with few generative steps, even one step only.
Collapse
Affiliation(s)
- Wenchao Du
- College of Computer Science, Sichuan University, Chengdu 610065, People's Republic of China
| | - HuanHuan Cui
- West China Hospital of Sichuan University, Chengdu 610041, People's Republic of China
| | - LinChao He
- College of Computer Science, Sichuan University, Chengdu 610065, People's Republic of China
| | - Hu Chen
- College of Computer Science, Sichuan University, Chengdu 610065, People's Republic of China
| | - Yi Zhang
- College of Computer Science, Sichuan University, Chengdu 610065, People's Republic of China
| | - Hongyu Yang
- College of Computer Science, Sichuan University, Chengdu 610065, People's Republic of China
| |
Collapse
|
12
|
Baldeon-Calisto M, Rivera-Velastegui F, Lai-Yuen SK, Riofrío D, Pérez-Pérez N, Benítez D, Flores-Moyano R. DistilIQA: Distilling Vision Transformers for no-reference perceptual CT image quality assessment. Comput Biol Med 2024; 177:108670. [PMID: 38838558 DOI: 10.1016/j.compbiomed.2024.108670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Revised: 04/25/2024] [Accepted: 05/26/2024] [Indexed: 06/07/2024]
Abstract
No-reference image quality assessment (IQA) is a critical step in medical image analysis, with the objective of predicting perceptual image quality without the need for a pristine reference image. The application of no-reference IQA to CT scans is valuable in providing an automated and objective approach to assessing scan quality, optimizing radiation dose, and improving overall healthcare efficiency. In this paper, we introduce DistilIQA, a novel distilled Vision Transformer network designed for no-reference CT image quality assessment. DistilIQA integrates convolutional operations and multi-head self-attention mechanisms by incorporating a powerful convolutional stem at the beginning of the traditional ViT network. Additionally, we present a two-step distillation methodology aimed at improving network performance and efficiency. In the initial step, a "teacher ensemble network" is constructed by training five vision Transformer networks using a five-fold division schema. In the second step, a "student network", comprising of a single Vision Transformer, is trained using the original labeled dataset and the predictions generated by the teacher network as new labels. DistilIQA is evaluated in the task of quality score prediction from low-dose chest CT scans obtained from the LDCT and Projection data of the Cancer Imaging Archive, along with low-dose abdominal CT images from the LDCTIQAC2023 Grand Challenge. Our results demonstrate DistilIQA's remarkable performance in both benchmarks, surpassing the capabilities of various CNNs and Transformer architectures. Moreover, our comprehensive experimental analysis demonstrates the effectiveness of incorporating convolutional operations within the ViT architecture and highlights the advantages of our distillation methodology.
Collapse
Affiliation(s)
- Maria Baldeon-Calisto
- Departamento de Ingeniería Industrial and Instituto de Innovación en Productividad y Logística CATENA-USFQ, Universidad San Francisco de Quito USFQ, Quito, 170157, Ecuador; Colegio de Ciencias e Ingenierías "El Politécnico", Universidad San Francisco de Quito USFQ, Quito, 170157, Ecuador.
| | | | - Susana K Lai-Yuen
- Department of Industrial and Management Systems Engineering, University of South Florida, Tampa, 33620, FL, USA.
| | - Daniel Riofrío
- Colegio de Ciencias e Ingenierías "El Politécnico", Universidad San Francisco de Quito USFQ, Quito, 170157, Ecuador.
| | - Noel Pérez-Pérez
- Colegio de Ciencias e Ingenierías "El Politécnico", Universidad San Francisco de Quito USFQ, Quito, 170157, Ecuador.
| | - Diego Benítez
- Colegio de Ciencias e Ingenierías "El Politécnico", Universidad San Francisco de Quito USFQ, Quito, 170157, Ecuador.
| | - Ricardo Flores-Moyano
- Colegio de Ciencias e Ingenierías "El Politécnico", Universidad San Francisco de Quito USFQ, Quito, 170157, Ecuador.
| |
Collapse
|
13
|
彭 声, 王 永, 边 兆, 马 建, 黄 静. [A dual-domain cone beam computed tomography reconstruction framework with improved differentiable domain transform for cone-angle artifact correction]. NAN FANG YI KE DA XUE XUE BAO = JOURNAL OF SOUTHERN MEDICAL UNIVERSITY 2024; 44:1188-1197. [PMID: 38977350 PMCID: PMC11237300 DOI: 10.12122/j.issn.1673-4254.2024.06.21] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Indexed: 07/10/2024]
Abstract
OBJECTIVE We propose a dual-domain cone beam computed tomography (CBCT) reconstruction framework DualCBR-Net based on improved differentiable domain transform for cone-angle artifact correction. METHODS The proposed CBCT dual-domain reconstruction framework DualCBR-Net consists of 3 individual modules: projection preprocessing, differentiable domain transform, and image post-processing. The projection preprocessing module first extends the original projection data in the row direction to ensure full coverage of the scanned object by X-ray. The differentiable domain transform introduces the FDK reconstruction and forward projection operators to complete the forward and gradient backpropagation processes, where the geometric parameters correspond to the extended data dimension to provide crucial prior information in the forward pass of the network and ensure the accuracy in the gradient backpropagation, thus enabling precise learning of cone-beam region data. The image post-processing module further fine-tunes the domain-transformed image to remove residual artifacts and noises. RESULTS The results of validation experiments conducted on Mayo's public chest dataset showed that the proposed DualCBR-Net framework was superior to other comparison methods in terms of artifact removal and structural detail preservation. Compared with the latest methods, the DualCBR-Net framework improved the PSNR and SSIM by 0.6479 and 0.0074, respectively. CONCLUSION The proposed DualCBR-Net framework for cone-angle artifact correction allows effective joint training of the CBCT dual-domain network and is especially effective for large cone-angle region.
Collapse
|
14
|
汪 辰, 蒙 铭, 李 明, 王 永, 曾 栋, 边 兆, 马 建. [Reconstruction from CT truncated data based on dual-domain transformer coupled feature learning]. NAN FANG YI KE DA XUE XUE BAO = JOURNAL OF SOUTHERN MEDICAL UNIVERSITY 2024; 44:950-959. [PMID: 38862453 PMCID: PMC11166716 DOI: 10.12122/j.issn.1673-4254.2024.05.17] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Indexed: 06/13/2024]
Abstract
OBJECTIVE To propose a CT truncated data reconstruction model (DDTrans) based on projection and image dualdomain Transformer coupled feature learning for reducing truncation artifacts and image structure distortion caused by insufficient field of view (FOV) in CT scanning. METHODS Transformer was adopted to build projection domain and image domain restoration models, and the long-range dependency modeling capability of the Transformer attention module was used to capture global structural features to restore the projection data information and enhance the reconstructed images. We constructed a differentiable Radon back-projection operator layer between the projection domain and image domain networks to enable end-to-end training of DDTrans. Projection consistency loss was introduced to constrain the image forwardprojection results to further improve the accuracy of image reconstruction. RESULTS The experimental results with Mayo simulation data showed that for both partial truncation and interior scanning data, the proposed DDTrans method showed better performance than the comparison algorithms in removing truncation artifacts at the edges and restoring the external information of the FOV. CONCLUSION The DDTrans method can effectively remove CT truncation artifacts to ensure accurate reconstruction of the data within the FOV and achieve approximate reconstruction of data outside the FOV.
Collapse
|
15
|
Chen Z, Niu C, Gao Q, Wang G, Shan H. LIT-Former: Linking In-Plane and Through-Plane Transformers for Simultaneous CT Image Denoising and Deblurring. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1880-1894. [PMID: 38194396 DOI: 10.1109/tmi.2024.3351723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2024]
Abstract
This paper studies 3D low-dose computed tomography (CT) imaging. Although various deep learning methods were developed in this context, typically they focus on 2D images and perform denoising due to low-dose and deblurring for super-resolution separately. Up to date, little work was done for simultaneous in-plane denoising and through-plane deblurring, which is important to obtain high-quality 3D CT images with lower radiation and faster imaging speed. For this task, a straightforward method is to directly train an end-to-end 3D network. However, it demands much more training data and expensive computational costs. Here, we propose to link in-plane and through-plane transformers for simultaneous in-plane denoising and through-plane deblurring, termed as LIT-Former, which can efficiently synergize in-plane and through-plane sub-tasks for 3D CT imaging and enjoy the advantages of both convolution and transformer networks. LIT-Former has two novel designs: efficient multi-head self-attention modules (eMSM) and efficient convolutional feed-forward networks (eCFN). First, eMSM integrates in-plane 2D self-attention and through-plane 1D self-attention to efficiently capture global interactions of 3D self-attention, the core unit of transformer networks. Second, eCFN integrates 2D convolution and 1D convolution to extract local information of 3D convolution in the same fashion. As a result, the proposed LIT-Former synergizes these two sub-tasks, significantly reducing the computational complexity as compared to 3D counterparts and enabling rapid convergence. Extensive experimental results on simulated and clinical datasets demonstrate superior performance over state-of-the-art models. The source code is made available at https://github.com/hao1635/LIT-Former.
Collapse
|
16
|
Li X, Jing K, Yang Y, Wang Y, Ma J, Zheng H, Xu Z. Noise-Generating and Imaging Mechanism Inspired Implicit Regularization Learning Network for Low Dose CT Reconstrution. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1677-1689. [PMID: 38145543 DOI: 10.1109/tmi.2023.3347258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Low-dose computed tomography (LDCT) helps to reduce radiation risks in CT scanning while maintaining image quality, which involves a consistent pursuit of lower incident rays and higher reconstruction performance. Although deep learning approaches have achieved encouraging success in LDCT reconstruction, most of them treat the task as a general inverse problem in either the image domain or the dual (sinogram and image) domains. Such frameworks have not considered the original noise generation of the projection data and suffer from limited performance improvement for the LDCT task. In this paper, we propose a novel reconstruction model based on noise-generating and imaging mechanism in full-domain, which fully considers the statistical properties of intrinsic noises in LDCT and prior information in sinogram and image domains. To solve the model, we propose an optimization algorithm based on the proximal gradient technique. Specifically, we derive the approximate solutions of the integer programming problem on the projection data theoretically. Instead of hand-crafting the sinogram and image regularizers, we propose to unroll the optimization algorithm to be a deep network. The network implicitly learns the proximal operators of sinogram and image regularizers with two deep neural networks, providing a more interpretable and effective reconstruction procedure. Numerical results demonstrate our proposed method improvements of > 2.9 dB in peak signal to noise ratio, > 1.4% promotion in structural similarity metric, and > 9 HU decrements in root mean square error over current state-of-the-art LDCT methods.
Collapse
|
17
|
Ji X, Zhuo X, Lu Y, Mao W, Zhu S, Quan G, Xi Y, Lyu T, Chen Y. Image Domain Multi-Material Decomposition Noise Suppression Through Basis Transformation and Selective Filtering. IEEE J Biomed Health Inform 2024; 28:2891-2903. [PMID: 38363665 DOI: 10.1109/jbhi.2023.3348135] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/18/2024]
Abstract
Spectral CT can provide material characterization ability to offer more precise material information for diagnosis purposes. However, the material decomposition process generally leads to amplification of noise which significantly limits the utility of the material basis images. To mitigate such problem, an image domain noise suppression method was proposed in this work. The method performs basis transformation of the material basis images based on a singular value decomposition. The noise variances of the original spectral CT images were incorporated in the matrix to be decomposed to ensure that the transformed basis images are statistically uncorrelated. Due to the difference in noise amplitudes in the transformed basis images, a selective filtering method was proposed with the low-noise transformed basis image as guidance. The method was evaluated using both numerical simulation and real clinical dual-energy CT data. Results demonstrated that compared with existing methods, the proposed method performs better in preserving the spatial resolution and the soft tissue contrast while suppressing the image noise. The proposed method is also computationally efficient and can realize real-time noise suppression for clinical spectral CT images.
Collapse
|
18
|
Lee J, Baek J. Iterative reconstruction for limited-angle CT using implicit neural representation. Phys Med Biol 2024; 69:105008. [PMID: 38593820 DOI: 10.1088/1361-6560/ad3c8e] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2023] [Accepted: 04/09/2024] [Indexed: 04/11/2024]
Abstract
Objective.Limited-angle computed tomography (CT) presents a challenge due to its ill-posed nature. In such scenarios, analytical reconstruction methods often exhibit severe artifacts. To tackle this inverse problem, several supervised deep learning-based approaches have been proposed. However, they are constrained by limitations such as generalization issue and the difficulty of acquiring a large amount of paired CT images.Approach.In this work, we propose an iterative neural reconstruction framework designed for limited-angle CT. By leveraging a coordinate-based neural representation, we formulate tomographic reconstruction as a convex optimization problem involving a deep neural network. We then employ differentiable projection layer to optimize this network by minimizing the discrepancy between the predicted and measured projection data. In addition, we introduce a prior-based weight initialization method to ensure the network starts optimization with an informed initial guess. This strategic initialization significantly improves the quality of iterative reconstruction by stabilizing the divergent behavior in ill-posed neural fields. Our method operates in a self-supervised manner, thereby eliminating the need for extensive data.Main results.The proposed method outperforms other iterative and learning-based methods. Experimental results on XCAT and Mayo Clinic datasets demonstrate the effectiveness of our approach in restoring anatomical features as well as structures. This finding was substantiated by visual inspections and quantitative evaluations using NRMSE, PSNR, and SSIM. Moreover, we conduct a comprehensive investigation into the divergent behavior of iterative neural reconstruction, thus revealing its suboptimal convergence when starting from scratch. In contrast, our method consistently produced accurate images by incorporating an initial estimate as informed initialization.Significance.This work showcases the feasibility to reconstruct high-fidelity CT images from limited-angle x-ray projections. The proposed methodology introduces a novel data-free approach to enhance medical imaging, holding promise across various clinical applications.
Collapse
Affiliation(s)
- Jooho Lee
- Department of Artificial Intelligence, Yonsei University, Seoul, Republic of Korea
| | - Jongduk Baek
- Department of Artificial Intelligence, Yonsei University, Seoul, Republic of Korea
| |
Collapse
|
19
|
Wang L, Meng M, Chen S, Bian Z, Zeng D, Meng D, Ma J. Semi-supervised iterative adaptive network for low-dose CT sinogram recovery. Phys Med Biol 2024; 69:085013. [PMID: 38422540 DOI: 10.1088/1361-6560/ad2ee7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Accepted: 02/29/2024] [Indexed: 03/02/2024]
Abstract
Background.Concern has been expressed regarding the risk of carcinogenesis from medical computed tomography (CT) radiation. Lowering radiation in CT without appropriate modifications often leads to severe noise-induced artifacts in the images. The utilization of deep learning (DL) techniques has achieved promising reconstruction performance in low-dose CT (LDCT) imaging. However, most DL-based algorithms require the pre-collection of a large set of image pairs (low-dose/standard-dose) and the training of networks in an end-to-end supervised manner. Meanwhile, securing such a large volume of paired, well-registered training data in clinical practice is challenging. Moreover, these algorithms often overlook the potential to utilize the abundant information in a large collection of LDCT-only images/sinograms.Methods.In this paper, we introduce a semi-supervised iterative adaptive network (SIA-Net) for LDCT imaging, utilizing both labeled and unlabeled sinograms in a cohesive network framework, integrating supervised and unsupervised learning processes. Specifically, the supervised process captures critical features (i.e. noise distribution and tissue characteristics) latent in the paired sinograms, while the unsupervised process effectively learns these features in the unlabeled low-dose sinograms, employing a conventional weighted least-squares model with a regularization term. Furthermore, the SIA-Net method is designed to adaptively transfer the learned feature distribution from the supervised to the unsupervised process, thereby obtaining a high-fidelity sinogram through iterative adaptive learning. Finally, high-quality CT images can be reconstructed from the refined sinogram using the filtered back-projection algorithm.Results.Experimental results on two clinical datasets indicate that the proposed SIA-Net method achieves competitive performance in terms of noise reduction and structure preservation in LDCT imaging, when compared to traditional supervised learning methods.
Collapse
Affiliation(s)
- Lei Wang
- School of Future Technology, Xi'an Jiaotong University, Xi'an 710049, People's Republic of China
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Mingqiang Meng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
- Pazhou Lab (Huangpu), Guangdong, People's Republic of China
| | - Shixuan Chen
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Zhaoying Bian
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
- Pazhou Lab (Huangpu), Guangdong, People's Republic of China
| | - Dong Zeng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
- Pazhou Lab (Huangpu), Guangdong, People's Republic of China
- Department of Radiology, Zhujiang Hospital, Southern Medical University, Guangdong, People's Republic of China
| | - Deyu Meng
- School of Mathematics and Statistics, Xi'an Jiaotong University, Xi'an 710049, People's Republic of China
| | - Jianhua Ma
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
- School of Life Science and Technology, Xi'an Jiaotong University, Xi'an, Shaanxi, People's Republic of China
| |
Collapse
|
20
|
Han M, Baek J. Direct estimation of the noise power spectrum from patient data to generate synthesized CT noise for denoising network training. Med Phys 2024; 51:1637-1652. [PMID: 38289987 DOI: 10.1002/mp.16963] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Revised: 12/12/2023] [Accepted: 01/18/2024] [Indexed: 02/01/2024] Open
Abstract
BACKGROUND Developing a deep-learning network for denoising low-dose CT (LDCT) images necessitates paired computed tomography (CT) images acquired at different dose levels. However, it is challenging to obtain these images from the same patient. PURPOSE In this study, we introduce a novel approach to generate CT images at different dose levels. METHODS Our method involves the direct estimation of the quantum noise power spectrum (NPS) from patient CT images without the need for prior information. By modeling the anatomical NPS using a power-law function and estimating the quantum NPS from the measured NPS after removing the anatomical NPS, we create synthesized quantum noise by applying the estimated quantum NPS as a filter to random noise. By adding synthesized noise to CT images, synthesized CT images can be generated as if these are obtained at a lower dose. This leads to the generation of paired images at different dose levels for training denoising networks. RESULTS The proposed method accurately estimates the reference quantum NPS. The denoising network trained with paired data generated using synthesized quantum noise achieves denoising performance comparable to networks trained using Mayo Clinic data, as justified by the mean-squared-error (MSE), structural similarity index (SSIM)and peak signal-to-noise ratio (PSNR) scores. CONCLUSIONS This approach offers a promising solution for LDCT image denoising network development without the need for multiple scans of the same patient at different doses.
Collapse
Affiliation(s)
- Minah Han
- Department of Artificial Intelligence, Yonsei University, Seoul, South Korea
- Bareunex Imaging Inc., Incheon, South Korea
| | - Jongduk Baek
- Department of Artificial Intelligence, Yonsei University, Seoul, South Korea
- Bareunex Imaging Inc., Incheon, South Korea
| |
Collapse
|
21
|
Zhang J, Gong W, Ye L, Wang F, Shangguan Z, Cheng Y. A Review of deep learning methods for denoising of medical low-dose CT images. Comput Biol Med 2024; 171:108112. [PMID: 38387380 DOI: 10.1016/j.compbiomed.2024.108112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Revised: 01/18/2024] [Accepted: 02/04/2024] [Indexed: 02/24/2024]
Abstract
To prevent patients from being exposed to excess of radiation in CT imaging, the most common solution is to decrease the radiation dose by reducing the X-ray, and thus the quality of the resulting low-dose CT images (LDCT) is degraded, as evidenced by more noise and streaking artifacts. Therefore, it is important to maintain high quality CT image while effectively reducing radiation dose. In recent years, with the rapid development of deep learning technology, deep learning-based LDCT denoising methods have become quite popular because of their data-driven and high-performance features to achieve excellent denoising results. However, to our knowledge, no relevant article has so far comprehensively introduced and reviewed advanced deep learning denoising methods such as Transformer structures in LDCT denoising tasks. Therefore, based on the literatures related to LDCT image denoising published from year 2016-2023, and in particular from 2020 to 2023, this study presents a systematic survey of current situation, and challenges and future research directions in LDCT image denoising field. Four types of denoising networks are classified according to the network structure: CNN-based, Encoder-Decoder-based, GAN-based, and Transformer-based denoising networks, and each type of denoising network is described and summarized from the perspectives of structural features and denoising performances. Representative deep-learning denoising methods for LDCT are experimentally compared and analyzed. The study results show that CNN-based denoising methods capture image details efficiently through multi-level convolution operation, demonstrating superior denoising effects and adaptivity. Encoder-decoder networks with MSE loss, achieve outstanding results in objective metrics. GANs based methods, employing innovative generators and discriminators, obtain denoised images that exhibit perceptually a closeness to NDCT. Transformer-based methods have potential for improving denoising performances due to their powerful capability in capturing global information. Challenges and opportunities for deep learning based LDCT denoising are analyzed, and future directions are also presented.
Collapse
Affiliation(s)
- Ju Zhang
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, China.
| | - Weiwei Gong
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Lieli Ye
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, China.
| | - Fanghong Wang
- Zhijiang College, Zhejiang University of Technology, Shaoxing, China.
| | - Zhibo Shangguan
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Yun Cheng
- Department of Medical Imaging, Zhejiang Hospital, Hangzhou, China.
| |
Collapse
|
22
|
Kim W, Lee J, Choi JH. An unsupervised two-step training framework for low-dose computed tomography denoising. Med Phys 2024; 51:1127-1144. [PMID: 37432026 DOI: 10.1002/mp.16628] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2023] [Revised: 06/25/2023] [Accepted: 06/25/2023] [Indexed: 07/12/2023] Open
Abstract
BACKGROUND Although low-dose computed tomography (CT) imaging has been more widely adopted in clinical practice to reduce radiation exposure to patients, the reconstructed CT images tend to have more noise, which impedes accurate diagnosis. Recently, deep neural networks using convolutional neural networks to reduce noise in the reconstructed low-dose CT images have shown considerable improvement. However, they need a large number of paired normal- and low-dose CT images to fully train the network via supervised learning methods. PURPOSE To propose an unsupervised two-step training framework for image denoising that uses low-dose CT images of one dataset and unpaired high-dose CT images from another dataset. METHODS Our proposed framework trains the denoising network in two steps. In the first training step, we train the network using 3D volumes of CT images and predict the center CT slice from them. This pre-trained network is used in the second training step to train the denoising network and is combined with the memory-efficient denoising generative adversarial network (DenoisingGAN), which further enhances both objective and perceptual quality. RESULTS The experimental results on phantom and clinical datasets show superior performance over the existing traditional machine learning and self-supervised deep learning methods, and the results are comparable to the fully supervised learning methods. CONCLUSIONS We proposed a new unsupervised learning framework for low-dose CT denoising, convincingly improving noisy CT images from both objective and perceptual quality perspectives. Because our denoising framework does not require physics-based noise models or system-dependent assumptions, our proposed method can be easily reproduced; consequently, it can also be generally applicable to various CT scanners or dose levels.
Collapse
Affiliation(s)
- Wonjin Kim
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| | - Jaayeon Lee
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| | - Jang-Hwan Choi
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| |
Collapse
|
23
|
Nelson BJ, Kc P, Badal A, Jiang L, Masters SC, Zeng R. Pediatric evaluations for deep learning CT denoising. Med Phys 2024; 51:978-990. [PMID: 38127330 DOI: 10.1002/mp.16901] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Revised: 11/13/2023] [Accepted: 12/04/2023] [Indexed: 12/23/2023] Open
Abstract
BACKGROUND Deep learning (DL) CT denoising models have the potential to improve image quality for lower radiation dose exams. These models are generally trained with large quantities of adult patient image data. However, CT, and increasingly DL denoising methods, are used in both adult and pediatric populations. Pediatric body habitus and size can differ significantly from adults and vary dramatically from newborns to adolescents. Ensuring that pediatric subgroups of different body sizes are not disadvantaged by DL methods requires evaluations capable of assessing performance in each subgroup. PURPOSE To assess DL CT denoising in pediatric and adult-sized patients, we built a framework of computer simulated image quality (IQ) control phantoms and evaluation methodology. METHODS The computer simulated IQ phantoms in the framework featured pediatric-sized versions of standard CatPhan 600 and MITA-LCD phantoms with a range of diameters matching the mean effective diameters of pediatric patients ranging from newborns to 18 years old. These phantoms were used in simulating CT images that were then inputs for a DL denoiser to evaluate performance in different sized patients. Adult CT test images were simulated using standard-sized phantoms scanned with adult scan protocols. Pediatric CT test images were simulated with pediatric-sized phantoms and adjusted pediatric protocols. The framework's evaluation methodology consisted of denoising both adult and pediatric test images then assessing changes in image quality, including noise, image sharpness, CT number accuracy, and low contrast detectability. To demonstrate the use of the framework, a REDCNN denoising model trained on adult patient images was evaluated. To validate that the DL model performance measured with the proposed pediatric IQ phantoms was representative of performance in more realistic patient anatomy, anthropomorphic pediatric XCAT phantoms of the same age range were also used to compare noise reduction performance. RESULTS Using the proposed pediatric-sized IQ phantom framework, size differences between adult and pediatric-sized phantoms were observed to substantially influence the adult trained DL denoising model's performance. When applied to adult images, the DL model achieved a 60% reduction in noise standard deviation without substantial loss in sharpness in mid or high spatial frequencies. However, in smaller phantoms the denoising performance dropped due to different image noise textures resulting from the smaller field of view (FOV) between adult and pediatric protocols. In the validation study, noise reduction trends in the pediatric-sized IQ phantoms were found to be consistent with those found in anthropomorphic phantoms. CONCLUSION We developed a framework of using pediatric-sized IQ phantoms for pediatric subgroup evaluation of DL denoising models. Using the framework, we found the performance of an adult trained DL denoiser did not generalize well in the smaller diameter phantoms corresponding to younger pediatric patient sizes. Our work suggests noise texture differences from FOV changes between adult and pediatric protocols can contribute to poor generalizability in DL denoising and that the proposed framework is an effective means to identify these performance disparities for a given model.
Collapse
Affiliation(s)
- Brandon J Nelson
- Center for Devices and Radiological Health, Office of Science and Engineering Labs, Division of Imaging, Diagnostics, and Software Reliability, U.S. Food and Drug Administration, Silver Spring, Maryland, USA
| | - Prabhat Kc
- Center for Devices and Radiological Health, Office of Science and Engineering Labs, Division of Imaging, Diagnostics, and Software Reliability, U.S. Food and Drug Administration, Silver Spring, Maryland, USA
| | - Andreu Badal
- Center for Devices and Radiological Health, Office of Science and Engineering Labs, Division of Imaging, Diagnostics, and Software Reliability, U.S. Food and Drug Administration, Silver Spring, Maryland, USA
| | - Lu Jiang
- Center for Devices and Radiological Health, Office of Product Evaluation and Quality, Office of Radiological Health, U.S. Food and Drug Administration, Silver Spring, Maryland, USA
| | - Shane C Masters
- Center for Drug Evaluation and Research, Office of Specialty Medicine, Division of Imaging and Radiation Medicine, U.S. Food and Drug Administration, Silver Spring, Maryland, USA
| | - Rongping Zeng
- Center for Devices and Radiological Health, Office of Science and Engineering Labs, Division of Imaging, Diagnostics, and Software Reliability, U.S. Food and Drug Administration, Silver Spring, Maryland, USA
| |
Collapse
|
24
|
Sadia RT, Chen J, Zhang J. CT image denoising methods for image quality improvement and radiation dose reduction. J Appl Clin Med Phys 2024; 25:e14270. [PMID: 38240466 PMCID: PMC10860577 DOI: 10.1002/acm2.14270] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Revised: 12/15/2023] [Accepted: 12/28/2023] [Indexed: 02/13/2024] Open
Abstract
With the ever-increasing use of computed tomography (CT), concerns about its radiation dose have become a significant public issue. To address the need for radiation dose reduction, CT denoising methods have been widely investigated and applied in low-dose CT images. Numerous noise reduction algorithms have emerged, such as iterative reconstruction and most recently, deep learning (DL)-based approaches. Given the rapid advancements in Artificial Intelligence techniques, we recognize the need for a comprehensive review that emphasizes the most recently developed methods. Hence, we have performed a thorough analysis of existing literature to provide such a review. Beyond directly comparing the performance, we focus on pivotal aspects, including model training, validation, testing, generalizability, vulnerability, and evaluation methods. This review is expected to raise awareness of the various facets involved in CT image denoising and the specific challenges in developing DL-based models.
Collapse
Affiliation(s)
- Rabeya Tus Sadia
- Department of Computer ScienceUniversity of KentuckyLexingtonKentuckyUSA
| | - Jin Chen
- Department of Medicine‐NephrologyUniversity of Alabama at BirminghamBirminghamAlabamaUSA
| | - Jie Zhang
- Department of RadiologyUniversity of KentuckyLexingtonKentuckyUSA
| |
Collapse
|
25
|
Azad R, Kazerouni A, Heidari M, Aghdam EK, Molaei A, Jia Y, Jose A, Roy R, Merhof D. Advances in medical image analysis with vision Transformers: A comprehensive review. Med Image Anal 2024; 91:103000. [PMID: 37883822 DOI: 10.1016/j.media.2023.103000] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 09/30/2023] [Accepted: 10/11/2023] [Indexed: 10/28/2023]
Abstract
The remarkable performance of the Transformer architecture in natural language processing has recently also triggered broad interest in Computer Vision. Among other merits, Transformers are witnessed as capable of learning long-range dependencies and spatial correlations, which is a clear advantage over convolutional neural networks (CNNs), which have been the de facto standard in Computer Vision problems so far. Thus, Transformers have become an integral part of modern medical image analysis. In this review, we provide an encyclopedic review of the applications of Transformers in medical imaging. Specifically, we present a systematic and thorough review of relevant recent Transformer literature for different medical image analysis tasks, including classification, segmentation, detection, registration, synthesis, and clinical report generation. For each of these applications, we investigate the novelty, strengths and weaknesses of the different proposed strategies and develop taxonomies highlighting key properties and contributions. Further, if applicable, we outline current benchmarks on different datasets. Finally, we summarize key challenges and discuss different future research directions. In addition, we have provided cited papers with their corresponding implementations in https://github.com/mindflow-institue/Awesome-Transformer.
Collapse
Affiliation(s)
- Reza Azad
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Amirhossein Kazerouni
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Moein Heidari
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | | | - Amirali Molaei
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Yiwei Jia
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Abin Jose
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Rijo Roy
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Dorit Merhof
- Faculty of Informatics and Data Science, University of Regensburg, Regensburg, Germany; Fraunhofer Institute for Digital Medicine MEVIS, Bremen, Germany.
| |
Collapse
|
26
|
Yun S, Jeong U, Lee D, Kim H, Cho S. Image quality improvement in bowtie-filter-equipped cone-beam CT using a dual-domain neural network. Med Phys 2023; 50:7498-7512. [PMID: 37669510 DOI: 10.1002/mp.16693] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Revised: 08/08/2023] [Accepted: 08/09/2023] [Indexed: 09/07/2023] Open
Abstract
BACKGROUND The bowtie-filter in cone-beam CT (CBCT) causes spatially nonuniform x-ray beam often leading to eclipse artifacts in the reconstructed image. The artifacts are further confounded by the patient scatter, which is therefore patient-dependent as well as system-specific. PURPOSE In this study, we propose a dual-domain network for reducing the bowtie-filter-induced artifacts in CBCT images. METHODS In the projection domain, the network compensates for the filter-induced beam-hardening that are highly related to the eclipse artifacts. The output of the projection-domain network was used for image reconstruction and the reconstructed images were fed into the image-domain network. In the image domain, the network further reduces the remaining cupping artifacts that are associated with the scatter. A single image-domain-only network was also implemented for comparison. RESULTS The proposed approach successfully enhanced soft-tissue contrast with much-reduced image artifacts. In the numerical study, the proposed method decreased perceptual loss and root-mean-square-error (RMSE) of the images by 84.5% and 84.9%, respectively, and increased the structure similarity index measure (SSIM) by 0.26 compared to the original input images on average. In the experimental study, the proposed method decreased perceptual loss and RMSE of the images by 87.2% and 92.1%, respectively, and increased SSIM by 0.58 compared to the original input images on average. CONCLUSIONS We have proposed a deep-learning-based dual-domain framework to reduce the bowtie-filter artifacts and to increase the soft-tissue contrast in CBCT images. The performance of the proposed method has been successfully demonstrated in both numerical and experimental studies.
Collapse
Affiliation(s)
- Sungho Yun
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
| | - Uijin Jeong
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
| | - Donghyeon Lee
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
| | - Hyeongseok Kim
- KAIST Institute for Artificial Intelligence, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
| | - Seungryong Cho
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
- KAIST Institute for Artificial Intelligence, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
- KAIST Institute for Health Science and Technology, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
- KAIST Institute for IT Convergence, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South Korea
| |
Collapse
|
27
|
Patel R, Provenzano D, Loew M. Anonymization and validation of three-dimensional volumetric renderings of computed tomography data using commercially available T1-weighted magnetic resonance imaging-based algorithms. J Med Imaging (Bellingham) 2023; 10:066501. [PMID: 38074629 PMCID: PMC10704182 DOI: 10.1117/1.jmi.10.6.066501] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2023] [Revised: 11/03/2023] [Accepted: 11/07/2023] [Indexed: 02/12/2024] Open
Abstract
Purpose Previous studies have demonstrated that three-dimensional (3D) volumetric renderings of magnetic resonance imaging (MRI) brain data can be used to identify patients using facial recognition. We have shown that facial features can be identified on simulation-computed tomography (CT) images for radiation oncology and mapped to face images from a database. We aim to determine whether CT images can be anonymized using anonymization software that was designed for T1-weighted MRI data. Approach Our study examines (1) the ability of off-the-shelf anonymization algorithms to anonymize CT data and (2) the ability of facial recognition algorithms to identify whether faces could be detected from a database of facial images. Our study generated 3D renderings from 57 head CT scans from The Cancer Imaging Archive database. Data were anonymized using AFNI (deface, reface, and 3Dskullstrip) and FSL's BET. Anonymized data were compared to the original renderings and passed through facial recognition algorithms (VGG-Face, FaceNet, DLib, and SFace) using a facial database (labeled faces in the wild) to determine what matches could be found. Results Our study found that all modules were able to process CT data and that AFNI's 3Dskullstrip and FSL's BET data consistently showed lower reidentification rates compared to the original. Conclusions The results from this study highlight the potential usage of anonymization algorithms as a clinical standard for deidentifying brain CT data. Our study demonstrates the importance of continued vigilance for patient privacy in publicly shared datasets and the importance of continued evaluation of anonymization methods for CT data.
Collapse
Affiliation(s)
- Rahil Patel
- George Washington University School of Engineering and Applied Science, Department of Biomedical Engineering, Washington, District of Columbia, United States
| | - Destie Provenzano
- George Washington University School of Engineering and Applied Science, Department of Biomedical Engineering, Washington, District of Columbia, United States
| | - Murray Loew
- George Washington University School of Engineering and Applied Science, Department of Biomedical Engineering, Washington, District of Columbia, United States
| |
Collapse
|
28
|
Li H, Yang X, Yang S, Wang D, Jeon G. Transformer With Double Enhancement for Low-Dose CT Denoising. IEEE J Biomed Health Inform 2023; 27:4660-4671. [PMID: 36279348 DOI: 10.1109/jbhi.2022.3216887] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/06/2023]
Abstract
Increasingly serious health problems have made the usage of computed tomography surge. Therefore, algorithms for processing CT images are becoming more and more abundant. These algorithms can lessen the harm of cumulative radiation in CT technology for the patient while eliminating the noise of image caused by dose reduction. However, the mainstream CNN-based algorithms are inefficient when dealing with features in broad regions. Inspired by the large receptive field of transformer framework, this paper designs an end-to-end low-dose CT (LDCT) denoising network based on the transformer. The overall network contains a main branch and dual side branches. Specifically, the overlapping-free window-based self-attention transformer block is adopted on the main branch to realize image denoising. On the dual side branches, we propose double enhancement module to enrich edge, texture, and context information of LDCT images. Meanwhile, the receptive field of network is further enlarged after processing, which is helpful for building model's long-range dependencies. The outputs of the side branches are concatenated for enhancing information and generating high-quality CT images. In addition, to better train the network, we introduce a compound loss function including mean squared error (MSE), multi-scale perceptual (MSP), and Sobel-L1 (SL) to make the denoised image closer to the targeted norm-dose CT (NDCT) image. Lastly, we conducted experiments on two clinical datasets including abdomen, head, and chest LDCT images with 25%, 25%, and 10% of the full dose, respectively. The experimental results demonstrated that the proposed DEformer achieved better denoising performance than the existing algorithms.
Collapse
|
29
|
Lepcha DC, Dogra A, Goyal B, Goyal V, Kukreja V, Bavirisetti DP. A constructive non-local means algorithm for low-dose computed tomography denoising with morphological residual processing. PLoS One 2023; 18:e0291911. [PMID: 37756296 PMCID: PMC10529561 DOI: 10.1371/journal.pone.0291911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2023] [Accepted: 09/08/2023] [Indexed: 09/29/2023] Open
Abstract
Low-dose computed tomography (LDCT) has attracted significant attention in the domain of medical imaging due to the inherent risks of normal-dose computed tomography (NDCT) based X-ray radiations to patients. However, reducing radiation dose in CT imaging produces noise and artifacts that degrade image quality and subsequently hinders medical disease diagnostic performance. In order to address these problems, this research article presents a competent low-dose computed tomography image denoising algorithm based on a constructive non-local means algorithm with morphological residual processing to achieve the task of removing noise from the LDCT images. We propose an innovative constructive non-local image filtering algorithm by means of applications in low-dose computed tomography technology. The nonlocal mean filter that was recently proposed was modified to construct our denoising algorithm. It constructs the discrete property of neighboring filtering to enable rapid vectorized and parallel implantation in contemporary shared memory computer platforms while simultaneously decreases computing complexity. Subsequently, the proposed method performs faster computation compared to a non-vectorized and serial implementation in terms of speed and scales linearly with image dimension. In addition, the morphological residual processing is employed for the purpose of edge-preserving image processing. It combines linear lowpass filtering with a nonlinear technique that enables the extraction of meaningful regions where edges could be preserved while removing residual artifacts from the images. Experimental results demonstrate that the proposed algorithm preserves more textural and structural features while reducing noise, enhances edges and significantly improves image quality more effectively. The proposed research article obtains better results both qualitatively and quantitively when compared to other comparative algorithms on publicly accessible datasets.
Collapse
Affiliation(s)
| | - Ayush Dogra
- Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, India
| | - Bhawna Goyal
- Department of ECE and UCRD, Chandigarh University, Mohali, Punjab, India
| | | | - Vinay Kukreja
- Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, India
| | - Durga Prasad Bavirisetti
- Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| |
Collapse
|
30
|
Kiss MB, Coban SB, Batenburg KJ, van Leeuwen T, Lucka F. 2DeteCT - A large 2D expandable, trainable, experimental Computed Tomography dataset for machine learning. Sci Data 2023; 10:576. [PMID: 37666897 PMCID: PMC10477177 DOI: 10.1038/s41597-023-02484-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Accepted: 08/16/2023] [Indexed: 09/06/2023] Open
Abstract
Recent research in computational imaging largely focuses on developing machine learning (ML) techniques for image reconstruction, which requires large-scale training datasets consisting of measurement data and ground-truth images. However, suitable experimental datasets for X-ray Computed Tomography (CT) are scarce, and methods are often developed and evaluated only on simulated data. We fill this gap by providing the community with a versatile, open 2D fan-beam CT dataset suitable for developing ML techniques for a range of image reconstruction tasks. To acquire it, we designed a sophisticated, semi-automatic scan procedure that utilizes a highly-flexible laboratory X-ray CT setup. A diverse mix of samples with high natural variability in shape and density was scanned slice-by-slice (5,000 slices in total) with high angular and spatial resolution and three different beam characteristics: A high-fidelity, a low-dose and a beam-hardening-inflicted mode. In addition, 750 out-of-distribution slices were scanned with sample and beam variations to accommodate robustness and segmentation tasks. We provide raw projection data, reference reconstructions and segmentations based on an open-source data processing pipeline.
Collapse
Affiliation(s)
- Maximilian B Kiss
- Centrum Wiskunde & Informatica, Computational Imaging group, Amsterdam, 1098 XG, The Netherlands.
| | - Sophia B Coban
- Centrum Wiskunde & Informatica, Computational Imaging group, Amsterdam, 1098 XG, The Netherlands
- Department of Mathematics, University of Manchester, Oxford Road, Manchester, M13 9PL, United Kingdom
| | - K Joost Batenburg
- Centrum Wiskunde & Informatica, Computational Imaging group, Amsterdam, 1098 XG, The Netherlands
- Leiden University, LIACS, Leiden, 2300 RA, The Netherlands
| | - Tristan van Leeuwen
- Centrum Wiskunde & Informatica, Computational Imaging group, Amsterdam, 1098 XG, The Netherlands
- Utrecht University, Mathematical Institute, Utrecht, 3584 CD, The Netherlands
| | - Felix Lucka
- Centrum Wiskunde & Informatica, Computational Imaging group, Amsterdam, 1098 XG, The Netherlands.
| |
Collapse
|
31
|
Huang J, Chen K, Ren Y, Sun J, Wang Y, Tao T, Pu X. CDDnet: Cross-domain denoising network for low-dose CT image via local and global information alignment. Comput Biol Med 2023; 163:107219. [PMID: 37422942 DOI: 10.1016/j.compbiomed.2023.107219] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2023] [Revised: 05/21/2023] [Accepted: 06/25/2023] [Indexed: 07/11/2023]
Abstract
The domain shift problem has emerged as a challenge in cross-domain low-dose CT (LDCT) image denoising task, where the acquisition of a sufficient number of medical images from multiple sources may be constrained by privacy concerns. In this study, we propose a novel cross-domain denoising network (CDDnet) that incorporates both local and global information of CT images. To address the local component, a local information alignment module has been proposed to regularize the similarity between extracted target and source features from selected patches. To align the general information of the semantic structure from a global perspective, an autoencoder is adopted to learn the latent correlation between the source label and the estimated target label generated by the pre-trained denoiser. Experimental results demonstrate that our proposed CDDnet effectively alleviates the domain shift problem, outperforming other deep learning-based and domain adaptation-based methods under cross-domain scenarios.
Collapse
Affiliation(s)
- Jiaxin Huang
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Kecheng Chen
- Department of Electrical Engineering, City University of Hong Kong, 999077, Hong Kong Special Administrative Region of China
| | - Yazhou Ren
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China; Shenzhen Institute for Advanced Study, University of Electronic Science and Technology of China, Shenzhen, 518110, China
| | - Jiayu Sun
- West China Hospital, Sichuan University, Chengdu, 610044, China
| | - Yanmei Wang
- Institute of Traditional Chinese Medicine, Sichuan College of Traditional Chinese Medicine (Sichuan Second Hospital of TCM), Chengdu, 610075, China
| | - Tao Tao
- Institute of Traditional Chinese Medicine, Sichuan College of Traditional Chinese Medicine (Sichuan Second Hospital of TCM), Chengdu, 610075, China
| | - Xiaorong Pu
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China; Shenzhen Institute for Advanced Study, University of Electronic Science and Technology of China, Shenzhen, 518110, China; NHC Key Laboratory of Nuclear Technology Medical Transformation, Mianyang Central Hospital, Mianyang, 621000, China.
| |
Collapse
|
32
|
Liu Z, Lv Q, Yang Z, Li Y, Lee CH, Shen L. Recent progress in transformer-based medical image analysis. Comput Biol Med 2023; 164:107268. [PMID: 37494821 DOI: 10.1016/j.compbiomed.2023.107268] [Citation(s) in RCA: 10] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Revised: 05/30/2023] [Accepted: 07/16/2023] [Indexed: 07/28/2023]
Abstract
The transformer is primarily used in the field of natural language processing. Recently, it has been adopted and shows promise in the computer vision (CV) field. Medical image analysis (MIA), as a critical branch of CV, also greatly benefits from this state-of-the-art technique. In this review, we first recap the core component of the transformer, the attention mechanism, and the detailed structures of the transformer. After that, we depict the recent progress of the transformer in the field of MIA. We organize the applications in a sequence of different tasks, including classification, segmentation, captioning, registration, detection, enhancement, localization, and synthesis. The mainstream classification and segmentation tasks are further divided into eleven medical image modalities. A large number of experiments studied in this review illustrate that the transformer-based method outperforms existing methods through comparisons with multiple evaluation metrics. Finally, we discuss the open challenges and future opportunities in this field. This task-modality review with the latest contents, detailed information, and comprehensive comparison may greatly benefit the broad MIA community.
Collapse
Affiliation(s)
- Zhaoshan Liu
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Qiujie Lv
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore; School of Intelligent Systems Engineering, Sun Yat-sen University, No. 66, Gongchang Road, Guangming District, 518107, China.
| | - Ziduo Yang
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore; School of Intelligent Systems Engineering, Sun Yat-sen University, No. 66, Gongchang Road, Guangming District, 518107, China.
| | - Yifan Li
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Chau Hung Lee
- Department of Radiology, Tan Tock Seng Hospital, 11 Jalan Tan Tock Seng, Singapore, 308433, Singapore.
| | - Lei Shen
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| |
Collapse
|
33
|
Zhou Z, Inoue A, McCollough CH, Yu L. Self-trained deep convolutional neural network for noise reduction in CT. J Med Imaging (Bellingham) 2023; 10:044008. [PMID: 37636895 PMCID: PMC10449263 DOI: 10.1117/1.jmi.10.4.044008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2023] [Revised: 08/04/2023] [Accepted: 08/08/2023] [Indexed: 08/29/2023] Open
Abstract
Purpose Supervised deep convolutional neural network (CNN)-based methods have been actively used in clinical CT to reduce image noise. The networks of these methods are typically trained using paired high- and low-quality data from a massive number of patients and/or phantom images. This training process is tedious, and the network trained under a given condition may not be generalizable to patient images acquired and reconstructed under different conditions. We propose a self-trained deep CNN (ST_CNN) method for noise reduction in CT that does not rely on pre-existing training datasets. Approach The ST_CNN training was accomplished using extensive data augmentation in the projection domain, and the inference was applied to the data itself. Specifically, multiple independent noise insertions were applied to the original patient projection data to generate multiple realizations of low-quality projection data. Then, rotation augmentation was adopted for both the original and low-quality projection data by applying the rotation angle directly on the projection data so that images were rotated at arbitrary angles without introducing additional bias. A large number of paired low- and high-quality images from the same patient were reconstructed and paired for training the ST_CNN model. Results No significant difference was found between the ST_CNN and conventional CNN models in terms of the peak signal-to-noise ratio and structural similarity index measure. The ST_CNN model outperformed the conventional CNN model in terms of noise texture and homogeneity in liver parenchyma as well as better subjective visualization of liver lesions. The ST_CNN may sacrifice the sharpness of vessels slightly compared to the conventional CNN model but without affecting the visibility of peripheral vessels or diagnosis of vascular pathology. Conclusions The proposed ST_CNN method trained from the data itself may achieve similar image quality in comparison with conventional deep CNN denoising methods pre-trained on external datasets.
Collapse
Affiliation(s)
- Zhongxing Zhou
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | - Akitoshi Inoue
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| | | | - Lifeng Yu
- Mayo Clinic, Department of Radiology, Rochester, Minnesota, United States
| |
Collapse
|
34
|
Zhao F, Liu M, Gao Z, Jiang X, Wang R, Zhang L. Dual-scale similarity-guided cycle generative adversarial network for unsupervised low-dose CT denoising. Comput Biol Med 2023; 161:107029. [PMID: 37230021 DOI: 10.1016/j.compbiomed.2023.107029] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Revised: 04/10/2023] [Accepted: 05/09/2023] [Indexed: 05/27/2023]
Abstract
Removing the noise in low-dose CT (LDCT) is crucial to improving the diagnostic quality. Previously, many supervised or unsupervised deep learning-based LDCT denoising algorithms have been proposed. Unsupervised LDCT denoising algorithms are more practical than supervised ones since they do not need paired samples. However, unsupervised LDCT denoising algorithms are rarely used clinically due to their unsatisfactory denoising ability. In unsupervised LDCT denoising, the lack of paired samples makes the direction of gradient descent full of uncertainty. On the contrary, paired samples used in supervised denoising allow the parameters of networks to have a clear direction of gradient descent. To bridge the gap in performance between unsupervised and supervised LDCT denoising, we propose dual-scale similarity-guided cycle generative adversarial network (DSC-GAN). DSC-GAN uses similarity-based pseudo-pairing to better accomplish unsupervised LDCT denoising. We design a Vision Transformer-based global similarity descriptor and a residual neural network-based local similarity descriptor for DSC-GAN to effectively describe the similarity between two samples. During training, pseudo-pairs, i.e., similar LDCT samples and normal-dose CT (NDCT) samples, dominate parameter updates. Thus, the training can achieve equivalent effect as training with paired samples. Experiments on two datasets demonstrate that DSC-GAN beats the state-of-the-art unsupervised algorithms and reaches a level close to supervised LDCT denoising algorithms.
Collapse
Affiliation(s)
- Feixiang Zhao
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, Chengdu, 610000, China.
| | - Mingzhe Liu
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, Chengdu, 610000, China; School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, 325000, China.
| | - Zhihong Gao
- Department of Big Data in Health Science, First Affiliated Hospital of Wenzhou Medical University, Wenzhou, 325000, China.
| | - Xin Jiang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, 325000, China.
| | - Ruili Wang
- School of Mathematical and Computational Science, Massey University, Auckland, 0632, New Zealand.
| | - Lejun Zhang
- Cyberspace Institute of Advanced Technology, Guangzhou University, Guangzhou, 510006, China; College of Information Engineering, Yangzhou University, Yangzhou, 225127, China.
| |
Collapse
|
35
|
Liu Y, Wang C. An efficient 3D reconstruction method based on WT-TV denoising for low-dose CT images. Technol Health Care 2023; 31:463-475. [PMID: 37038798 DOI: 10.3233/thc-236040] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/12/2023]
Abstract
BACKGROUND In order to reduce the impact of CT radiation, low-dose CT is often used, but low-dose CT will bring more noise, affecting image quality and subsequent 3D reconstruction results. OBJECTIVE The study presents a reconstruction method based on wavelet transform-total variation (WT-TV) for low-dose CT. METHODS First, the low-dose CT images were denoised using WT and TV denoising methods. The WT method could preserve the features, and the TV method could preserve the edges and structures. Second, the two sets of denoised images were fused so that the features, edges, and structures could be preserved at the same time. Finally, FBP reconstruction was performed to obtain the final 3D reconstruction result. RESULTS The results show that The WT-TV method can effectively denoise low-dose CT and improve the clarity and accuracy of 3D reconstruction models. CONCLUSION Compared with other reconstruction methods, the proposed reconstruction method successfully addressed the issue of low-dose CT noising by further denoising the CT images before reconstruction. The denoising effect of low-dose CT images and the 3D reconstruction model were compared via experiments.
Collapse
|
36
|
Yang L, Li Z, Ge R, Zhao J, Si H, Zhang D. Low-Dose CT Denoising via Sinogram Inner-Structure Transformer. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:910-921. [PMID: 36331637 DOI: 10.1109/tmi.2022.3219856] [Citation(s) in RCA: 17] [Impact Index Per Article: 17.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Low-Dose Computed Tomography (LDCT) technique, which reduces the radiation harm to human bodies, is now attracting increasing interest in the medical imaging field. As the image quality is degraded by low dose radiation, LDCT exams require specialized reconstruction methods or denoising algorithms. However, most of the recent effective methods overlook the inner-structure of the original projection data (sinogram) which limits their denoising ability. The inner-structure of the sinogram represents special characteristics of the data in the sinogram domain. By maintaining this structure while denoising, the noise can be obviously restrained. Therefore, we propose an LDCT denoising network namely Sinogram Inner-Structure Transformer (SIST) to reduce the noise by utilizing the inner-structure in the sinogram domain. Specifically, we study the CT imaging mechanism and statistical characteristics of sinogram to design the sinogram inner-structure loss including the global and local inner-structure for restoring high-quality CT images. Besides, we propose a sinogram transformer module to better extract sinogram features. The transformer architecture using a self-attention mechanism can exploit interrelations between projections of different view angles, which achieves an outstanding performance in sinogram denoising. Furthermore, in order to improve the performance in the image domain, we propose the image reconstruction module to complementarily denoise both in the sinogram and image domain.
Collapse
|
37
|
Takata T, Yamada K, Yamamoto M, Kondo H. REBOA Zone Estimation from the Body Surface Using Semantic Segmentation. J Med Syst 2023; 47:42. [PMID: 36995484 DOI: 10.1007/s10916-023-01938-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Accepted: 03/06/2023] [Indexed: 03/31/2023]
Abstract
Resuscitative endovascular balloon occlusion of the aorta (REBOA) is an endovascular procedure for hemorrhage control. In REBOA, the balloon must be placed in the precise place, but it may be performed without X-ray fluoroscopy. This study aimed to estimate the REBOA zones from the body surface using deep learning for safe balloon placement. A total of 198 abdominal computed tomography (CT) datasets containing the regions of the REBOA zones were collected from open data libraries. Then, depth images of the body surface generated from the CT datasets and the images corresponding to the zones were labeled for deep learning training and validation. DeepLabV3+, a deep learning semantic segmentation model, was employed to estimate the zones. We used 176 depth images as training data and 22 images as validation data. A nine-fold cross-validation was performed to generalize the performance of the network. The median Dice coefficients for Zones 1-3 were 0.94 (inter-quarter range: 0.90-0.96), 0.77 (0.60-0.86), and 0.83 (0.74-0.89), respectively. The median displacements of the zone boundaries were 11.34 mm (5.90-19.45), 11.40 mm (4.88-20.23), and 14.17 mm (6.89-23.70) for the boundary between Zones 1 and 2, between Zones 2 and 3, and between Zone 3 and out of zone, respectively. This study examined the feasibility of REBOA zone estimation from the body surface only using deep learning-based segmentation without aortography.
Collapse
Affiliation(s)
- Takeshi Takata
- Advanced Comprehensive Research Organization, Teikyo University, Tokyo, Japan.
| | - Kentaro Yamada
- Dotter Interventional Institute, Oregon Health & Science University, Portland, OR, USA
| | - Masayoshi Yamamoto
- Department of Radiology, Teikyo University School of Medicine, Tokyo, Japan
| | - Hiroshi Kondo
- Department of Radiology, Teikyo University School of Medicine, Tokyo, Japan
| |
Collapse
|
38
|
Zhu M, Zhu Q, Song Y, Guo Y, Zeng D, Bian Z, Wang Y, Ma J. Physics-informed sinogram completion for metal artifact reduction in CT imaging. Phys Med Biol 2023; 68. [PMID: 36808913 DOI: 10.1088/1361-6560/acbddf] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2022] [Accepted: 02/21/2023] [Indexed: 02/23/2023]
Abstract
Objective.Metal artifacts in the computed tomography (CT) imaging are unavoidably adverse to the clinical diagnosis and treatment outcomes. Most metal artifact reduction (MAR) methods easily result in the over-smoothing problem and loss of structure details near the metal implants, especially for these metal implants with irregular elongated shapes. To address this problem, we present the physics-informed sinogram completion (PISC) method for MAR in CT imaging, to reduce metal artifacts and recover more structural textures.Approach.Specifically, the original uncorrected sinogram is firstly completed by the normalized linear interpolation algorithm to reduce metal artifacts. Simultaneously, the uncorrected sinogram is also corrected based on the beam-hardening correction physical model, to recover the latent structure information in metal trajectory region by leveraging the attenuation characteristics of different materials. Both corrected sinograms are fused with the pixel-wise adaptive weights, which are manually designed according to the shape and material information of metal implants. To furtherly reduce artifacts and improve the CT image quality, a post-processing frequency split algorithm is adopted to yield the final corrected CT image after reconstructing the fused sinogram.Main results.We qualitatively and quantitatively evaluated the presented PISC method on two simulated datasets and three real datasets. All results demonstrate that the presented PISC method can effectively correct the metal implants with various shapes and materials, in terms of artifact suppression and structure preservation.Significance.We proposed a sinogram-domain MAR method to compensate for the over-smoothing problem existing in most MAR methods by taking advantage of the physical prior knowledge, which has the potential to improve the performance of the deep learning based MAR approaches.
Collapse
Affiliation(s)
- Manman Zhu
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Qisen Zhu
- Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Yuyan Song
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Yi Guo
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Dong Zeng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Zhaoying Bian
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Yongbo Wang
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| | - Jianhua Ma
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China.,Pazhou Lab (Huangpu), Guangzhou 510700, People's Republic of China
| |
Collapse
|
39
|
Lu Z, Xia W, Huang Y, Hou M, Chen H, Zhou J, Shan H, Zhang Y. M 3NAS: Multi-Scale and Multi-Level Memory-Efficient Neural Architecture Search for Low-Dose CT Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:850-863. [PMID: 36327187 DOI: 10.1109/tmi.2022.3219286] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Lowering the radiation dose in computed tomography (CT) can greatly reduce the potential risk to public health. However, the reconstructed images from dose-reduced CT or low-dose CT (LDCT) suffer from severe noise which compromises the subsequent diagnosis and analysis. Recently, convolutional neural networks have achieved promising results in removing noise from LDCT images. The network architectures that are used are either handcrafted or built on top of conventional networks such as ResNet and U-Net. Recent advances in neural network architecture search (NAS) have shown that the network architecture has a dramatic effect on the model performance. This indicates that current network architectures for LDCT may be suboptimal. Therefore, in this paper, we make the first attempt to apply NAS to LDCT and propose a multi-scale and multi-level memory-efficient NAS for LDCT denoising, termed M3NAS. On the one hand, the proposed M3NAS fuses features extracted by different scale cells to capture multi-scale image structural details. On the other hand, the proposed M3NAS can search a hybrid cell- and network-level structure for better performance. In addition, M3NAS can effectively reduce the number of model parameters and increase the speed of inference. Extensive experimental results on two different datasets demonstrate that the proposed M3NAS can achieve better performance and fewer parameters than several state-of-the-art methods. In addition, we also validate the effectiveness of the multi-scale and multi-level architecture for LDCT denoising, and present further analysis for different configurations of super-net.
Collapse
|
40
|
Fu Y, Dong S, Niu M, Xue L, Guo H, Huang Y, Xu Y, Yu T, Shi K, Yang Q, Shi Y, Zhang H, Tian M, Zhuo C. AIGAN: Attention-encoding Integrated Generative Adversarial Network for the reconstruction of low-dose CT and low-dose PET images. Med Image Anal 2023; 86:102787. [PMID: 36933386 DOI: 10.1016/j.media.2023.102787] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2022] [Revised: 11/05/2022] [Accepted: 02/22/2023] [Indexed: 03/04/2023]
Abstract
X-ray computed tomography (CT) and positron emission tomography (PET) are two of the most commonly used medical imaging technologies for the evaluation of many diseases. Full-dose imaging for CT and PET ensures the image quality but usually raises concerns about the potential health risks of radiation exposure. The contradiction between reducing the radiation exposure and remaining diagnostic performance can be addressed effectively by reconstructing the low-dose CT (L-CT) and low-dose PET (L-PET) images to the same high-quality ones as full-dose (F-CT and F-PET). In this paper, we propose an Attention-encoding Integrated Generative Adversarial Network (AIGAN) to achieve efficient and universal full-dose reconstruction for L-CT and L-PET images. AIGAN consists of three modules: the cascade generator, the dual-scale discriminator and the multi-scale spatial fusion module (MSFM). A sequence of consecutive L-CT (L-PET) slices is first fed into the cascade generator that integrates with a generation-encoding-generation pipeline. The generator plays the zero-sum game with the dual-scale discriminator for two stages: the coarse and fine stages. In both stages, the generator generates the estimated F-CT (F-PET) images as like the original F-CT (F-PET) images as possible. After the fine stage, the estimated fine full-dose images are then fed into the MSFM, which fully explores the inter- and intra-slice structural information, to output the final generated full-dose images. Experimental results show that the proposed AIGAN achieves the state-of-the-art performances on commonly used metrics and satisfies the reconstruction needs for clinical standards.
Collapse
Affiliation(s)
- Yu Fu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China; Binjiang Institute, Zhejiang University, Hangzhou, China
| | - Shunjie Dong
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Meng Niu
- Department of Radiology, The First Hospital of Lanzhou University, Lanzhou, China
| | - Le Xue
- Department of Nuclear Medicine and Medical PET Center The Second Hospital of Zhejiang University School of Medicine, Hangzhou, China
| | - Hanning Guo
- Institute of Neuroscience and Medicine, Medical Imaging Physics (INM-4), Forschungszentrum Jülich, Jülich, Germany
| | - Yanyan Huang
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Yuanfan Xu
- Hangzhou Universal Medical Imaging Diagnostic Center, Hangzhou, China
| | - Tianbai Yu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Kuangyu Shi
- Department of Nuclear Medicine, University Hospital Bern, Bern, Switzerland
| | - Qianqian Yang
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Yiyu Shi
- Department of Computer Science and Engineering, University of Notre Dame, Notre Dame, IN, USA
| | - Hong Zhang
- Binjiang Institute, Zhejiang University, Hangzhou, China; Department of Nuclear Medicine and Medical PET Center The Second Hospital of Zhejiang University School of Medicine, Hangzhou, China
| | - Mei Tian
- Human Phenome Institute, Fudan University, Shanghai, China.
| | - Cheng Zhuo
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China; Key Laboratory of Collaborative Sensing and Autonomous Unmanned Systems of Zhejiang Province, Hangzhou, China.
| |
Collapse
|
41
|
Zhu L, Han Y, Xi X, Fu H, Tan S, Liu M, Yang S, Liu C, Li L, Yan B. STEDNet: Swin transformer-based encoder-decoder network for noise reduction in low-dose CT. Med Phys 2023. [PMID: 36708286 DOI: 10.1002/mp.16249] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2022] [Revised: 11/04/2022] [Accepted: 01/17/2023] [Indexed: 01/29/2023] Open
Abstract
BACKGROUND Low-dose computed tomography (LDCT) can reduce the dose of X-ray radiation, making it increasingly significant for routine clinical diagnosis and treatment planning. However, the noise introduced by low-dose X-ray exposure degrades the quality of CT images, affecting the accuracy of clinical diagnosis. Purpose The noises, artifacts, and high-frequency components are similarly distributed in LDCT images. Transformer can capture global context information in an attentional manner to create distant dependencies on targets and extract more powerful features. In this paper, we reduce the impact of image errors on the ability to retain detailed information and improve the noise suppression performance by fully mining the distribution characteristics of image information. METHODS This paper proposed an LDCT noise and artifact suppressing network based on Swin Transformer. The network includes a noise extraction sub-network and a noise removal sub-network. The noise extraction and removal capability are improved using a coarse extraction network of high-frequency features based on full convolution. The noise removal sub-network improves the network's ability to extract relevant image features by using a Swin Transformer with a shift window as an encoder-decoder and skip connections for global feature fusion. Also, the perceptual field is extended by extracting multi-scale features of the images to recover the spatial resolution of the feature maps. The network uses a loss constraint with a combination of L1 and MS-SSIM to improve and ensure the stability and denoising effect of the network. RESULTS The denoising ability and clinical applicability of the methods were tested using clinical datasets. Compared with DnCNN, RED-CNN, CBDNet and TSCN, the STEDNet method shows a better denoising effect on RMSE and PSNR. The STEDNet method effectively removes image noise and preserves the image structure to the maximum extent, making the reconstructed image closest to the NDCT image. The subjective and objective analysis of several sets of experiments shows that the method in this paper can effectively maintain the structure, edges, and textures of the denoised images while having good noise suppression performance. In the real data evaluation, the RMSE of this method is reduced by 18.82%, 15.15%, 2.25%, and 1.10% on average compared with DnCNN, RED-CNN, CBDNet, and TSCNN, respectively. The average improvement of PSNR is 9.53%, 7.33%, 2.65%, and 3.69%, respectively. CONCLUSIONS This paper proposed a LDCT image denoising algorithm based on end-to-end training. The method in this paper can effectively improve the diagnostic performance of CT images by constraining the details of the images and restoring the LDCT image structure. The problem of increased noise and artifacts in CT images can be solved while maintaining the integrity of CT image tissue structure and pathological information. Compared with other algorithms, this method has better denoising effects both quantitatively and qualitatively.
Collapse
Affiliation(s)
- Linlin Zhu
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Yu Han
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Xiaoqi Xi
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Huijuan Fu
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Siyu Tan
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Mengnan Liu
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Shuangzhan Yang
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Chang Liu
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China.,School of Cyber Science and Engineering, Zhengzhou University, Zhengzhou, China
| | - Lei Li
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| | - Bin Yan
- Henan Key Laboratory of Imaging and Intelligent Processing, PLA Strategic Support Force Information Engineering University, Zhengzhou, China
| |
Collapse
|
42
|
Yang S, Pu Q, Lei C, Zhang Q, Jeon S, Yang X. Low-dose CT denoising with a high-level feature refinement and dynamic convolution network. Med Phys 2022. [PMID: 36542402 DOI: 10.1002/mp.16175] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2022] [Revised: 10/31/2022] [Accepted: 12/08/2022] [Indexed: 12/24/2022] Open
Abstract
BACKGROUND Since the potential health risks of the radiation generated by computer tomography (CT), concerns have been expressed on reducing the radiation dose. However, low-dose CT (LDCT) images contain complex noise and artifacts, bringing uncertainty to medical diagnosis. PURPOSE Existing deep learning (DL)-based denoising methods are difficult to fully exploit hierarchical features of different levels, limiting the effect of denoising. Moreover, the standard convolution kernel is parameter sharing and cannot be adjusted dynamically with input change. This paper proposes an LDCT denoising network using high-level feature refinement and multiscale dynamic convolution to mitigate these problems. METHODS The dual network structure proposed in this paper consists of the feature refinement network (FRN) and the dynamic perception network (DPN). The FDN extracts features of different levels through residual dense connections. The high-level hierarchical information is transmitted to DPN to improve the low-level representations. In DPN, the two networks' features are fused by local channel attention (LCA) to assign weights in different regions and handle CT images' delicate tissues better. Then, the dynamic dilated convolution (DDC) with multibranch and multiscale receptive fields is proposed to enhance the expression and processing ability of the denoising network. The experiments were trained and tested on the dataset "NIH-AAPM-Mayo Clinic Low-Dose CT Grand Challenge," consisting of 10 anonymous patients with normal-dose abdominal CT and LDCT at 25% dose. In addition, external validation was performed on the dataset "Low Dose CT Image and Projection Data," which included 300 chest CT images at 10% dose and 300 head CT images at 25% dose. RESULTS Proposed method compared with seven mainstream LDCT denoising algorithms. On the Mayo dataset, achieved peak signal-to-noise ratio (PSNR): 46.3526 dB (95% CI: 46.0121-46.6931 dB) and structural similarity (SSIM): 0.9844 (95% CI: 0.9834-0.9854). Compared with LDCT, the average increase was 3.4159 dB and 0.0239, respectively. The results are relatively optimal and statistically significant compared with other methods. In external verification, our algorithm can cope well with ultra-low-dose chest CT images at 10% dose and obtain PSNR: 28.6130 (95% CI: 28.1680-29.0580 dB) and SSIM: 0.7201 (95% CI: 0.7101-0.7301). Compared with LDCT, PSNR/SSIM is increased by 3.6536dB and 0.2132, respectively. In addition, the quality of LDCT can also be improved in head CT denoising. CONCLUSIONS This paper proposes a DL-based LDCT denoising algorithm, which utilizes high-level features and multiscale dynamic convolution to optimize the network's denoising effect. This method can realize speedy denoising and performs well in noise suppression and detail preservation, which can be helpful for the diagnosis of LDCT.
Collapse
Affiliation(s)
- Sihan Yang
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, China.,School of Aeronautics and Astronautics, Sichuan University, Chengdu, Sichuan, China
| | - Qiang Pu
- Department of Thoracic Surgery, West China Hospital, Sichuan University, Chengdu, Sichuan, China
| | - Chunting Lei
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, China.,School of Aeronautics and Astronautics, Sichuan University, Chengdu, Sichuan, China
| | - Qiao Zhang
- Macro Net Communication Co., Ltd., Chongqing, China
| | - Seunggil Jeon
- Samsung Electronics, Suwon-si, Gyeonggi-do, Republic of Korea
| | - Xiaomin Yang
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, China.,School of Aeronautics and Astronautics, Sichuan University, Chengdu, Sichuan, China
| |
Collapse
|
43
|
Zhu M, Mao Z, Li D, Wang Y, Zeng D, Bian Z, Ma J. Structure-preserved meta-learning uniting network for improving low-dose CT quality. Phys Med Biol 2022; 67. [PMID: 36351294 DOI: 10.1088/1361-6560/aca194] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2022] [Accepted: 11/09/2022] [Indexed: 11/10/2022]
Abstract
Objective.Deep neural network (DNN) based methods have shown promising performances for low-dose computed tomography (LDCT) imaging. However, most of the DNN-based methods are trained on simulated labeled datasets, and the low-dose simulation algorithms are usually designed based on simple statistical models which deviate from the real clinical scenarios, which could lead to issues of overfitting, instability and poor robustness. To address these issues, in this work, we present a structure-preserved meta-learning uniting network (shorten as 'SMU-Net') to suppress noise-induced artifacts and preserve structure details in the unlabeled LDCT imaging task in real scenarios.Approach.Specifically, the presented SMU-Net contains two networks, i.e., teacher network and student network. The teacher network is trained on simulated labeled dataset and then helps the student network train with the unlabeled LDCT images via the meta-learning strategy. The student network is trained on real LDCT dataset with the pseudo-labels generated by the teacher network. Moreover, the student network adopts the Co-teaching strategy to improve the robustness of the presented SMU-Net.Main results.We validate the proposed SMU-Net method on three public datasets and one real low-dose dataset. The visual image results indicate that the proposed SMU-Net has superior performance on reducing noise-induced artifacts and preserving structure details. And the quantitative results exhibit that the presented SMU-Net method generally obtains the highest signal-to-noise ratio (PSNR), the highest structural similarity index measurement (SSIM), and the lowest root-mean-square error (RMSE) values or the lowest natural image quality evaluator (NIQE) scores.Significance.We propose a meta learning strategy to obtain high-quality CT images in the LDCT imaging task, which is designed to take advantage of unlabeled CT images to promote the reconstruction performance in the LDCT environments.
Collapse
Affiliation(s)
- Manman Zhu
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Zerui Mao
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Danyang Li
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Yongbo Wang
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Dong Zeng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Zhaoying Bian
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| | - Jianhua Ma
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, People's Republic of China
| |
Collapse
|
44
|
Kim B, Kim J, Ye JC. Task-Agnostic Vision Transformer for Distributed Learning of Image Processing. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; PP:203-218. [PMID: 37015481 DOI: 10.1109/tip.2022.3226892] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Recently, distributed learning approaches have been studied for using data from multiple sources without sharing them, but they are not usually suitable in applications where each client carries out different tasks. Meanwhile, Transformer has been widely explored in computer vision area due to its capability to learn the common representation through global attention. By leveraging the advantages of Transformer, here we present a new distributed learning framework for multiple image processing tasks, allowing clients to learn distinct tasks with their local data. This arises from a disentangled representation of local and non-local features using a task-specific head/tail and a task-agnostic Vision Transformer. Each client learns a translation from its own task to a common representation using the task-specific networks, while the Transformer body on the server learns global attention between the features embedded in the representation. To enable decomposition between the task-specific and common representations, we propose an alternating training strategy between clients and server. Experimental results on distributed learning for various tasks show that our method synergistically improves the performance of each client with its own data.
Collapse
|
45
|
Wagner F, Thies M, Denzinger F, Gu M, Patwari M, Ploner S, Maul N, Pfaff L, Huang Y, Maier A. Trainable joint bilateral filters for enhanced prediction stability in low-dose CT. Sci Rep 2022; 12:17540. [PMID: 36266416 PMCID: PMC9585057 DOI: 10.1038/s41598-022-22530-4] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2022] [Accepted: 10/17/2022] [Indexed: 01/13/2023] Open
Abstract
Low-dose computed tomography (CT) denoising algorithms aim to enable reduced patient dose in routine CT acquisitions while maintaining high image quality. Recently, deep learning (DL)-based methods were introduced, outperforming conventional denoising algorithms on this task due to their high model capacity. However, for the transition of DL-based denoising to clinical practice, these data-driven approaches must generalize robustly beyond the seen training data. We, therefore, propose a hybrid denoising approach consisting of a set of trainable joint bilateral filters (JBFs) combined with a convolutional DL-based denoising network to predict the guidance image. Our proposed denoising pipeline combines the high model capacity enabled by DL-based feature extraction with the reliability of the conventional JBF. The pipeline's ability to generalize is demonstrated by training on abdomen CT scans without metal implants and testing on abdomen scans with metal implants as well as on head CT data. When embedding RED-CNN/QAE, two well-established DL-based denoisers in our pipeline, the denoising performance is improved by 10%/82% (RMSE) and 3%/81% (PSNR) in regions containing metal and by 6%/78% (RMSE) and 2%/4% (PSNR) on head CT data, compared to the respective vanilla model. Concluding, the proposed trainable JBFs limit the error bound of deep neural networks to facilitate the applicability of DL-based denoisers in low-dose CT pipelines.
Collapse
Affiliation(s)
- Fabian Wagner
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Mareike Thies
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Felix Denzinger
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Mingxuan Gu
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Mayank Patwari
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Stefan Ploner
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Noah Maul
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Laura Pfaff
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Yixing Huang
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| | - Andreas Maier
- grid.5330.50000 0001 2107 3311Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, 91058 Erlangen, Germany
| |
Collapse
|
46
|
Qi Y, He P, Zhu J, Wang Y, Zhao H, Chen J. Application of Low-Dose CT and MRI in the Evaluation of Soft Tissue Injury in Tibial Plateau Fractures. SCANNING 2022; 2022:7686485. [PMID: 36189142 PMCID: PMC9507771 DOI: 10.1155/2022/7686485] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Revised: 08/31/2022] [Accepted: 09/05/2022] [Indexed: 06/16/2023]
Abstract
Objective To explore the application value of low-dose CT and MRI in the evaluation of soft tissue injury in tibial plateau fractures. Methods This study included 89 patients with high suspicion of TPF and KI admitted to our hospital from July 2015 to May 2021. After arthroscopy, 81 patients were diagnosed with FTP combined with KI. The Schatzker classification based on X-ray and CT plain scan combined with three-dimensional reconstruction was recorded, and the soft tissue injury was recorded according to the MRI examination of the affected knee joint. Results With the results of pathological examination and arthroscopic surgery as the gold standard, the results of MRI and pathological examination and arthroscopic examination were in good agreement (Kappa = 0.857, 0.844), and CT was moderately in agreement (Kappa = 0.697, 0.694). In KI examination, CT and MRI had no difference in the evaluation of ligament injury and bone injury (P > 0.05), but MRI had better diagnostic effect on meniscus injury (P < 0.05). Finally, the satisfaction survey showed that patients in the CT group were more satisfied with clinical services (P < 0.05). Conclusion Both CT and MRI have certain diagnostic value for occult tibial plateau fractures, among which CT examination is more advantageous for trabecular bone fractures, MRI examination is more advantageous for cortical bone fractures, and MRI examination can improve occult tibial plateau fracture inspection accuracy.
Collapse
Affiliation(s)
- Yinping Qi
- Department of Radiology, The Second Hospital of Yinzhou, Ningbo, Zhejiang 315100, China
| | - Peipei He
- Department of Radiology, The Second Hospital of Yinzhou, Ningbo, Zhejiang 315100, China
| | - Jianping Zhu
- Department of Radiology, The Second Hospital of Yinzhou, Ningbo, Zhejiang 315100, China
| | - Yanan Wang
- Department of Radiology, The Second Hospital of Yinzhou, Ningbo, Zhejiang 315100, China
| | - Hong Zhao
- Department of Radiology, The Second Hospital of Yinzhou, Ningbo, Zhejiang 315100, China
| | - Junbo Chen
- Department of Radiology, The Second Hospital of Yinzhou, Ningbo, Zhejiang 315100, China
| |
Collapse
|
47
|
Kandarpa VSS, Perelli A, Bousse A, Visvikis D. LRR-CED: low-resolution reconstruction-aware convolutional encoder–decoder network for direct sparse-view CT image reconstruction. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac7bce] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Accepted: 06/23/2022] [Indexed: 11/12/2022]
Abstract
Abstract
Objective. Sparse-view computed tomography (CT) reconstruction has been at the forefront of research in medical imaging. Reducing the total x-ray radiation dose to the patient while preserving the reconstruction accuracy is a big challenge. The sparse-view approach is based on reducing the number of rotation angles, which leads to poor quality reconstructed images as it introduces several artifacts. These artifacts are more clearly visible in traditional reconstruction methods like the filtered-backprojection (FBP) algorithm. Approach. Over the years, several model-based iterative and more recently deep learning-based methods have been proposed to improve sparse-view CT reconstruction. Many deep learning-based methods improve FBP-reconstructed images as a post-processing step. In this work, we propose a direct deep learning-based reconstruction that exploits the information from low-dimensional scout images, to learn the projection-to-image mapping. This is done by concatenating FBP scout images at multiple resolutions in the decoder part of a convolutional encoder–decoder (CED). Main results. This approach is investigated on two different networks, based on Dense Blocks and U-Net to show that a direct mapping can be learned from a sinogram to an image. The results are compared to two post-processing deep learning methods (FBP-ConvNet and DD-Net) and an iterative method that uses a total variation (TV) regularization. Significance. This work presents a novel method that uses information from both sinogram and low-resolution scout images for sparse-view CT image reconstruction. We also generalize this idea by demonstrating results with two different neural networks. This work is in the direction of exploring deep learning across the various stages of the image reconstruction pipeline involving data correction, domain transfer and image improvement.
Collapse
|
48
|
Zavala-Mondragon LA, Rongen P, Bescos JO, de With PHN, van der Sommen F. Noise Reduction in CT Using Learned Wavelet-Frame Shrinkage Networks. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:2048-2066. [PMID: 35201984 DOI: 10.1109/tmi.2022.3154011] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Encoding-decoding (ED) CNNs have demonstrated state-of-the-art performance for noise reduction over the past years. This has triggered the pursuit of better understanding the inner workings of such architectures, which has led to the theory of deep convolutional framelets (TDCF), revealing important links between signal processing and CNNs. Specifically, the TDCF demonstrates that ReLU CNNs induce low-rankness, since these models often do not satisfy the necessary redundancy to achieve perfect reconstruction (PR). In contrast, this paper explores CNNs that do meet the PR conditions. We demonstrate that in these type of CNNs soft shrinkage and PR can be assumed. Furthermore, based on our explorations we propose the learned wavelet-frame shrinkage network, or LWFSN and its residual counterpart, the rLWFSN. The ED path of the (r)LWFSN complies with the PR conditions, while the shrinkage stage is based on the linear expansion of thresholds proposed Blu and Luisier. In addition, the LWFSN has only a fraction of the training parameters (<1%) of conventional CNNs, very small inference times, low memory footprint, while still achieving performance close to state-of-the-art alternatives, such as the tight frame (TF) U-Net and FBPConvNet, in low-dose CT denoising.
Collapse
|
49
|
Limited-Angle CT Reconstruction with Generative Adversarial Network Sinogram Inpainting and Unsupervised Artifact Removal. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12126268] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/10/2022]
Abstract
High-quality limited-angle computed tomography (CT) reconstruction is in high demand in the medical field. Being unlimited by the pairing of sinogram and the reconstructed image, unsupervised methods have attracted wide attention from researchers. The reconstruction limit of the existing unsupervised reconstruction methods, however, is to use [0°, 120°] of projection data, and the quality of the reconstruction still has room for improvement. In this paper, we propose a limited-angle CT reconstruction generative adversarial network based on sinogram inpainting and unsupervised artifact removal to further reduce the angle range limit and to improve the image quality. We collected a large number of CT lung and head images and Radon transformed them into missing sinograms. Sinogram inpainting network is developed to complete missing sinograms, based on which the filtered back projection algorithm can output images with most artifacts removed; then, these images are mapped to artifact-free images by using artifact removal network. Finally, we generated reconstruction results sized 512×512 that are comparable to full-scan reconstruction using only [0°, 90°] of limited sinogram projection data. Compared with the current unsupervised methods, the proposed method can reconstruct images of higher quality.
Collapse
|
50
|
Inkinen SI, Mäkelä T, Kaasalainen T, Peltonen J, Kangasniemi M, Kortesniemi M. Automatic head computed tomography image noise quantification with deep learning. Phys Med 2022; 99:102-112. [PMID: 35671678 DOI: 10.1016/j.ejmp.2022.05.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/20/2022] [Revised: 04/02/2022] [Accepted: 05/25/2022] [Indexed: 10/18/2022] Open
Abstract
PURPOSE Computed tomography (CT) image noise is usually determined by standard deviation (SD) of pixel values from uniform image regions. This study investigates how deep learning (DL) could be applied in head CT image noise estimation. METHODS Two approaches were investigated for noise image estimation of a single acquisition image: direct noise image estimation using supervised DnCNN convolutional neural network (CNN) architecture, and subtraction of a denoised image estimated with denoising UNet-CNN experimented with supervised and unsupervised noise2noise training approaches. Noise was assessed with local SD maps using 3D- and 2D-CNN architectures. Anthropomorphic phantom CT image dataset (N = 9 scans, 3 repetitions) was used for DL-model comparisons. Mean square error (MSE) and mean absolute percentage errors (MAPE) of SD values were determined using the SD values of subtraction images as ground truth. Open-source clinical head CT low-dose dataset (Ntrain = 37, Ntest = 10 subjects) were used to demonstrate DL applicability in noise estimation from manually labeled uniform regions and in automated noise and contrast assessment. RESULTS The direct SD estimation using 3D-CNN was the most accurate assessment method when comparing in phantom dataset (MAPE = 15.5%, MSE = 6.3HU). Unsupervised noise2noise approach provided only slightly inferior results (MAPE = 20.2%, MSE = 13.7HU). 2DCNN and unsupervised UNet models provided the smallest MSE on clinical labeled uniform regions. CONCLUSIONS DL-based clinical image assessment is feasible and provides acceptable accuracy as compared to true image noise. Noise2noise approach may be feasible in clinical use where no ground truth data is available. Noise estimation combined with tissue segmentation may enable more comprehensive image quality characterization.
Collapse
Affiliation(s)
- Satu I Inkinen
- HUS Diagnostic Center, Radiology, Helsinki University and Helsinki University Hospital, Haartmaninkatu 4, 00290 Helsinki, Finland.
| | - Teemu Mäkelä
- HUS Diagnostic Center, Radiology, Helsinki University and Helsinki University Hospital, Haartmaninkatu 4, 00290 Helsinki, Finland; Department of Physics, University of Helsinki, P.O. Box 64, FI-00014 Helsinki, Finland
| | - Touko Kaasalainen
- HUS Diagnostic Center, Radiology, Helsinki University and Helsinki University Hospital, Haartmaninkatu 4, 00290 Helsinki, Finland
| | - Juha Peltonen
- HUS Diagnostic Center, Radiology, Helsinki University and Helsinki University Hospital, Haartmaninkatu 4, 00290 Helsinki, Finland
| | - Marko Kangasniemi
- HUS Diagnostic Center, Radiology, Helsinki University and Helsinki University Hospital, Haartmaninkatu 4, 00290 Helsinki, Finland
| | - Mika Kortesniemi
- HUS Diagnostic Center, Radiology, Helsinki University and Helsinki University Hospital, Haartmaninkatu 4, 00290 Helsinki, Finland
| |
Collapse
|