1
|
Bajaj S, Bala M, Angurala M. A comparative analysis of different augmentations for brain images. Med Biol Eng Comput 2024; 62:3123-3150. [PMID: 38782880 DOI: 10.1007/s11517-024-03127-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 05/10/2024] [Indexed: 05/25/2024]
Abstract
Deep learning (DL) requires a large amount of training data to improve performance and prevent overfitting. To overcome these difficulties, we need to increase the size of the training dataset. This can be done by augmentation on a small dataset. The augmentation approaches must enhance the model's performance during the learning period. There are several types of transformations that can be applied to medical images. These transformations can be applied to the entire dataset or to a subset of the data, depending on the desired outcome. In this study, we categorize data augmentation methods into four groups: Absent augmentation, where no modifications are made; basic augmentation, which includes brightness and contrast adjustments; intermediate augmentation, encompassing a wider array of transformations like rotation, flipping, and shifting in addition to brightness and contrast adjustments; and advanced augmentation, where all transformation layers are employed. We plan to conduct a comprehensive analysis to determine which group performs best when applied to brain CT images. This evaluation aims to identify the augmentation group that produces the most favorable results in terms of improving model accuracy, minimizing diagnostic errors, and ensuring the robustness of the model in the context of brain CT image analysis.
Collapse
Affiliation(s)
- Shilpa Bajaj
- Applied Sciences (Computer Applications), I.K. Gujral Punjab Technical University, Jalandhar, Kapurthala, India.
| | - Manju Bala
- Department of Computer Science and Engineering, Khalsa College of Engineering and Technology, Amritsar, India
| | - Mohit Angurala
- Apex Institute of Technology (CSE), Chandigarh University, Gharuan, Mohali, Punjab, India
| |
Collapse
|
2
|
Gao C, Wang H. Intelligent Stroke Disease Prediction Model Using Deep Learning Approaches. Stroke Res Treat 2024; 2024:4523388. [PMID: 38817540 PMCID: PMC11139533 DOI: 10.1155/2024/4523388] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2024] [Revised: 04/24/2024] [Accepted: 04/27/2024] [Indexed: 06/01/2024] Open
Abstract
Stroke is a high morbidity and mortality disease that poses a serious threat to people's health. Early recognition of the various warning signs of stroke is necessary so that timely clinical intervention can help reduce the severity of stroke. Deep neural networks have powerful feature representation capabilities and can automatically learn discriminant features from large amounts of data. This paper uses a range of physiological characteristic parameters and collaborates with deep neural networks, such as the Wasserstein generative adversarial networks with gradient penalty and regression network, to construct a stroke prediction model. Firstly, to address the problem of imbalance between positive and negative samples in the stroke public data set, we performed positive sample data augmentation and utilized WGAN-GP to generate stroke data with high fidelity and used it for the training of the prediction network model. Then, the relationship between observable physiological characteristic parameters and the predicted risk of suffering a stroke was modeled as a nonlinear mapping transformation, and a stroke prediction model based on a deep regression network was designed. Finally, the proposed method is compared with commonly used machine learning-based classification algorithms such as decision tree, random forest, support vector machine, and artificial neural networks. The prediction results of the proposed method are optimal in the comprehensive measurement index F. Further ablation experiments also show that the designed prediction model has certain robustness and can effectively predict stroke diseases.
Collapse
Affiliation(s)
- Chunhua Gao
- School of Tourism and Physical Health, Hezhou University, Hezhou 542899, China
| | - Hui Wang
- School of Artificial Intelligence, Hezhou University, Hezhou 542899, China
| |
Collapse
|
3
|
Roy R, Mazumdar S, Chowdhury AS. ADGAN: Attribute-Driven Generative Adversarial Network for Synthesis and Multiclass Classification of Pulmonary Nodules. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:2484-2495. [PMID: 35853058 DOI: 10.1109/tnnls.2022.3190331] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Lung cancer is the leading cause of cancer-related deaths worldwide. According to the American Cancer Society, early diagnosis of pulmonary nodules in computed tomography (CT) scans can improve the five-year survival rate up to 70% with proper treatment planning. In this article, we propose an attribute-driven Generative Adversarial Network (ADGAN) for synthesis and multiclass classification of Pulmonary Nodules. A self-attention U-Net (SaUN) architecture is proposed to improve the generation mechanism of the network. The generator is designed with two modules, namely, self-attention attribute module (SaAM) and a self-attention spatial module (SaSM). SaAM generates a nodule image based on given attributes whereas SaSM specifies the nodule region of the input image to be altered. A reconstruction loss along with an attention localization loss (AL) is used to produce an attention map prioritizing the nodule regions. To avoid resemblance between a generated image and a real image, we further introduce an adversarial loss containing a regularization term based on KL divergence. The discriminator part of the proposed model is designed to achieve the multiclass nodule classification task. Our proposed approach is validated over two challenging publicly available datasets, namely LIDC-IDRI and LUNGX. Exhaustive experimentation on these two datasets clearly indicate that we have achieved promising classification accuracy as compared to other state-of-the-art methods.
Collapse
|
4
|
Chen X, Liu X, Wu Y, Wang Z, Wang SH. Research related to the diagnosis of prostate cancer based on machine learning medical images: A review. Int J Med Inform 2024; 181:105279. [PMID: 37977054 DOI: 10.1016/j.ijmedinf.2023.105279] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2023] [Revised: 09/06/2023] [Accepted: 10/29/2023] [Indexed: 11/19/2023]
Abstract
BACKGROUND Prostate cancer is currently the second most prevalent cancer among men. Accurate diagnosis of prostate cancer can provide effective treatment for patients and greatly reduce mortality. The current medical imaging tools for screening prostate cancer are mainly MRI, CT and ultrasound. In the past 20 years, these medical imaging methods have made great progress with machine learning, especially the rise of deep learning has led to a wider application of artificial intelligence in the use of image-assisted diagnosis of prostate cancer. METHOD This review collected medical image processing methods, prostate and prostate cancer on MR images, CT images, and ultrasound images through search engines such as web of science, PubMed, and Google Scholar, including image pre-processing methods, segmentation of prostate gland on medical images, registration between prostate gland on different modal images, detection of prostate cancer lesions on the prostate. CONCLUSION Through these collated papers, it is found that the current research on the diagnosis and staging of prostate cancer using machine learning and deep learning is in its infancy, and most of the existing studies are on the diagnosis of prostate cancer and classification of lesions, and the accuracy is low, with the best results having an accuracy of less than 0.95. There are fewer studies on staging. The research is mainly focused on MR images and much less on CT images, ultrasound images. DISCUSSION Machine learning and deep learning combined with medical imaging have a broad application prospect for the diagnosis and staging of prostate cancer, but the research in this area still has more room for development.
Collapse
Affiliation(s)
- Xinyi Chen
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China.
| | - Xiang Liu
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China.
| | - Yuke Wu
- School of Electronic and Electrical Engineering, Shanghai University of Engineering Science, Shanghai 201620, China.
| | - Zhenglei Wang
- Department of Medical Imaging, Shanghai Electric Power Hospital, Shanghai 201620, China.
| | - Shuo Hong Wang
- Department of Molecular and Cellular Biology and Center for Brain Science, Harvard University, Cambridge, MA 02138, USA.
| |
Collapse
|
5
|
Goceri E. Medical image data augmentation: techniques, comparisons and interpretations. Artif Intell Rev 2023; 56:1-45. [PMID: 37362888 PMCID: PMC10027281 DOI: 10.1007/s10462-023-10453-z] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/27/2023] [Indexed: 03/29/2023]
Abstract
Designing deep learning based methods with medical images has always been an attractive area of research to assist clinicians in rapid examination and accurate diagnosis. Those methods need a large number of datasets including all variations in their training stages. On the other hand, medical images are always scarce due to several reasons, such as not enough patients for some diseases, patients do not want to allow their images to be used, lack of medical equipment or equipment, inability to obtain images that meet the desired criteria. This issue leads to bias in datasets, overfitting, and inaccurate results. Data augmentation is a common solution to overcome this issue and various augmentation techniques have been applied to different types of images in the literature. However, it is not clear which data augmentation technique provides more efficient results for which image type since different diseases are handled, different network architectures are used, and these architectures are trained and tested with different numbers of data sets in the literature. Therefore, in this work, the augmentation techniques used to improve performances of deep learning based diagnosis of the diseases in different organs (brain, lung, breast, and eye) from different imaging modalities (MR, CT, mammography, and fundoscopy) have been examined. Also, the most commonly used augmentation methods have been implemented, and their effectiveness in classifications with a deep network has been discussed based on quantitative performance evaluations. Experiments indicated that augmentation techniques should be chosen carefully according to image types.
Collapse
Affiliation(s)
- Evgin Goceri
- Department of Biomedical Engineering, Engineering Faculty, Akdeniz University, Antalya, Turkey
| |
Collapse
|
6
|
Li X, Zhang R, Wang Q, Duan X, Sun Y, Wang J. SAR-CGAN: Improved generative adversarial network for EIT reconstruction of lung diseases. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022]
|
7
|
Osuala R, Kushibar K, Garrucho L, Linardos A, Szafranowska Z, Klein S, Glocker B, Diaz O, Lekadir K. Data synthesis and adversarial networks: A review and meta-analysis in cancer imaging. Med Image Anal 2023; 84:102704. [PMID: 36473414 DOI: 10.1016/j.media.2022.102704] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2021] [Revised: 11/02/2022] [Accepted: 11/21/2022] [Indexed: 11/26/2022]
Abstract
Despite technological and medical advances, the detection, interpretation, and treatment of cancer based on imaging data continue to pose significant challenges. These include inter-observer variability, class imbalance, dataset shifts, inter- and intra-tumour heterogeneity, malignancy determination, and treatment effect uncertainty. Given the recent advancements in image synthesis, Generative Adversarial Networks (GANs), and adversarial training, we assess the potential of these technologies to address a number of key challenges of cancer imaging. We categorise these challenges into (a) data scarcity and imbalance, (b) data access and privacy, (c) data annotation and segmentation, (d) cancer detection and diagnosis, and (e) tumour profiling, treatment planning and monitoring. Based on our analysis of 164 publications that apply adversarial training techniques in the context of cancer imaging, we highlight multiple underexplored solutions with research potential. We further contribute the Synthesis Study Trustworthiness Test (SynTRUST), a meta-analysis framework for assessing the validation rigour of medical image synthesis studies. SynTRUST is based on 26 concrete measures of thoroughness, reproducibility, usefulness, scalability, and tenability. Based on SynTRUST, we analyse 16 of the most promising cancer imaging challenge solutions and observe a high validation rigour in general, but also several desirable improvements. With this work, we strive to bridge the gap between the needs of the clinical cancer imaging community and the current and prospective research on data synthesis and adversarial networks in the artificial intelligence community.
Collapse
Affiliation(s)
- Richard Osuala
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain.
| | - Kaisar Kushibar
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Lidia Garrucho
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Akis Linardos
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Zuzanna Szafranowska
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Stefan Klein
- Biomedical Imaging Group Rotterdam, Department of Radiology & Nuclear Medicine, Erasmus MC, Rotterdam, The Netherlands
| | - Ben Glocker
- Biomedical Image Analysis Group, Department of Computing, Imperial College London, UK
| | - Oliver Diaz
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Karim Lekadir
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| |
Collapse
|
8
|
Wang H, Xiao N, Luo S, Li R, Zhao J, Ma Y, Zhao J, Qiang Y, Wang L, Lian J. Multi-scale dense selective network based on border modeling for lung nodule segmentation. Int J Comput Assist Radiol Surg 2023; 18:845-853. [PMID: 36637749 DOI: 10.1007/s11548-022-02817-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2022] [Accepted: 12/20/2022] [Indexed: 01/14/2023]
Abstract
PURPOSE Accurate quantification of pulmonary nodules helps physicians to accurately diagnose and treat lung cancer. We try to improve the segmentation efficiency of irregular nodules while maintaining the segmentation accuracy of simple types of nodules. METHODS In this paper, we obtain the unique edge part of pulmonary nodules and process it as a single branch stream, i.e., border stream, to explicitly model the nodule edge information. We propose a multi-scale dense selective network based on border modeling (BorDenNet). Its overall framework consists of a dual-branch encoder-decoder, which achieves parallel processing of classical image stream and border stream. We design a dense attention module to facilitate a strongly coupled status of feature images to focus on key regions of pulmonary nodules. Then, during the process of model decoding, the multi-scale selective attention module is proposed to establish long-range correlation relationships between different scale features, which further achieves finer feature discrimination and spatial recovery. We introduce border context enhancement module to mutually fuse and enhance the edge-related voxel features contained in the image stream and border stream and finally achieve the accurate segmentation of pulmonary nodules. RESULTS We evaluate the BorDenNet rigorously on the lung public dataset LIDC-IDRI. For the segmentation of the target nodules, the average Dice score is 92.78[Formula: see text], the average sensitivity is 91.37[Formula: see text], and the average Hausdorff distance is 3.06 mm. We further test on a private dataset from Shanxi Provincial People's Hospital, which verifies the excellent generalization of BorDenNet. Our BorDenNet relatively improves the segmentation efficiency for multi-type nodules such as adherent pulmonary nodules and ground-glass pulmonary nodules. CONCLUSION Accurate segmentation of irregular pulmonary nodules can obtain important clinical parameters, which can be used as a guide for clinicians and improve clinical efficiency.
Collapse
Affiliation(s)
- Hexi Wang
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Ning Xiao
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Shichao Luo
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Runrui Li
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Jun Zhao
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Yulan Ma
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Juanjuan Zhao
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China.
- College of Information, Jinzhong College of Information, Jinzhong, 030600, Shanxi, China.
| | - Yan Qiang
- College of Information and Computer, Taiyuan University of Technology, Taiyuan, 030000, Shanxi, China
| | - Long Wang
- College of Information, Jinzhong College of Information, Jinzhong, 030600, Shanxi, China
| | - Jianhong Lian
- Cancer Hospital, Shanxi Cancer Hospital, Taiyuan, 030000, Shanxi, China
| |
Collapse
|
9
|
Tyagi S, Talbar SN. CSE-GAN: A 3D conditional generative adversarial network with concurrent squeeze-and-excitation blocks for lung nodule segmentation. Comput Biol Med 2022; 147:105781. [DOI: 10.1016/j.compbiomed.2022.105781] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2022] [Revised: 06/16/2022] [Accepted: 06/19/2022] [Indexed: 11/03/2022]
|
10
|
Ferreira A, Pereira T, Silva F, Vilares AT, Silva MC, Cunha A, Oliveira HP. Synthesizing 3D Lung CT scans with Generative Adversarial Networks. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:2033-2036. [PMID: 36085795 DOI: 10.1109/embc48229.2022.9871481] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
In the healthcare domain, datasets are often private and lack large amounts of samples, making it difficult to cope with the inherent patient data heterogeneity. As an attempt to mitigate data scarcity, generative models are being used due to their ability to produce new data, using a dataset as a reference. However, synthesis studies often rely on a 2D representation of data, a seriously limited form of information when it comes to lung computed tomography scans where, for example, pathologies like nodules can manifest anywhere in the organ. Here, we develop a 3D Progressive Growing Generative Adversarial Network capable of generating thoracic CT volumes at a resolution of 1283, and analyze the model outputs through a quantitative metric (3D Muli-Scale Structural Similarity) and a Visual Turing Test. Clinical relevance - This paper is a novel application of the 3D PGGAN model to synthesize CT lung scans. This preliminary study focuses on synthesizing the entire volume of the lung rather than just the lung nodules. The synthesized data represent an attempt to mitigate data scarcity which is one of the major limitations to create learning models with good generalization in healthcare.
Collapse
|
11
|
Huang Y, Holcombe SA, Zhou Q, Wang SC, Tang J, Nie B. A GAN based approach for inferring progression trajectories of costal cartilage calcification from cross-sectional data at image level. Comput Biol Med 2022; 146:105647. [PMID: 35617729 DOI: 10.1016/j.compbiomed.2022.105647] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2022] [Revised: 05/02/2022] [Accepted: 05/14/2022] [Indexed: 11/03/2022]
Abstract
BACKGROUND Costal cartilage calcification (CCC) increases with age and presents differently for men and women. In individuals, however, the cross-sectional studies that show such trends do not reveal the geometric trajectories through which calcification might accumulate across a lifetime. Generative adversarial networks have the potential to reveal such trajectories from cross-sectional data by learning population trends and synthesizing individualized images at progressive levels of calcification. METHODS Chest wall mid-surface CT images with normalized cartilage morphologies were produced for 379 subjects aged 6 to 90, and labeled by sex and calcification severity. A conditional GAN with added loss terms to favor one-way accumulation of CCC was trained using organized image batches. GAN performance was assessed by comparing the distributions of images between the training and synthetic groups. RESULTS Synthetic images generated from a common seed for a given sex and at successive calcification severity levels showed incremental and regional growth of calcification sites. CCC patterns for synthetic male and female images matched known sex-based differences, and individual CCC growth in synthetic images was consistent with previously observed population trends. These trends in the synthetic images were also quantified by structural similarity scores. Synthetic images generated from different input seeds further showed individual variance in specific regions and trajectories of CCC accumulation. CONCLUSION This study inferred individual progression of CCC accumulation from uncalcified to severely calcified using cross-sectional image data. This information can inform computational models of the changing chest wall biomechanics with age, and the GAN-based technique shows potential for inferring longitudinal data from population trends in other clinical areas.
Collapse
Affiliation(s)
- Yuan Huang
- State Key Laboratory of Automotive Safety and Energy, Tsinghua University, China; International Center for Automotive Medicine (ICAM), University of Michigan, USA.
| | - Sven A Holcombe
- International Center for Automotive Medicine (ICAM), University of Michigan, USA.
| | - Qing Zhou
- State Key Laboratory of Automotive Safety and Energy, Tsinghua University, China.
| | - Stewart C Wang
- International Center for Automotive Medicine (ICAM), University of Michigan, USA.
| | - Jisi Tang
- State Key Laboratory of Automotive Safety and Energy, Tsinghua University, China.
| | - Bingbing Nie
- State Key Laboratory of Automotive Safety and Energy, Tsinghua University, China.
| |
Collapse
|
12
|
DAS-Net: A lung nodule segmentation method based on adaptive dual-branch attention and shadow mapping. APPL INTELL 2022. [DOI: 10.1007/s10489-021-03038-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
13
|
Chen Y, Yang XH, Wei Z, Heidari AA, Zheng N, Li Z, Chen H, Hu H, Zhou Q, Guan Q. Generative Adversarial Networks in Medical Image augmentation: A review. Comput Biol Med 2022; 144:105382. [PMID: 35276550 DOI: 10.1016/j.compbiomed.2022.105382] [Citation(s) in RCA: 74] [Impact Index Per Article: 37.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2022] [Revised: 02/25/2022] [Accepted: 03/02/2022] [Indexed: 12/31/2022]
Abstract
OBJECT With the development of deep learning, the number of training samples for medical image-based diagnosis and treatment models is increasing. Generative Adversarial Networks (GANs) have attracted attention in medical image processing due to their excellent image generation capabilities and have been widely used in data augmentation. In this paper, a comprehensive and systematic review and analysis of medical image augmentation work are carried out, and its research status and development prospects are reviewed. METHOD This paper reviews 105 medical image augmentation related papers, which mainly collected by ELSEVIER, IEEE Xplore, and Springer from 2018 to 2021. We counted these papers according to the parts of the organs corresponding to the images, and sorted out the medical image datasets that appeared in them, the loss function in model training, and the quantitative evaluation metrics of image augmentation. At the same time, we briefly introduce the literature collected in three journals and three conferences that have received attention in medical image processing. RESULT First, we summarize the advantages of various augmentation models, loss functions, and evaluation metrics. Researchers can use this information as a reference when designing augmentation tasks. Second, we explore the relationship between augmented models and the amount of the training set, and tease out the role that augmented models may play when the quality of the training set is limited. Third, the statistical number of papers shows that the development momentum of this research field remains strong. Furthermore, we discuss the existing limitations of this type of model and suggest possible research directions. CONCLUSION We discuss GAN-based medical image augmentation work in detail. This method effectively alleviates the challenge of limited training samples for medical image diagnosis and treatment models. It is hoped that this review will benefit researchers interested in this field.
Collapse
Affiliation(s)
- Yizhou Chen
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Xu-Hua Yang
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Zihan Wei
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Ali Asghar Heidari
- School of Surveying and Geospatial Engineering, College of Engineering, University of Tehran, Tehran, Iran; Department of Computer Science, School of Computing, National University of Singapore, Singapore, Singapore.
| | - Nenggan Zheng
- Qiushi Academy for Advanced Studies, Zhejiang University, Hangzhou, Zhejiang, China.
| | - Zhicheng Li
- Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China.
| | - Huiling Chen
- College of Computer Science and Artificial Intelligence, Wenzhou University, Wenzhou, Zhejiang, 325035, China.
| | - Haigen Hu
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Qianwei Zhou
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Qiu Guan
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| |
Collapse
|
14
|
Nishio M, Fujimoto K, Matsuo H, Muramatsu C, Sakamoto R, Fujita H. Lung Cancer Segmentation With Transfer Learning: Usefulness of a Pretrained Model Constructed From an Artificial Dataset Generated Using a Generative Adversarial Network. Front Artif Intell 2021; 4:694815. [PMID: 34337394 PMCID: PMC8322116 DOI: 10.3389/frai.2021.694815] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2021] [Accepted: 06/21/2021] [Indexed: 12/14/2022] Open
Abstract
Purpose: The purpose of this study was to develop and evaluate lung cancer segmentation with a pretrained model and transfer learning. The pretrained model was constructed from an artificial dataset generated using a generative adversarial network (GAN). Materials and Methods: Three public datasets containing images of lung nodules/lung cancers were used: LUNA16 dataset, Decathlon lung dataset, and NSCLC radiogenomics. The LUNA16 dataset was used to generate an artificial dataset for lung cancer segmentation with the help of the GAN and 3D graph cut. Pretrained models were then constructed from the artificial dataset. Subsequently, the main segmentation model was constructed from the pretrained models and the Decathlon lung dataset. Finally, the NSCLC radiogenomics dataset was used to evaluate the main segmentation model. The Dice similarity coefficient (DSC) was used as a metric to evaluate the segmentation performance. Results: The mean DSC for the NSCLC radiogenomics dataset improved overall when using the pretrained models. At maximum, the mean DSC was 0.09 higher with the pretrained model than that without it. Conclusion: The proposed method comprising an artificial dataset and a pretrained model can improve lung cancer segmentation as confirmed in terms of the DSC metric. Moreover, the construction of the artificial dataset for the segmentation using the GAN and 3D graph cut was found to be feasible.
Collapse
Affiliation(s)
- Mizuho Nishio
- Department of Diagnostic Imaging and Nuclear Medicine, Kyoto University Graduate School of Medicine, Kyoto, Japan.,Department of Radiology, Kobe University Hospital, Kobe, Japan
| | - Koji Fujimoto
- Department of Diagnostic Imaging and Nuclear Medicine, Kyoto University Graduate School of Medicine, Kyoto, Japan.,Department of Real World Data Research and Development, Kyoto University Graduate School of Medicine, Hikone, Japan
| | | | | | - Ryo Sakamoto
- Department of Diagnostic Imaging and Nuclear Medicine, Kyoto University Graduate School of Medicine, Kyoto, Japan.,Preemptive Medicine and Lifestyle-Related Disease Research Center, Kyoto University Hospital, Kyoto, Japan
| | - Hiroshi Fujita
- Department of Electrical, Electronic and Computer Engineering, Faculty of Engineering, Gifu University, Gifu, Japan
| |
Collapse
|
15
|
Wang Y, Zhang J. CMMCSegNet: Cross-Modality Multicascade Indirect LGE Segmentation on Multimodal Cardiac MR. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2021; 2021:9942149. [PMID: 34194539 PMCID: PMC8203380 DOI: 10.1155/2021/9942149] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/13/2021] [Accepted: 05/24/2021] [Indexed: 11/18/2022]
Abstract
Since Late-Gadolinium Enhancement (LGE) of cardiac magnetic resonance (CMR) visualizes myocardial infarction, and the balanced-Steady State Free Precession (bSSFP) cine sequence can capture cardiac motions and present clear boundaries; multimodal CMR segmentation has played an important role in the assessment of myocardial viability and clinical diagnosis, while automatic and accurate CMR segmentation still remains challenging due to a very small amount of labeled LGE data and the relatively low contrasts of LGE. The main purpose of our work is to learn the real/fake bSSFP modality with ground truths to indirectly segment the LGE modality of cardiac MR by using a proposed cross-modality multicascade framework: cross-modality translation network and automatic segmentation network, respectively. In the segmentation stage, a novel multicascade pix2pix network is designed to segment the fake bSSFP sequence obtained from a cross-modality translation network. Moreover, we propose perceptual loss measuring features between ground truth and prediction, which are extracted from the pretrained vgg network in the segmentation stage. We evaluate the performance of the proposed method on the multimodal CMR dataset and verify its superiority over other state-of-the-art approaches under different network structures and different types of adversarial losses in terms of dice accuracy in testing. Therefore, the proposed network is promising for Indirect Cardiac LGE Segmentation in clinical applications.
Collapse
Affiliation(s)
- Yu Wang
- School of Mathematics and Computational Science, Xiangtan University, Xiangtan, Hunan 411105, China
| | - Jianping Zhang
- School of Mathematics and Computational Science, Xiangtan University, Xiangtan, Hunan 411105, China
| |
Collapse
|