1
|
Waisberg E, Ong J, Kamran SA, Masalkhi M, Paladugu P, Zaman N, Lee AG, Tavakkoli A. Generative artificial intelligence in ophthalmology. Surv Ophthalmol 2024:S0039-6257(24)00044-4. [PMID: 38762072 DOI: 10.1016/j.survophthal.2024.04.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Revised: 04/28/2024] [Accepted: 04/29/2024] [Indexed: 05/20/2024]
Abstract
Generative AI has revolutionized medicine over the past several years. A generative adversarial network (GAN) is a deep learning framework that has become a powerful technique in medicine, particularly in ophthalmology and image analysis. In this paper we review the current ophthalmic literature involving GANs, and highlight key contributions in the field. We briefly touch on ChatGPT, another application of generative AI, and its potential in ophthalmology. We also explore the potential uses for GANs in ocular imaging, with a specific emphasis on 3 primary domains: image enhancement, disease identification, and generating of synthetic data. PubMed, Ovid MEDLINE, Google Scholar were searched from inception to October 30, 2022 to identify applications of GAN in ophthalmology. A total of 40 papers were included in this review. We cover various applications of GANs in ophthalmic-related imaging including optical coherence tomography, orbital magnetic resonance imaging, fundus photography, and ultrasound; however, we also highlight several challenges, that resulted in the generation of inaccurate and atypical results during certain iterations. Finally, we examine future directions and considerations for generative AI in ophthalmology.
Collapse
Affiliation(s)
- Ethan Waisberg
- Department of Ophthalmology, University of Cambridge, Cambridge, United Kingdom.
| | - Joshua Ong
- Michigan Medicine, University of Michigan, Ann Arbor, United States
| | - Sharif Amit Kamran
- School of Medicine, University College Dublin, Belfield, Dublin, Ireland
| | - Mouayad Masalkhi
- School of Medicine, University College Dublin, Belfield, Dublin, Ireland
| | - Phani Paladugu
- Sidney Kimmel Medical College, Thomas Jefferson University, Philadelphia, Pennsylvania, United States; Brigham and Women's Hospital, Harvard Medical School, Boston, Massachusetts, United States
| | - Nasif Zaman
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, Nevada, United States
| | - Andrew G Lee
- Center for Space Medicine, Baylor College of Medicine, Houston, Texas, United States; Department of Ophthalmology, Blanton Eye Institute, Houston Methodist Hospital, Houston, Texas, United States; The Houston Methodist Research Institute, Houston Methodist Hospital, Houston, Texas, United States; Departments of Ophthalmology, Neurology, and Neurosurgery, Weill Cornell Medicine, New York, New York, United States; Department of Ophthalmology, University of Texas Medical Branch, Galveston, Texas, United States; University of Texas MD Anderson Cancer Center, Houston, Texas, United States; Texas A&M College of Medicine, Texas, United States; Department of Ophthalmology, The University of Iowa Hospitals and Clinics, Iowa City, Iowa, United States
| | - Alireza Tavakkoli
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, Nevada, United States
| |
Collapse
|
2
|
Alizadeh Eghtedar R, Vard A, Malekahmadi M, Peyman A. A new computer-aided diagnosis tool based on deep learning methods for automatic detection of retinal disorders from OCT images. Int Ophthalmol 2024; 44:110. [PMID: 38396074 DOI: 10.1007/s10792-024-03033-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Accepted: 01/11/2024] [Indexed: 02/25/2024]
Abstract
PURPOSE Early detection of retinal disorders using optical coherence tomography (OCT) images can prevent vision loss. Since manual screening can be time-consuming, tedious, and fallible, we present a reliable computer-aided diagnosis (CAD) software based on deep learning. Also, we made efforts to increase the interpretability of the deep learning methods, overcome their vague and black box nature, and also understand their behavior in the diagnosis. METHODS We propose a novel method to improve the interpretability of the used deep neural network by embedding the rich semantic information of abnormal areas based on the ophthalmologists' interpretations and medical descriptions in the OCT images. Finally, we trained the classification network on a small subset of the online publicly available University of California San Diego (UCSD) dataset with an overall of 29,800 OCT images. RESULTS The experimental results on the 1000 test OCT images show that the proposed method achieves the overall precision, accuracy, sensitivity, and f1-score of 97.6%, 97.6%, 97.6%, and 97.59%, respectively. Also, the heat map images provide a clear region of interest which indicates that the interpretability of the proposed method is increased dramatically. CONCLUSION The proposed software can help ophthalmologists in providing a second opinion to make a decision, and primitive automated diagnoses of retinal diseases and even it can be used as a screening tool, in eye clinics. Also, the improvement of the interpretability of the proposed method causes to increase in the model generalization, and therefore, it will work properly on a wide range of other OCT datasets.
Collapse
Affiliation(s)
- Reza Alizadeh Eghtedar
- Department of Bioelectrics and Biomedical Engineering, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
- Student Research Committee, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Alireza Vard
- Department of Bioelectrics and Biomedical Engineering, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran.
- Medical Image & Signal Processing Research Center, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran.
| | - Mohammad Malekahmadi
- Department of Ophthalmology, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
- Isfahan Eye Research Center, Department of Ophthalmology, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Alireza Peyman
- Department of Ophthalmology, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
- Isfahan Eye Research Center, Department of Ophthalmology, Isfahan University of Medical Sciences, Isfahan, Iran
| |
Collapse
|
3
|
Ait Hammou B, Antaki F, Boucher MC, Duval R. MBT: Model-Based Transformer for retinal optical coherence tomography image and video multi-classification. Int J Med Inform 2023; 178:105178. [PMID: 37657204 DOI: 10.1016/j.ijmedinf.2023.105178] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Revised: 07/13/2023] [Accepted: 08/06/2023] [Indexed: 09/03/2023]
Abstract
BACKGROUND AND OBJECTIVE The detection of retinal diseases using optical coherence tomography (OCT) images and videos is a concrete example of a data classification problem. In recent years, Transformer architectures have been successfully applied to solve a variety of real-world classification problems. Although they have shown impressive discriminative abilities compared to other state-of-the-art models, improving their performance is essential, especially in healthcare-related problems. METHODS This paper presents an effective technique named model-based transformer (MBT). It is based on popular pre-trained transformer models, particularly, vision transformer, swin transformer for OCT image classification, and multiscale vision transformer for OCT video classification. The proposed approach is designed to represent OCT data by taking advantage of an approximate sparse representation technique. Then, it estimates the optimal features, and performs data classification. RESULTS The experiments are carried out using three real-world retinal datasets. The experimental results on OCT image and OCT video datasets show that the proposed method outperforms existing state-of-the-art deep learning approaches in terms of classification accuracy, precision, recall, and f1-score, kappa, AUC-ROC, and AUC-PR. It can also boost the performance of existing transformer models, including Vision transformer and Swin transformer for OCT image classification, and Multiscale Vision Transformers for OCT video classification. CONCLUSIONS This work presents an approach for the automated detection of retinal diseases. Although deep neural networks have proven great potential in ophthalmology applications, our findings demonstrate for the first time a new way to identify retinal pathologies using OCT videos instead of images. Moreover, our proposal can help researchers enhance the discriminative capacity of a variety of powerful deep learning models presented in published papers. This can be valuable for future directions in medical research and clinical practice.
Collapse
Affiliation(s)
- Badr Ait Hammou
- Department of Ophthalmology, Université de Montréal, Montreal, Québec, Canada; Centre Universitaire d'Ophtalmologie (CUO), Hôpital Maisonneuve-Rosemont, CIUSSS de l'Est-de-l'Île-de-Montréal, Montréal, Québec, Canada.
| | - Fares Antaki
- Department of Ophthalmology, Université de Montréal, Montreal, Québec, Canada; Centre Universitaire d'Ophtalmologie (CUO), Hôpital Maisonneuve-Rosemont, CIUSSS de l'Est-de-l'Île-de-Montréal, Montréal, Québec, Canada; Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Quebec, Canada
| | - Marie-Carole Boucher
- Department of Ophthalmology, Université de Montréal, Montreal, Québec, Canada; Centre Universitaire d'Ophtalmologie (CUO), Hôpital Maisonneuve-Rosemont, CIUSSS de l'Est-de-l'Île-de-Montréal, Montréal, Québec, Canada
| | - Renaud Duval
- Department of Ophthalmology, Université de Montréal, Montreal, Québec, Canada; Centre Universitaire d'Ophtalmologie (CUO), Hôpital Maisonneuve-Rosemont, CIUSSS de l'Est-de-l'Île-de-Montréal, Montréal, Québec, Canada
| |
Collapse
|
4
|
Tripathi A, Kumar P, Mayya V, Tulsani A. Generating OCT B-Scan DME images using optimized Generative Adversarial Networks (GANs). Heliyon 2023; 9:e18773. [PMID: 37609420 PMCID: PMC10440457 DOI: 10.1016/j.heliyon.2023.e18773] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2023] [Revised: 07/26/2023] [Accepted: 07/27/2023] [Indexed: 08/24/2023] Open
Abstract
Diabetic Macular Edema (DME) represents a significant visual impairment among individuals with diabetes, leading to a dramatic reduction in visual acuity and potentially resulting in irreversible vision loss. Optical Coherence Tomography (OCT), a technique that produces high-resolution retinal images, plays a vital role in the clinical assessment of this condition. Physicians typically rely on OCT B-Scan images to evaluate DME severity. However, manual interpretation of these images is susceptible to errors, which can lead to detrimental consequences, such as misdiagnosis and improper treatment strategies. Hence, there is a critical need for more reliable diagnostic methods. This study aims to address this gap by proposing an automated model based on Generative Adversarial Networks (GANs) to generate OCT B-Scan images of DME. The model synthesizes images from patients' baseline OCT B-Scan images, which could potentially enhance the robustness of DME detection systems. We employ five distinct GANs in this study: Deep Convolutional GAN, Conditional GAN, CycleGAN, StyleGAN2, and StyleGAN3, drawing comparisons across their performance. Subsequently, the hyperparameters of the best-performing GAN are fine-tuned using Particle Swarm Optimization (PSO) to produce more realistic OCT images. This comparative analysis not only serves to improve the detection of DME severity using OCT images but also provides insights into the appropriate choice of GANs for the effective generation of realistic OCT images from the baseline OCT datasets.
Collapse
Affiliation(s)
- Aditya Tripathi
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Preetham Kumar
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Veena Mayya
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Akshat Tulsani
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| |
Collapse
|
5
|
Tripathi A, Kumar P, Tulsani A, Chakrapani PK, Maiya G, Bhandary SV, Mayya V, Pathan S, Achar R, Acharya UR. Fuzzy Logic-Based System for Identifying the Severity of Diabetic Macular Edema from OCT B-Scan Images Using DRIL, HRF, and Cystoids. Diagnostics (Basel) 2023; 13:2550. [PMID: 37568913 PMCID: PMC10416860 DOI: 10.3390/diagnostics13152550] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2023] [Revised: 07/19/2023] [Accepted: 07/21/2023] [Indexed: 08/13/2023] Open
Abstract
Diabetic Macular Edema (DME) is a severe ocular complication commonly found in patients with diabetes. The condition can precipitate a significant drop in VA and, in extreme cases, may result in irreversible vision loss. Optical Coherence Tomography (OCT), a technique that yields high-resolution retinal images, is often employed by clinicians to assess the extent of DME in patients. However, the manual interpretation of OCT B-scan images for DME identification and severity grading can be error-prone, with false negatives potentially resulting in serious repercussions. In this paper, we investigate an Artificial Intelligence (AI) driven system that offers an end-to-end automated model, designed to accurately determine DME severity using OCT B-Scan images. This model operates by extracting specific biomarkers such as Disorganization of Retinal Inner Layers (DRIL), Hyper Reflective Foci (HRF), and cystoids from the OCT image, which are then utilized to ascertain DME severity. The rules guiding the fuzzy logic engine are derived from contemporary research in the field of DME and its association with various biomarkers evident in the OCT image. The proposed model demonstrates high efficacy, identifying images with DRIL with 93.3% accuracy and successfully segmenting HRF and cystoids from OCT images with dice similarity coefficients of 91.30% and 95.07% respectively. This study presents a comprehensive system capable of accurately grading DME severity using OCT B-scan images, serving as a potentially invaluable tool in the clinical assessment and treatment of DME.
Collapse
Affiliation(s)
- Aditya Tripathi
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Preetham Kumar
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Akshat Tulsani
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Pavithra Kodiyalbail Chakrapani
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Geetha Maiya
- Department of Computer Science and Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Sulatha V. Bhandary
- Department of Ophthalmology, Kasturba Medical College, Manipal Academy of Higher Education, Manipal 576104, India
| | - Veena Mayya
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Sameena Pathan
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - Raghavendra Achar
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal 576104, India
| | - U. Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield Central, QLD 4300, Australia
| |
Collapse
|
6
|
Image-to-image translation with Generative Adversarial Networks via retinal masks for realistic Optical Coherence Tomography imaging of Diabetic Macular Edema disorders. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104098] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
7
|
Pavithra K, Kumar P, Geetha M, Bhandary SV. Computer aided diagnosis of diabetic macular edema in retinal fundus and OCT images: A review. Biocybern Biomed Eng 2023. [DOI: 10.1016/j.bbe.2022.12.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
|
8
|
Sreejith Kumar AJ, Chong RS, Crowston JG, Chua J, Bujor I, Husain R, Vithana EN, Girard MJA, Ting DSW, Cheng CY, Aung T, Popa-Cherecheanu A, Schmetterer L, Wong D. Evaluation of Generative Adversarial Networks for High-Resolution Synthetic Image Generation of Circumpapillary Optical Coherence Tomography Images for Glaucoma. JAMA Ophthalmol 2022; 140:974-981. [PMID: 36048435 PMCID: PMC9437828 DOI: 10.1001/jamaophthalmol.2022.3375] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/14/2022]
Abstract
Importance Deep learning (DL) networks require large data sets for training, which can be challenging to collect clinically. Generative models could be used to generate large numbers of synthetic optical coherence tomography (OCT) images to train such DL networks for glaucoma detection. Objective To assess whether generative models can synthesize circumpapillary optic nerve head OCT images of normal and glaucomatous eyes and determine the usability of synthetic images for training DL models for glaucoma detection. Design, Setting, and Participants Progressively growing generative adversarial network models were trained to generate circumpapillary OCT scans. Image gradeability and authenticity were evaluated on a clinical set of 100 real and 100 synthetic images by 2 clinical experts. DL networks for glaucoma detection were trained with real or synthetic images and evaluated on independent internal and external test data sets of 140 and 300 real images, respectively. Main Outcomes and Measures Evaluations of the clinical set between the experts were compared. Glaucoma detection performance of the DL networks was assessed using area under the curve (AUC) analysis. Class activation maps provided visualizations of the regions contributing to the respective classifications. Results A total of 990 normal and 862 glaucomatous eyes were analyzed. Evaluations of the clinical set were similar for gradeability (expert 1: 92.0%; expert 2: 93.0%) and authenticity (expert 1: 51.8%; expert 2: 51.3%). The best-performing DL network trained on synthetic images had AUC scores of 0.97 (95% CI, 0.95-0.99) on the internal test data set and 0.90 (95% CI, 0.87-0.93) on the external test data set, compared with AUCs of 0.96 (95% CI, 0.94-0.99) on the internal test data set and 0.84 (95% CI, 0.80-0.87) on the external test data set for the network trained with real images. An increase in the AUC for the synthetic DL network was observed with the use of larger synthetic data set sizes. Class activation maps showed that the regions of the synthetic images contributing to glaucoma detection were generally similar to that of real images. Conclusions and Relevance DL networks trained with synthetic OCT images for glaucoma detection were comparable with networks trained with real images. These results suggest potential use of generative models in the training of DL networks and as a means of data sharing across institutions without patient information confidentiality issues.
Collapse
Affiliation(s)
- Ashish Jith Sreejith Kumar
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,SERI-NTU Advanced Ocular Engineering (STANCE), Singapore, Singapore.,Institute for Infocomm Research, A*STAR, Singapore
| | - Rachel S Chong
- Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Jonathan G Crowston
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Jacqueline Chua
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Inna Bujor
- Carol Davila University of Medicine and Pharmacy, Bucharest, Romania
| | - Rahat Husain
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Eranga N Vithana
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Michaël J A Girard
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore.,Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland
| | - Daniel S W Ting
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Ching-Yu Cheng
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore.,Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland
| | - Tin Aung
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore.,Department of Ophthalmology, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
| | - Alina Popa-Cherecheanu
- Carol Davila University of Medicine and Pharmacy, Bucharest, Romania.,Emergency University Hospital, Department of Ophthalmology, Bucharest, Romania
| | - Leopold Schmetterer
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,Academic Clinical Program, Duke-NUS Medical School, Singapore.,SERI-NTU Advanced Ocular Engineering (STANCE), Singapore, Singapore.,Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland.,Department of Ophthalmology and Optometry, Medical University Vienna, Vienna, Austria.,School of Chemical and Biomedical Engineering, Nanyang Technological University, Singapore.,Department of Clinical Pharmacology, Medical University Vienna, Vienna, Austria.,Center for Medical Physics and Biomedical Engineering, Medical University Vienna, Vienna, Austria
| | - Damon Wong
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore.,SERI-NTU Advanced Ocular Engineering (STANCE), Singapore, Singapore.,School of Chemical and Biomedical Engineering, Nanyang Technological University, Singapore
| |
Collapse
|
9
|
Ma Z, Xie Q, Xie P, Fan F, Gao X, Zhu J. HCTNet: A Hybrid ConvNet-Transformer Network for Retinal Optical Coherence Tomography Image Classification. BIOSENSORS 2022; 12:542. [PMID: 35884345 PMCID: PMC9313149 DOI: 10.3390/bios12070542] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/09/2022] [Revised: 07/13/2022] [Accepted: 07/18/2022] [Indexed: 06/15/2023]
Abstract
Automatic and accurate optical coherence tomography (OCT) image classification is of great significance to computer-assisted diagnosis of retinal disease. In this study, we propose a hybrid ConvNet-Transformer network (HCTNet) and verify the feasibility of a Transformer-based method for retinal OCT image classification. The HCTNet first utilizes a low-level feature extraction module based on the residual dense block to generate low-level features for facilitating the network training. Then, two parallel branches of the Transformer and the ConvNet are designed to exploit the global and local context of the OCT images. Finally, a feature fusion module based on an adaptive re-weighting mechanism is employed to combine the extracted global and local features for predicting the category of OCT images in the testing datasets. The HCTNet combines the advantage of the convolutional neural network in extracting local features and the advantage of the vision Transformer in establishing long-range dependencies. A verification on two public retinal OCT datasets shows that our HCTNet method achieves an overall accuracy of 91.56% and 86.18%, respectively, outperforming the pure ViT and several ConvNet-based classification methods.
Collapse
Affiliation(s)
- Zongqing Ma
- Key Laboratory of the Ministry of Education for Optoelectronic Measurement Technology and Instrument, Beijing Information Science and Technology University, Beijing 100192, China; (Z.M.); (Q.X.); (F.F.)
- Beijing Laboratory of Biomedical Testing Technology and Instruments, Beijing Information Science and Technology University, Beijing 100192, China
| | - Qiaoxue Xie
- Key Laboratory of the Ministry of Education for Optoelectronic Measurement Technology and Instrument, Beijing Information Science and Technology University, Beijing 100192, China; (Z.M.); (Q.X.); (F.F.)
- Beijing Laboratory of Biomedical Testing Technology and Instruments, Beijing Information Science and Technology University, Beijing 100192, China
| | - Pinxue Xie
- Beijing Anzhen Hospital, Capital Medical University, Beijing 100029, China; (P.X.); (X.G.)
| | - Fan Fan
- Key Laboratory of the Ministry of Education for Optoelectronic Measurement Technology and Instrument, Beijing Information Science and Technology University, Beijing 100192, China; (Z.M.); (Q.X.); (F.F.)
- Beijing Laboratory of Biomedical Testing Technology and Instruments, Beijing Information Science and Technology University, Beijing 100192, China
| | - Xinxiao Gao
- Beijing Anzhen Hospital, Capital Medical University, Beijing 100029, China; (P.X.); (X.G.)
| | - Jiang Zhu
- Key Laboratory of the Ministry of Education for Optoelectronic Measurement Technology and Instrument, Beijing Information Science and Technology University, Beijing 100192, China; (Z.M.); (Q.X.); (F.F.)
- Beijing Laboratory of Biomedical Testing Technology and Instruments, Beijing Information Science and Technology University, Beijing 100192, China
| |
Collapse
|
10
|
Li F. Construction of Computer Microscope Image Segmentation Model Based on Fourth-Order Partial Differential Equation Smoothing. SCANNING 2022; 2022:4355184. [PMID: 35937671 PMCID: PMC9296342 DOI: 10.1155/2022/4355184] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/05/2022] [Revised: 06/26/2022] [Accepted: 07/01/2022] [Indexed: 06/15/2023]
Abstract
In order to solve the problem of image noise, the author proposes a computer microscope image segmentation model based on the smoothing of fourth-order partial differential equations. On the basis of the functional describing the smoothness of the image by the directional curvature modulus, the author deduces a fourth-order partial differential equation (PDE) image noise reduction model, while effectively reducing noise, the edges are well preserved. The processing result of this method is a piecewise linear image, and there is a step in the gradient at the edge of the target. Taking advantage of this feature of the noise reduction results, the author proposes a new geodesic active contour model. The experimental results show that the reference method directly segments the results, iterates 10 times, and takes 160.721 seconds. Using the noise reduction model in the paper to preprocess and then using the reference method to segment the result, iterating 8 times, it takes 32.347 seconds. Conclusion. The new model is not only stable but also has strong contour extraction ability and fast convergence speed.
Collapse
Affiliation(s)
- Feng Li
- Yellow River Conservancy Technical Institute, Kaifeng Henan 475004, China
| |
Collapse
|
11
|
Design and Implementation of Obstetric Central Monitoring System Based on Medical Image Segmentation Algorithm. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:3545831. [PMID: 35529540 PMCID: PMC9072048 DOI: 10.1155/2022/3545831] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/10/2021] [Accepted: 08/13/2021] [Indexed: 11/17/2022]
Abstract
At present, the incidence of emergencies in obstetric care environment is gradually increasing, and different obstetric wards often have a variety of situations. Therefore, it can provide great help in clinical medicine to give early warning and plan coping plans according to different situations. This paper studied an obstetrics central surveillance system based on a medical image segmentation algorithm. Images obtained by central obstetrics monitoring are segmented, magnified in detail, and image features are extracted, collated, and trained. The normal distribution rule is used to classify the features, which are included in the feature library of the obstetric central monitoring system. In the gray space of the medical image, the statistical distribution of gray features of the medical image is described by the mixture model of Rayleigh distribution and Gaussian distribution. In the gray space of the medical image, Taylor series expansion is used to describe the linear geometric structure of medicine. The eigenvalues of Hessian matrix are introduced to obtain high-order multiscale features of medicine. The multiscale feature energy function is introduced into Markov random energy objective function to realize medical image segmentation. Compared with other segmentation algorithms, the accuracy and sensitivity of the proposed algorithm are 87.98% and 86.58%, respectively, which can clearly segment small medical features.
Collapse
|
12
|
Sotoudeh-Paima S, Jodeiri A, Hajizadeh F, Soltanian-Zadeh H. Multi-scale convolutional neural network for automated AMD classification using retinal OCT images. Comput Biol Med 2022; 144:105368. [DOI: 10.1016/j.compbiomed.2022.105368] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2021] [Revised: 02/28/2022] [Accepted: 02/28/2022] [Indexed: 11/29/2022]
|
13
|
You A, Kim JK, Ryu IH, Yoo TK. Application of generative adversarial networks (GAN) for ophthalmology image domains: a survey. EYE AND VISION (LONDON, ENGLAND) 2022; 9:6. [PMID: 35109930 PMCID: PMC8808986 DOI: 10.1186/s40662-022-00277-3] [Citation(s) in RCA: 48] [Impact Index Per Article: 24.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/15/2021] [Accepted: 01/11/2022] [Indexed: 12/12/2022]
Abstract
BACKGROUND Recent advances in deep learning techniques have led to improved diagnostic abilities in ophthalmology. A generative adversarial network (GAN), which consists of two competing types of deep neural networks, including a generator and a discriminator, has demonstrated remarkable performance in image synthesis and image-to-image translation. The adoption of GAN for medical imaging is increasing for image generation and translation, but it is not familiar to researchers in the field of ophthalmology. In this work, we present a literature review on the application of GAN in ophthalmology image domains to discuss important contributions and to identify potential future research directions. METHODS We performed a survey on studies using GAN published before June 2021 only, and we introduced various applications of GAN in ophthalmology image domains. The search identified 48 peer-reviewed papers in the final review. The type of GAN used in the analysis, task, imaging domain, and the outcome were collected to verify the usefulness of the GAN. RESULTS In ophthalmology image domains, GAN can perform segmentation, data augmentation, denoising, domain transfer, super-resolution, post-intervention prediction, and feature extraction. GAN techniques have established an extension of datasets and modalities in ophthalmology. GAN has several limitations, such as mode collapse, spatial deformities, unintended changes, and the generation of high-frequency noises and artifacts of checkerboard patterns. CONCLUSIONS The use of GAN has benefited the various tasks in ophthalmology image domains. Based on our observations, the adoption of GAN in ophthalmology is still in a very early stage of clinical validation compared with deep learning classification techniques because several problems need to be overcome for practical use. However, the proper selection of the GAN technique and statistical modeling of ocular imaging will greatly improve the performance of each image analysis. Finally, this survey would enable researchers to access the appropriate GAN technique to maximize the potential of ophthalmology datasets for deep learning research.
Collapse
Affiliation(s)
- Aram You
- School of Architecture, Kumoh National Institute of Technology, Gumi, Gyeongbuk, South Korea
| | - Jin Kuk Kim
- B&VIIT Eye Center, Seoul, South Korea
- VISUWORKS, Seoul, South Korea
| | - Ik Hee Ryu
- B&VIIT Eye Center, Seoul, South Korea
- VISUWORKS, Seoul, South Korea
| | - Tae Keun Yoo
- B&VIIT Eye Center, Seoul, South Korea.
- Department of Ophthalmology, Aerospace Medical Center, Republic of Korea Air Force, 635 Danjae-ro, Namil-myeon, Cheongwon-gun, Cheongju, Chungcheongbuk-do, 363-849, South Korea.
| |
Collapse
|
14
|
Barua PD, Chan WY, Dogan S, Baygin M, Tuncer T, Ciaccio EJ, Islam N, Cheong KH, Shahid ZS, Acharya UR. Multilevel Deep Feature Generation Framework for Automated Detection of Retinal Abnormalities Using OCT Images. ENTROPY (BASEL, SWITZERLAND) 2021; 23:1651. [PMID: 34945957 PMCID: PMC8700736 DOI: 10.3390/e23121651] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Revised: 11/22/2021] [Accepted: 11/25/2021] [Indexed: 01/04/2023]
Abstract
Optical coherence tomography (OCT) images coupled with many learning techniques have been developed to diagnose retinal disorders. This work aims to develop a novel framework for extracting deep features from 18 pre-trained convolutional neural networks (CNN) and to attain high performance using OCT images. In this work, we have developed a new framework for automated detection of retinal disorders using transfer learning. This model consists of three phases: deep fused and multilevel feature extraction, using 18 pre-trained networks and tent maximal pooling, feature selection with ReliefF, and classification using the optimized classifier. The novelty of this proposed framework is the feature generation using widely used CNNs and to select the most suitable features for classification. The extracted features using our proposed intelligent feature extractor are fed to iterative ReliefF (IRF) to automatically select the best feature vector. The quadratic support vector machine (QSVM) is utilized as a classifier in this work. We have developed our model using two public OCT image datasets, and they are named database 1 (DB1) and database 2 (DB2). The proposed framework can attain 97.40% and 100% classification accuracies using the two OCT datasets, DB1 and DB2, respectively. These results illustrate the success of our model.
Collapse
Affiliation(s)
- Prabal Datta Barua
- School of Management & Enterprise, University of Southern Queensland, Toowoomba, QLD 4350, Australia;
- Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia
- Cogninet Brain Team, Cogninet Australia, Sydney, NSW 2010, Australia
| | - Wai Yee Chan
- University Malaya Research Imaging Centre, Department of Biomedical Imaging, Faculty of Medicine, University of Malaya, Kuala Lumpur 59100, Malaysia;
| | - Sengul Dogan
- Department of Digital Forensics Engineering, College of Technology, Firat University, Elazig 23002, Turkey; (S.D.); (T.T.)
| | - Mehmet Baygin
- Department of Computer Engineering, College of Engineering, Ardahan University, Ardahan 75000, Turkey;
| | - Turker Tuncer
- Department of Digital Forensics Engineering, College of Technology, Firat University, Elazig 23002, Turkey; (S.D.); (T.T.)
| | - Edward J. Ciaccio
- Department of Medicine, Columbia University Irving Medical Center, New York, NY 10032-3784, USA;
| | - Nazrul Islam
- Glaucoma Faculty, Bangladesh Eye Hospital & Institute, Dhaka 1206, Bangladesh;
| | - Kang Hao Cheong
- Science, Mathematics and Technology Cluster, Singapore University of Technology and Design, Singapore 487372, Singapore
| | - Zakia Sultana Shahid
- Department of Ophthalmology, Anwer Khan Modern Medical College, Dhaka 1205, Bangladesh;
| | - U. Rajendra Acharya
- Department of Electronics and Computer Engineering, Ngee Ann Polytechnic, Singapore 599489, Singapore
- Department of Biomedical Engineering, School of Science and Technology, SUSS University, Singapore 129799, Singapore
- Department of Biomedical Informatics and Medical Engineering, Asia University, Taichung 41354, Taiwan
| |
Collapse
|
15
|
Zhang Y, Li M, Ji Z, Fan W, Yuan S, Liu Q, Chen Q. Twin self-supervision based semi-supervised learning (TS-SSL): Retinal anomaly classification in SD-OCT images. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.08.051] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/29/2023]
|
16
|
Glaucoma classification based on scanning laser ophthalmoscopic images using a deep learning ensemble method. PLoS One 2021; 16:e0252339. [PMID: 34086716 PMCID: PMC8177489 DOI: 10.1371/journal.pone.0252339] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2021] [Accepted: 05/12/2021] [Indexed: 12/21/2022] Open
Abstract
This study aimed to assess the utility of optic nerve head (onh) en-face images, captured with scanning laser ophthalmoscopy (slo) during standard optical coherence tomography (oct) imaging of the posterior segment, and demonstrate the potential of deep learning (dl) ensemble method that operates in a low data regime to differentiate glaucoma patients from healthy controls. The two groups of subjects were initially categorized based on a range of clinical tests including measurements of intraocular pressure, visual fields, oct derived retinal nerve fiber layer (rnfl) thickness and dilated stereoscopic examination of onh. 227 slo images of 227 subjects (105 glaucoma patients and 122 controls) were used. A new task-specific convolutional neural network architecture was developed for slo image-based classification. To benchmark the results of the proposed method, a range of classifiers were tested including five machine learning methods to classify glaucoma based on rnfl thickness—a well-known biomarker in glaucoma diagnostics, ensemble classifier based on inception v3 architecture, and classifiers based on features extracted from the image. The study shows that cross-validation dl ensemble based on slo images achieved a good discrimination performance with up to 0.962 of balanced accuracy, outperforming all of the other tested classifiers.
Collapse
|
17
|
Various Generative Adversarial Networks Model for Synthetic Prohibitory Sign Image Generation. APPLIED SCIENCES-BASEL 2021. [DOI: 10.3390/app11072913] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
A synthetic image is a critical issue for computer vision. Traffic sign images synthesized from standard models are commonly used to build computer recognition algorithms for acquiring more knowledge on various and low-cost research issues. Convolutional Neural Network (CNN) achieves excellent detection and recognition of traffic signs with sufficient annotated training data. The consistency of the entire vision system is dependent on neural networks. However, locating traffic sign datasets from most countries in the world is complicated. This work uses various generative adversarial networks (GAN) models to construct intricate images, such as Least Squares Generative Adversarial Networks (LSGAN), Deep Convolutional Generative Adversarial Networks (DCGAN), and Wasserstein Generative Adversarial Networks (WGAN). This paper also discusses, in particular, the quality of the images produced by various GANs with different parameters. For processing, we use a picture with a specific number and scale. The Structural Similarity Index (SSIM) and Mean Squared Error (MSE) will be used to measure image consistency. Between the generated image and the corresponding real image, the SSIM values will be compared. As a result, the images display a strong similarity to the real image when using more training images. LSGAN outperformed other GAN models in the experiment with maximum SSIM values achieved using 200 images as inputs, 2000 epochs, and size 32 × 32.
Collapse
|