1
|
Mohseni A, Ghotbi E, Kazemi F, Shababi A, Jahan SC, Mohseni A, Shababi N. Artificial Intelligence in Radiology: What Is Its True Role at Present, and Where Is the Evidence? Radiol Clin North Am 2024; 62:935-947. [PMID: 39393852 DOI: 10.1016/j.rcl.2024.03.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/13/2024]
Abstract
The integration of artificial intelligence (AI) in radiology has brought about substantial advancements and transformative potential in diagnostic imaging practices. This study presents an overview of the current research on the application of AI in radiology, highlighting key insights from recent studies and surveys. These recent studies have explored the expected impact of AI, encompassing machine learning and deep learning, on the work volume of diagnostic radiologists. The present and future role of AI in radiology holds great promise for enhancing diagnostic capabilities, improving workflow efficiency, and ultimately, advancing patient care.
Collapse
Affiliation(s)
- Alireza Mohseni
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA.
| | - Elena Ghotbi
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA
| | - Foad Kazemi
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA
| | - Amirali Shababi
- School of Medicine, Iran University of Medical Sciences, Hemat Highway next to Milad Tower 14535, Tehran, Iran
| | - Shayan Chashm Jahan
- Department of Computer Science, University of Maryland, 8125 Paint Branch Drive College Park, MD 20742, USA
| | - Anita Mohseni
- Azad University Tehran Medical Branch, Danesh, Shariati Street, Tehran, Iran 19395/1495
| | - Niloufar Shababi
- Johns Hopkins University School of Medicine, 600 N. Wolfe Street / Phipps 446, Baltimore, MD 21287, USA
| |
Collapse
|
2
|
Linde G, Rodrigues de Souza W, Chalakkal R, Danesh-Meyer HV, O'Keeffe B, Chiong Hong S. A comparative evaluation of deep learning approaches for ophthalmology. Sci Rep 2024; 14:21829. [PMID: 39294275 PMCID: PMC11410932 DOI: 10.1038/s41598-024-72752-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Accepted: 09/09/2024] [Indexed: 09/20/2024] Open
Abstract
There is a growing number of publicly available ophthalmic imaging datasets and open-source code for Machine Learning algorithms. This allows ophthalmic researchers and practitioners to independently perform various deep-learning tasks. With the advancement in artificial intelligence (AI) and in the field of imaging, the choice of the most appropriate AI architecture for different tasks will vary greatly. The best-performing AI-dataset combination will depend on the specific problem that needs to be solved and the type of data available. The article discusses different machine learning models and deep learning architectures currently used for various ophthalmic imaging modalities and for different machine learning tasks. It also proposes the most appropriate models based on accuracy and other important factors such as training time, the ability to deploy the model on clinical devices/smartphones, heatmaps that enhance the self-explanatory nature of classification decisions, and the ability to train/adapt on small image datasets to determine if further data collection is worthwhile. The article extensively reviews the existing state-of-the-art AI methods focused on useful machine-learning applications for ophthalmology. It estimates their performance and viability through training and evaluating architectures with different public and private image datasets of different modalities, such as full-color retinal images, OCT images, and 3D OCT scans. The article is expected to benefit the readers by enriching their knowledge of artificial intelligence applied to ophthalmology.
Collapse
Affiliation(s)
- Glenn Linde
- oDocs Eye Care Research, Dunedin, New Zealand
| | - Waldir Rodrigues de Souza
- Department of Ophthalmology, Dunedin Hospital, Te Whatu Ora Southern, Dunedin, New Zealand
- Department of Medicine, Ophthalmology Section, University of Otago, Dunedin, New Zealand
| | | | | | | | - Sheng Chiong Hong
- oDocs Eye Care Research, Dunedin, New Zealand
- Department of Ophthalmology, Dunedin Hospital, Te Whatu Ora Southern, Dunedin, New Zealand
| |
Collapse
|
3
|
Likassa HT, Chen DG, Chen K, Wang Y, Zhu W. Robust PCA with Lw,∗ and L2,1 Norms: A Novel Method for Low-Quality Retinal Image Enhancement. J Imaging 2024; 10:151. [PMID: 39057722 PMCID: PMC11277667 DOI: 10.3390/jimaging10070151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2024] [Revised: 06/12/2024] [Accepted: 06/17/2024] [Indexed: 07/28/2024] Open
Abstract
Nonmydriatic retinal fundus images often suffer from quality issues and artifacts due to ocular or systemic comorbidities, leading to potential inaccuracies in clinical diagnoses. In recent times, deep learning methods have been widely employed to improve retinal image quality. However, these methods often require large datasets and lack robustness in clinical settings. Conversely, the inherent stability and adaptability of traditional unsupervised learning methods, coupled with their reduced reliance on extensive data, render them more suitable for real-world clinical applications, particularly in the limited data context of high noise levels or a significant presence of artifacts. However, existing unsupervised learning methods encounter challenges such as sensitivity to noise and outliers, reliance on assumptions like cluster shapes, and difficulties with scalability and interpretability, particularly when utilized for retinal image enhancement. To tackle these challenges, we propose a novel robust PCA (RPCA) method with low-rank sparse decomposition that also integrates affine transformations τi, weighted nuclear norm, and the L2,1 norms, aiming to overcome existing method limitations and to achieve image quality improvement unseen by these methods. We employ the weighted nuclear norm (Lw,∗) to assign weights to singular values to each retinal images and utilize the L2,1 norm to eliminate correlated samples and outliers in the retinal images. Moreover, τi is employed to enhance retinal image alignment, making the new method more robust to variations, outliers, noise, and image blurring. The Alternating Direction Method of Multipliers (ADMM) method is used to optimally determine parameters, including τi, by solving an optimization problem. Each parameter is addressed separately, harnessing the benefits of ADMM. Our method introduces a novel parameter update approach and significantly improves retinal image quality, detecting cataracts, and diabetic retinopathy. Simulation results confirm our method's superiority over existing state-of-the-art methods across various datasets.
Collapse
Affiliation(s)
- Habte Tadesse Likassa
- Department of Biostatistics, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
| | - Ding-Geng Chen
- Department of Biostatistics, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Department of Statistics, University of Pretoria, Pretoria 0028, South Africa
| | - Kewei Chen
- Department of Biostatistics, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
| | - Yalin Wang
- Computer Science and Engineering, School of Computing and Augmented Intelligence, Arizona State University, Phoenix, AZ 85287-8809, USA
| | - Wenhui Zhu
- Computer Science and Engineering, School of Computing and Augmented Intelligence, Arizona State University, Phoenix, AZ 85287-8809, USA
| |
Collapse
|
4
|
Habeb AAAA, Taresh MM, Li J, Gao Z, Zhu N. Enhancing Medical Image Classification with an Advanced Feature Selection Algorithm: A Novel Approach to Improving the Cuckoo Search Algorithm by Incorporating Caputo Fractional Order. Diagnostics (Basel) 2024; 14:1191. [PMID: 38893717 PMCID: PMC11172208 DOI: 10.3390/diagnostics14111191] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2024] [Revised: 05/23/2024] [Accepted: 05/29/2024] [Indexed: 06/21/2024] Open
Abstract
Glaucoma is a chronic eye condition that seriously impairs vision and requires early diagnosis and treatment. Automated detection techniques are essential for obtaining a timely diagnosis. In this paper, we propose a novel method for feature selection that integrates the cuckoo search algorithm with Caputo fractional order (CFO-CS) to enhance the performance of glaucoma classification. However, when using the infinite series, the Caputo definition has memory length truncation issues. Therefore, we suggest a fixed memory step and an adjustable term count for optimization. We conducted experiments integrating various feature extraction techniques, including histograms of oriented gradients (HOGs), local binary patterns (LBPs), and deep features from MobileNet and VGG19, to create a unified vector. We evaluate the informative features selected from the proposed method using the k-nearest neighbor. Furthermore, we use data augmentation to enhance the diversity and quantity of the training set. The proposed method enhances convergence speed and the attainment of optimal solutions during training. The results demonstrate superior performance on the test set, achieving 92.62% accuracy, 94.70% precision, 93.52% F1-Score, 92.98% specificity, 92.36% sensitivity, and 85.00% Matthew's correlation coefficient. The results confirm the efficiency of the proposed method, rendering it a generalizable and applicable technique in ophthalmology.
Collapse
Affiliation(s)
| | | | - Jintang Li
- College of Computer Science and Electronic Engineering, Hunan University, Changsha 410012, China; (A.A.A.A.H.); (J.L.); (N.Z.)
| | - Zhan Gao
- College of Computer Science and Electronic Engineering, Hunan University, Changsha 410012, China; (A.A.A.A.H.); (J.L.); (N.Z.)
| | - Ningbo Zhu
- College of Computer Science and Electronic Engineering, Hunan University, Changsha 410012, China; (A.A.A.A.H.); (J.L.); (N.Z.)
- Research Institute of Hunan University in Chongqing, Chongqing 400000, China
| |
Collapse
|
5
|
Rasel RK, Wu F, Chiariglione M, Choi SS, Doble N, Gao XR. Assessing the efficacy of 2D and 3D CNN algorithms in OCT-based glaucoma detection. Sci Rep 2024; 14:11758. [PMID: 38783015 PMCID: PMC11116516 DOI: 10.1038/s41598-024-62411-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Accepted: 05/16/2024] [Indexed: 05/25/2024] Open
Abstract
Glaucoma is a progressive neurodegenerative disease characterized by the gradual degeneration of retinal ganglion cells, leading to irreversible blindness worldwide. Therefore, timely and accurate diagnosis of glaucoma is crucial, enabling early intervention and facilitating effective disease management to mitigate further vision deterioration. The advent of optical coherence tomography (OCT) has marked a transformative era in ophthalmology, offering detailed visualization of the macula and optic nerve head (ONH) regions. In recent years, both 2D and 3D convolutional neural network (CNN) algorithms have been applied to OCT image analysis. While 2D CNNs rely on post-prediction aggregation of all B-scans within OCT volumes, 3D CNNs allow for direct glaucoma prediction from the OCT data. However, in the absence of extensively pre-trained 3D models, the comparative efficacy of 2D and 3D-CNN algorithms in detecting glaucoma from volumetric OCT images remains unclear. Therefore, this study explores the efficacy of glaucoma detection through volumetric OCT images using select state-of-the-art (SOTA) 2D-CNN models, 3D adaptations of these 2D-CNN models with specific weight transfer techniques, and a custom 5-layer 3D-CNN-Encoder algorithm. The performance across two distinct datasets is evaluated, each focusing on the macula and the ONH, to provide a comprehensive understanding of the models' capabilities in identifying glaucoma. Our findings demonstrate that the 2D-CNN algorithm consistently provided robust results compared to their 3D counterparts tested in this study for glaucoma detection, achieving AUC values of 0.960 and 0.943 for the macular and ONH OCT test images, respectively. Given the scarcity of pre-trained 3D models trained on extensive datasets, this comparative analysis underscores the overall utility of 2D and 3D-CNN algorithms in advancing glaucoma diagnostic systems in ophthalmology and highlights the potential of 2D algorithms for volumetric OCT image-based glaucoma detection.
Collapse
Affiliation(s)
- Rafiul Karim Rasel
- Department of Ophthalmology and Visual Sciences, The Ohio State University, Columbus, OH, 43212, USA
| | - Fengze Wu
- Department of Ophthalmology and Visual Sciences, The Ohio State University, Columbus, OH, 43212, USA
- Department of Biomedical Informatics, The Ohio State University, Columbus, OH, 43210, USA
| | - Marion Chiariglione
- Department of Ophthalmology and Visual Sciences, The Ohio State University, Columbus, OH, 43212, USA
| | - Stacey S Choi
- Department of Ophthalmology and Visual Sciences, The Ohio State University, Columbus, OH, 43212, USA
- College of Optometry, The Ohio State University, Columbus, OH, 43210, USA
| | - Nathan Doble
- Department of Ophthalmology and Visual Sciences, The Ohio State University, Columbus, OH, 43212, USA
- College of Optometry, The Ohio State University, Columbus, OH, 43210, USA
| | - Xiaoyi Raymond Gao
- Department of Ophthalmology and Visual Sciences, The Ohio State University, Columbus, OH, 43212, USA.
- Department of Biomedical Informatics, The Ohio State University, Columbus, OH, 43210, USA.
- Division of Human Genetics, The Ohio State University, Columbus, OH, 43210, USA.
| |
Collapse
|
6
|
Bowd C, Belghith A, Rezapour J, Christopher M, Jonas JB, Hyman L, Fazio MA, Weinreb RN, Zangwill LM. Multimodal Deep Learning Classifier for Primary Open Angle Glaucoma Diagnosis Using Wide-Field Optic Nerve Head Cube Scans in Eyes With and Without High Myopia. J Glaucoma 2023; 32:841-847. [PMID: 37523623 DOI: 10.1097/ijg.0000000000002267] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2023] [Accepted: 06/18/2023] [Indexed: 08/02/2023]
Abstract
PRCIS An optical coherence tomography (OCT)-based multimodal deep learning (DL) classification model, including texture information, is introduced that outperforms single-modal models and multimodal models without texture information for glaucoma diagnosis in eyes with and without high myopia. BACKGROUND/AIMS To evaluate the diagnostic accuracy of a multimodal DL classifier using wide OCT optic nerve head cube scans in eyes with and without axial high myopia. MATERIALS AND METHODS Three hundred seventy-one primary open angle glaucoma (POAG) eyes and 86 healthy eyes, all without axial high myopia [axial length (AL) ≤ 26 mm] and 92 POAG eyes and 44 healthy eyes, all with axial high myopia (AL > 26 mm) were included. The multimodal DL classifier combined features of 3 individual VGG-16 models: (1) texture-based en face image, (2) retinal nerve fiber layer (RNFL) thickness map image, and (3) confocal scanning laser ophthalmoscope (cSLO) image. Age, AL, and disc area adjusted area under the receiver operating curves were used to compare model accuracy. RESULTS Adjusted area under the receiver operating curve for the multimodal DL model was 0.91 (95% CI = 0.87, 0.95). This value was significantly higher than the values of individual models [0.83 (0.79, 0.86) for texture-based en face image; 0.84 (0.81, 0.87) for RNFL thickness map; and 0.68 (0.61, 0.74) for cSLO image; all P ≤ 0.05]. Using only highly myopic eyes, the multimodal DL model showed significantly higher diagnostic accuracy [0.89 (0.86, 0.92)] compared with texture en face image [0.83 (0.78, 0.85)], RNFL [0.85 (0.81, 0.86)] and cSLO image models [0.69 (0.63, 0.76)] (all P ≤ 0.05). CONCLUSIONS Combining OCT-based RNFL thickness maps with texture-based en face images showed a better ability to discriminate between healthy and POAG than thickness maps alone, particularly in high axial myopic eyes.
Collapse
Affiliation(s)
- Christopher Bowd
- Viterbi Family Department of Ophthalmology, Shiley Eye Institute, Hamilton Glaucoma Center
| | - Akram Belghith
- Viterbi Family Department of Ophthalmology, Shiley Eye Institute, Hamilton Glaucoma Center
| | - Jasmin Rezapour
- Viterbi Family Department of Ophthalmology, Shiley Eye Institute, Hamilton Glaucoma Center
- Department of Ophthalmology, University Medical Center of the Johannes Gutenberg University Mainz
| | - Mark Christopher
- Viterbi Family Department of Ophthalmology, Shiley Eye Institute, Hamilton Glaucoma Center
| | - Jost B Jonas
- Department of Ophthalmology, Heidelberg University, Mannheim, Germany
| | - Leslie Hyman
- Vickie and Jack Farber Vision Research Center, Wills Eye Hospital, Thomas Jefferson University, Philadelphia, PA
| | - Massimo A Fazio
- Department of Ophthalmology and Visual Sciences, The University of Alabama at Birmingham, Birmingham, AL
| | - Robert N Weinreb
- Viterbi Family Department of Ophthalmology, Shiley Eye Institute, Hamilton Glaucoma Center
| | - Linda M Zangwill
- Viterbi Family Department of Ophthalmology, Shiley Eye Institute, Hamilton Glaucoma Center
| |
Collapse
|
7
|
Valente J, António J, Mora C, Jardim S. Developments in Image Processing Using Deep Learning and Reinforcement Learning. J Imaging 2023; 9:207. [PMID: 37888314 PMCID: PMC10607786 DOI: 10.3390/jimaging9100207] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 09/24/2023] [Accepted: 09/28/2023] [Indexed: 10/28/2023] Open
Abstract
The growth in the volume of data generated, consumed, and stored, which is estimated to exceed 180 zettabytes in 2025, represents a major challenge both for organizations and for society in general. In addition to being larger, datasets are increasingly complex, bringing new theoretical and computational challenges. Alongside this evolution, data science tools have exploded in popularity over the past two decades due to their myriad of applications when dealing with complex data, their high accuracy, flexible customization, and excellent adaptability. When it comes to images, data analysis presents additional challenges because as the quality of an image increases, which is desirable, so does the volume of data to be processed. Although classic machine learning (ML) techniques are still widely used in different research fields and industries, there has been great interest from the scientific community in the development of new artificial intelligence (AI) techniques. The resurgence of neural networks has boosted remarkable advances in areas such as the understanding and processing of images. In this study, we conducted a comprehensive survey regarding advances in AI design and the optimization solutions proposed to deal with image processing challenges. Despite the good results that have been achieved, there are still many challenges to face in this field of study. In this work, we discuss the main and more recent improvements, applications, and developments when targeting image processing applications, and we propose future research directions in this field of constant and fast evolution.
Collapse
Affiliation(s)
- Jorge Valente
- Techframe-Information Systems, SA, 2785-338 São Domingos de Rana, Portugal; (J.V.); (J.A.)
| | - João António
- Techframe-Information Systems, SA, 2785-338 São Domingos de Rana, Portugal; (J.V.); (J.A.)
| | - Carlos Mora
- Smart Cities Research Center, Polytechnic Institute of Tomar, 2300-313 Tomar, Portugal;
| | - Sandra Jardim
- Smart Cities Research Center, Polytechnic Institute of Tomar, 2300-313 Tomar, Portugal;
| |
Collapse
|
8
|
Sangchocanonta S, Ingpochai S, Puangarom S, Munthuli A, Phienphanich P, Itthipanichpong R, Chansangpetch S, Manassakorn A, Ratanawongphaibul K, Tantisevi V, Rojanapongpun P, Tantibundhit C. Donut: Augmentation Technique for Enhancing The Efficacy of Glaucoma Suspect Screening. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2023; 2023:1-5. [PMID: 38083547 DOI: 10.1109/embc40787.2023.10341115] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2023]
Abstract
Glaucoma is the second most common cause of blindness. A glaucoma suspect has risk factors that increase the possibility of developing glaucoma. Evaluating a patient with suspected glaucoma is challenging. The "donut method" was developed in this study as an augmentation technique for obtaining high-quality fundus images for training ConvNeXt-Small model. Fundus images from GlauCUTU-DATA, labelled by randomizing at least 3 well-trained ophthalmologists (4 well-trained ophthalmologists in case of no majority agreement) with a unanimous agreement (3/3) and majority agreement (2/3), were used in the experiment. The experimental results from the proposed method showed the training model with the "donut method" increased the sensitivity of glaucoma suspects from 52.94% to 70.59% for the 3/3 data and increased the sensitivity of glaucoma suspects from 37.78% to 42.22% for the 2/3 data. This method enhanced the efficacy of classifying glaucoma suspects in both equalizing sensitivity and specificity sufficiently. Furthermore, three well-trained ophthalmologists agreed that the GradCAM++ heatmaps obtained from the training model using the proposed method highlighted the clinical criteria.Clinical relevance- The donut method for augmentation fundus images focuses on the optic nerve head region for enhancing efficacy of glaucoma suspect screening, and uses Grad-CAM++ to highlight the clinical criteria.
Collapse
|
9
|
Aldughayfiq B, Ashfaq F, Jhanjhi NZ, Humayun M. Explainable AI for Retinoblastoma Diagnosis: Interpreting Deep Learning Models with LIME and SHAP. Diagnostics (Basel) 2023; 13:1932. [PMID: 37296784 PMCID: PMC10253103 DOI: 10.3390/diagnostics13111932] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2023] [Revised: 05/19/2023] [Accepted: 05/26/2023] [Indexed: 06/12/2023] Open
Abstract
Retinoblastoma is a rare and aggressive form of childhood eye cancer that requires prompt diagnosis and treatment to prevent vision loss and even death. Deep learning models have shown promising results in detecting retinoblastoma from fundus images, but their decision-making process is often considered a "black box" that lacks transparency and interpretability. In this project, we explore the use of LIME and SHAP, two popular explainable AI techniques, to generate local and global explanations for a deep learning model based on InceptionV3 architecture trained on retinoblastoma and non-retinoblastoma fundus images. We collected and labeled a dataset of 400 retinoblastoma and 400 non-retinoblastoma images, split it into training, validation, and test sets, and trained the model using transfer learning from the pre-trained InceptionV3 model. We then applied LIME and SHAP to generate explanations for the model's predictions on the validation and test sets. Our results demonstrate that LIME and SHAP can effectively identify the regions and features in the input images that contribute the most to the model's predictions, providing valuable insights into the decision-making process of the deep learning model. In addition, the use of InceptionV3 architecture with spatial attention mechanism achieved high accuracy of 97% on the test set, indicating the potential of combining deep learning and explainable AI for improving retinoblastoma diagnosis and treatment.
Collapse
Affiliation(s)
- Bader Aldughayfiq
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia;
| | - Farzeen Ashfaq
- School of Computer Science, SCS, Taylor’s University, Subang Jaya 47500, Malaysia; (F.A.); (N.Z.J.)
| | - N. Z. Jhanjhi
- School of Computer Science, SCS, Taylor’s University, Subang Jaya 47500, Malaysia; (F.A.); (N.Z.J.)
| | - Mamoona Humayun
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia;
| |
Collapse
|
10
|
Bayhaqi YA, Hamidi A, Navarini AA, Cattin PC, Canbaz F, Zam A. Real-time closed-loop tissue-specific laser osteotomy using deep-learning-assisted optical coherence tomography. BIOMEDICAL OPTICS EXPRESS 2023; 14:2986-3002. [PMID: 37342720 PMCID: PMC10278623 DOI: 10.1364/boe.486660] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/30/2023] [Revised: 05/15/2023] [Accepted: 05/18/2023] [Indexed: 06/23/2023]
Abstract
This article presents a real-time noninvasive method for detecting bone and bone marrow in laser osteotomy. This is the first optical coherence tomography (OCT) implementation as an online feedback system for laser osteotomy. A deep-learning model has been trained to identify tissue types during laser ablation with a test accuracy of 96.28 %. For the hole ablation experiments, the average maximum depth of perforation and volume loss was 0.216 mm and 0.077 mm3, respectively. The contactless nature of OCT with the reported performance shows that it is becoming more feasible to utilize it as a real-time feedback system for laser osteotomy.
Collapse
Affiliation(s)
- Yakub. A. Bayhaqi
- Biomedical Laser and Optics Group (BLOG), Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland
| | - Arsham Hamidi
- Biomedical Laser and Optics Group (BLOG), Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland
| | - Alexander A. Navarini
- Digital Dermatology Group, Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland
| | - Philippe C. Cattin
- Center for medical Image Analysis and Navigation (CIAN), Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland
| | - Ferda Canbaz
- Biomedical Laser and Optics Group (BLOG), Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland
| | - Azhar Zam
- Biomedical Laser and Optics Group (BLOG), Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland
- Division of Engineering, New York University Abu Dhabi, Abu Dhabi, 129188, United Arab Emirates
- Tandon School of Engineering, New York University, Brooklyn, NY, 11201, USA
| |
Collapse
|
11
|
Rayavel P, Murukesh C. Comparative analysis of deep learning classifiers for diabetic retinopathy identification and detection. THE IMAGING SCIENCE JOURNAL 2023. [DOI: 10.1080/13682199.2023.2168851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/18/2023]
Affiliation(s)
- P. Rayavel
- Department of Computer Science and Engineering (Cybersecurity), Sri Sairam Institute of Technology, Chennai, Tamil Nadu, India
| | - C. Murukesh
- Department of Electronics and Communication Engineering, Velammal Engineering College, Chennai, Tamil Nadu, India
| |
Collapse
|
12
|
Agboola HA, Zaccheus JE. Wavelet image scattering based glaucoma detection. BMC Biomed Eng 2023; 5:1. [PMID: 36864533 PMCID: PMC9979468 DOI: 10.1186/s42490-023-00067-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2022] [Accepted: 02/06/2023] [Indexed: 03/04/2023] Open
Abstract
BACKGROUND The ever-growing need for cheap, simple, fast, and accurate healthcare solutions spurred a lot of research activities which are aimed at the reliable deployment of artificial intelligence in the medical fields. However, this has proved to be a daunting task especially when looking to make automated diagnoses using biomedical image data. Biomedical image data have complex patterns which human experts find very hard to comprehend. Against this backdrop, we applied a representation or feature learning algorithm: Invariant Scattering Convolution Network or Wavelet scattering Network to retinal fundus images and studied the the efficacy of the automatically extracted features therefrom for glaucoma diagnosis/detection. The influence of wavelet scattering network parameter settings as well as 2-D channel image type on the detection correctness is also examined. Our work is a distinct departure from the usual method where wavelet transform is applied to pre-processed retinal fundus images and handcrafted features are extracted from the decomposition results. Here, the RIM-ONE DL image dataset was fed into a wavelet scattering network developed in the Matlab environment to achieve a stage-wise decomposition process called wavelet scattering of the retinal fundus images thereby, automatically learning features from the images. These features were then used to build simple and computationally cheap classification algorithms. RESULTS Maximum detection correctness of 98% was achieved on the held-out test set. Detection correctness is highly sensitive to scattering network parameter setting and 2-D channel image type. CONCLUSION A superficial comparison of the classification results obtained from our work and those obtained using a convolutional neural network underscores the potentiality of the proposed method for glaucoma detection.
Collapse
|
13
|
Zhang S, Webers CAB, Berendschot TTJM. Luminosity rectified blind Richardson-Lucy deconvolution for single retinal image restoration. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 229:107297. [PMID: 36563648 DOI: 10.1016/j.cmpb.2022.107297] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/11/2022] [Revised: 11/14/2022] [Accepted: 11/29/2022] [Indexed: 06/17/2023]
Abstract
BACKGROUND AND OBJECTIVE Due to imperfect imaging conditions, retinal images can be degraded by uneven/insufficient illumination, blurriness caused by optical aberrations and unintentional motions. Degraded images reduce the effectiveness of diagnosis by an ophthalmologist. To restore the image quality, in this research we propose the luminosity rectified Richardson-Lucy (LRRL) blind deconvolution framework for single retinal image restoration. METHODS We established an image formation model based on the double-pass fundus reflection feature and developed a differentiable non-convex cost function that jointly achieves illumination correction and blind deconvolution. To solve this non-convex optimization problem, we derived the closed-form expression of the gradients and used gradient descent with Nesterov-accelerated adaptive momentum estimation to accelerate the optimization, which is more efficient than the traditional half quadratic splitting method. RESULTS The LRRL was tested on 1719 images from three public databases. Four image quality matrixes including image definition, image sharpness, image entropy, and image multiscale contrast were used for objective assessments. The LRRL was compared against the state-of-the-art retinal image blind deconvolution methods. CONCLUSIONS Our LRRL corrects the problematic illumination and improves the clarity of the retinal image simultaneously, showing its superiority in terms of restoration quality and implementation efficiency. The MATLAB code is available on Github.
Collapse
Affiliation(s)
- Shuhe Zhang
- University Eye Clinic Maastricht, Maastricht University Medical Center +, P.O. Box 5800, Maastricht, AZ 6202, the Netherlands.
| | - Carroll A B Webers
- University Eye Clinic Maastricht, Maastricht University Medical Center +, P.O. Box 5800, Maastricht, AZ 6202, the Netherlands
| | - Tos T J M Berendschot
- University Eye Clinic Maastricht, Maastricht University Medical Center +, P.O. Box 5800, Maastricht, AZ 6202, the Netherlands
| |
Collapse
|
14
|
Glaucoma Detection and Classification Using Improved U-Net Deep Learning Model. Healthcare (Basel) 2022; 10:healthcare10122497. [PMID: 36554021 PMCID: PMC9778546 DOI: 10.3390/healthcare10122497] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Revised: 11/19/2022] [Accepted: 12/05/2022] [Indexed: 12/14/2022] Open
Abstract
Glaucoma is prominent in a variety of nations, with the United States and Europe being two of the most famous. Glaucoma now affects around 78 million people throughout the world (2020). By the year 2040, it is expected that there will be 111.8 million cases of glaucoma worldwide. In countries that are still building enough healthcare infrastructure to cope with glaucoma, the ailment is misdiagnosed nine times out of ten. To aid in the early diagnosis of glaucoma, the creation of a detection system is necessary. In this work, the researchers propose using a technology known as deep learning to identify and predict glaucoma before symptoms appear. The glaucoma dataset is used in this deep learning algorithm that has been proposed for analyzing glaucoma images. To get the required results when using deep learning principles for the job of segmenting the optic cup, pretrained transfer learning models are integrated with the U-Net architecture. For feature extraction, the DenseNet-201 deep convolution neural network (DCNN) is used. The DCNN approach is used to determine whether a person has glaucoma. The fundamental goal of this line of research is to recognize glaucoma in retinal fundus images, which will aid in assessing whether a patient has the condition. Because glaucoma can affect the model in both positive and negative ways, the model's outcome might be either positive or negative. Accuracy, precision, recall, specificity, the F-measure, and the F-score are some of the metrics used in the model evaluation process. An extra comparison study is performed as part of the process of establishing whether the suggested model is accurate. The findings are compared to convolution neural network classification methods based on deep learning. When used for training, the suggested model has an accuracy of 98.82 percent and an accuracy of 96.90 percent when used for testing. All assessments show that the new paradigm that has been proposed is more successful than the one that is currently in use.
Collapse
|
15
|
Ong J, Tan G, Ang M, Chhablani J. Digital Advancements in Retinal Models of Care in the Post-COVID-19 Lockdown Era. Asia Pac J Ophthalmol (Phila) 2022; 11:403-407. [PMID: 36094383 DOI: 10.1097/apo.0000000000000533] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2021] [Accepted: 03/14/2022] [Indexed: 11/25/2022] Open
Abstract
The coronavirus disease-2019 (COVID-19) pandemic introduced unique barriers to retinal care including limited access to imaging modalities, ophthalmic clinicians, and direct medical interventions. These unprecedented barriers were met with the robust implementation of digital advances to aid in monitoring and efficiency of retinal care while taking into the account of public safety. Many of these innovations have been successful in maintaining efficiency and patient satisfaction and are likely to stay to help preserve vision in the future. In this article we highlight these advances implemented during the pandemic including telescreening triage, virtual retinal imaging clinics, at-home optical coherence tomography, mobile phone self-monitoring, and virtual reality monitoring technology. We also discuss advancing innovations including Internet of Things and Blockchain technology that will be critical for further implementation and security of these digital advancements.
Collapse
Affiliation(s)
- Joshua Ong
- Department of Ophthalmology, University of Pittsburgh School of Medicine, Pittsburgh, PA
| | - Gavin Tan
- Surgical Retinal Department of the Singapore National Eye Centre, Singapore
- Clinician Scientist, Singapore Eye Research Institute, Singapore
- Duke-NUS Medical School, National University of Singapore, Singapore
| | - Marcus Ang
- Duke-NUS Department of Ophthalmology and Visual Sciences, Singapore
| | - Jay Chhablani
- Department of Ophthalmology, University of Pittsburgh School of Medicine, Pittsburgh, PA
| |
Collapse
|
16
|
Alawad M, Aljouie A, Alamri S, Alghamdi M, Alabdulkader B, Alkanhal N, Almazroa A. Machine Learning and Deep Learning Techniques for Optic Disc and Cup Segmentation - A Review. Clin Ophthalmol 2022; 16:747-764. [PMID: 35300031 PMCID: PMC8923700 DOI: 10.2147/opth.s348479] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2021] [Accepted: 02/11/2022] [Indexed: 12/12/2022] Open
Abstract
Background Globally, glaucoma is the second leading cause of blindness. Detecting glaucoma in the early stages is essential to avoid disease complications, which lead to blindness. Thus, computer-aided diagnosis systems are powerful tools to overcome the shortage of glaucoma screening programs. Methods A systematic search of public databases, including PubMed, Google Scholar, and other sources, was performed to identify relevant studies to overview the publicly available fundus image datasets used to train, validate, and test machine learning and deep learning methods. Additionally, existing machine learning and deep learning methods for optic cup and disc segmentation were surveyed and critically reviewed. Results Eight fundus images datasets were publicly available with 15,445 images labeled with glaucoma or non-glaucoma, and manually annotated optic disc and cup boundaries were found. Five metrics were identified for evaluating the developed models. Finally, three main deep learning architectural designs were commonly used for optic disc and optic cup segmentation. Conclusion We provided future research directions to formulate robust optic cup and disc segmentation systems. Deep learning can be utilized in clinical settings for this task. However, many challenges need to be addressed before using this strategy in clinical trials. Finally, two deep learning architectural designs have been widely adopted, such as U-net and its variants.
Collapse
Affiliation(s)
- Mohammed Alawad
- Department of Biostatistics and Bioinformatics, King Abdullah International Medical Research Center, King Saud bin Abdulaziz University for Health Sciences, Riyadh, Saudi Arabia
| | - Abdulrhman Aljouie
- Department of Biostatistics and Bioinformatics, King Abdullah International Medical Research Center, King Saud bin Abdulaziz University for Health Sciences, Riyadh, Saudi Arabia
| | - Suhailah Alamri
- Department of Imaging Research, King Abdullah International Medical Research Center, King Saud bin Abdulaziz University for health Sciences, Riyadh, Saudi Arabia
- Research Labs, National Center for Artificial Intelligence, Riyadh, Saudi Arabia
| | - Mansour Alghamdi
- Department of Optometry and Vision Sciences College of Applied Medical Sciences, King Saud University, Riyadh, Saudi Arabia
| | - Balsam Alabdulkader
- Department of Optometry and Vision Sciences College of Applied Medical Sciences, King Saud University, Riyadh, Saudi Arabia
| | - Norah Alkanhal
- Department of Imaging Research, King Abdullah International Medical Research Center, King Saud bin Abdulaziz University for health Sciences, Riyadh, Saudi Arabia
| | - Ahmed Almazroa
- Department of Imaging Research, King Abdullah International Medical Research Center, King Saud bin Abdulaziz University for health Sciences, Riyadh, Saudi Arabia
| |
Collapse
|
17
|
Segmentation and Classification of Glaucoma Using U-Net with Deep Learning Model. JOURNAL OF HEALTHCARE ENGINEERING 2022; 2022:1601354. [PMID: 35222876 PMCID: PMC8866016 DOI: 10.1155/2022/1601354] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/02/2021] [Revised: 01/07/2022] [Accepted: 01/12/2022] [Indexed: 11/17/2022]
Abstract
Glaucoma is the second most common cause for blindness around the world and the third most common in Europe and the USA. Around 78 million people are presently living with glaucoma (2020). It is expected that 111.8 million people will have glaucoma by the year 2040. 90% of glaucoma is undetected in developing nations. It is essential to develop a glaucoma detection system for early diagnosis. In this research, early prediction of glaucoma using deep learning technique is proposed. In this proposed deep learning model, the ORIGA dataset is used for the evaluation of glaucoma images. The U-Net architecture based on deep learning algorithm is implemented for optic cup segmentation and a pretrained transfer learning model; DenseNet-201 is used for feature extraction along with deep convolution neural network (DCNN). The DCNN approach is used for the classification, where the final results will be representing whether the glaucoma infected or not. The primary objective of this research is to detect the glaucoma using the retinal fundus images, which can be useful to determine if the patient was affected by glaucoma or not. The result of this model can be positive or negative based on the outcome detected as infected by glaucoma or not. The model is evaluated using parameters such as accuracy, precision, recall, specificity, and F-measure. Also, a comparative analysis is conducted for the validation of the model proposed. The output is compared to other current deep learning models used for CNN classification, such as VGG-19, Inception ResNet, ResNet 152v2, and DenseNet-169. The proposed model achieved 98.82% accuracy in training and 96.90% in testing. Overall, the performance of the proposed model is better in all the analysis.
Collapse
|
18
|
Neto A, Camara J, Cunha A. Evaluations of Deep Learning Approaches for Glaucoma Screening Using Retinal Images from Mobile Device. SENSORS 2022; 22:s22041449. [PMID: 35214351 PMCID: PMC8874723 DOI: 10.3390/s22041449] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/29/2021] [Revised: 02/09/2022] [Accepted: 02/10/2022] [Indexed: 02/04/2023]
Abstract
Glaucoma is a silent disease that leads to vision loss or irreversible blindness. Current deep learning methods can help glaucoma screening by extending it to larger populations using retinal images. Low-cost lenses attached to mobile devices can increase the frequency of screening and alert patients earlier for a more thorough evaluation. This work explored and compared the performance of classification and segmentation methods for glaucoma screening with retinal images acquired by both retinography and mobile devices. The goal was to verify the results of these methods and see if similar results could be achieved using images captured by mobile devices. The used classification methods were the Xception, ResNet152 V2 and the Inception ResNet V2 models. The models’ activation maps were produced and analysed to support glaucoma classifier predictions. In clinical practice, glaucoma assessment is commonly based on the cup-to-disc ratio (CDR) criterion, a frequent indicator used by specialists. For this reason, additionally, the U-Net architecture was used with the Inception ResNet V2 and Inception V3 models as the backbone to segment and estimate CDR. For both tasks, the performance of the models reached close to that of state-of-the-art methods, and the classification method applied to a low-quality private dataset illustrates the advantage of using cheaper lenses.
Collapse
Affiliation(s)
- Alexandre Neto
- Escola de Ciências de Tecnologia, University of Trás-os-Montes and Alto Douro, Quinta de Prados, 5001-801 Vila Real, Portugal;
- INESC TEC—Institute for Systems and Computer Engineering, Technology and Science, 4200-465 Porto, Portugal;
| | - José Camara
- INESC TEC—Institute for Systems and Computer Engineering, Technology and Science, 4200-465 Porto, Portugal;
- Departamento de Ciências e Tecnologia, University Aberta, 1250-100 Lisboa, Portugal
| | - António Cunha
- Escola de Ciências de Tecnologia, University of Trás-os-Montes and Alto Douro, Quinta de Prados, 5001-801 Vila Real, Portugal;
- INESC TEC—Institute for Systems and Computer Engineering, Technology and Science, 4200-465 Porto, Portugal;
- Correspondence: ; Tel.: +351-931-636-373
| |
Collapse
|
19
|
Vinayaki VD, Kalaiselvi R. Multithreshold Image Segmentation Technique Using Remora Optimization Algorithm for Diabetic Retinopathy Detection from Fundus Images. Neural Process Lett 2022; 54:2363-2384. [PMID: 35095328 PMCID: PMC8784591 DOI: 10.1007/s11063-021-10734-0] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/24/2021] [Indexed: 12/21/2022]
Abstract
One of the most common complications of diabetes mellitus is diabetic retinopathy (DR), which produces lesions on the retina. A novel framework for DR detection and classification was proposed in this study. The proposed work includes four stages: pre-processing, segmentation, feature extraction, and classification. Initially, the image pre-processing is performed and after that, the Multi threshold-based Remora Optimization (MTRO) algorithm performs the vessel segmentation. The feature extraction and classification process are done by using a Region-based Convolution Neural Network (R-CNN) with Wild Geese Algorithm (WGA). Finally, the proposed R-CNN with WGA effectively classifies the different stages of DR including Non-DR, Proliferative DR, Severe, Moderate DR, Mild DR. The experimental images were collected from the DRIVE database, and the proposed framework exhibited superior DR detection performance. Compared to other existing methods like fully convolutional deep neural network (FCDNN), genetic-search feature selection (GSFS), Convolutional Neural Networks (CNN), and deep learning (DL) techniques, the proposed R-CNN with WGA provided 95.42% accuracy, 93.10% specificity, 93.20% sensitivity, and 98.28% F-score results.
Collapse
Affiliation(s)
- V. Desika Vinayaki
- Department of Computer Science and Engineering, Noorul Islam Centre for Higher Education, Kumaracoil, India
| | - R. Kalaiselvi
- Department of Computer Science and Engineering, Noorul Islam Centre for Higher Education, Kumaracoil, India
| |
Collapse
|
20
|
Review of Machine Learning Applications Using Retinal Fundus Images. Diagnostics (Basel) 2022; 12:diagnostics12010134. [PMID: 35054301 PMCID: PMC8774893 DOI: 10.3390/diagnostics12010134] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Revised: 01/03/2022] [Accepted: 01/03/2022] [Indexed: 02/04/2023] Open
Abstract
Automating screening and diagnosis in the medical field saves time and reduces the chances of misdiagnosis while saving on labor and cost for physicians. With the feasibility and development of deep learning methods, machines are now able to interpret complex features in medical data, which leads to rapid advancements in automation. Such efforts have been made in ophthalmology to analyze retinal images and build frameworks based on analysis for the identification of retinopathy and the assessment of its severity. This paper reviews recent state-of-the-art works utilizing the color fundus image taken from one of the imaging modalities used in ophthalmology. Specifically, the deep learning methods of automated screening and diagnosis for diabetic retinopathy (DR), age-related macular degeneration (AMD), and glaucoma are investigated. In addition, the machine learning techniques applied to the retinal vasculature extraction from the fundus image are covered. The challenges in developing these systems are also discussed.
Collapse
|
21
|
Saeed AQ, Sheikh Abdullah SNH, Che-Hamzah J, Abdul Ghani AT. Accuracy of Using Generative Adversarial Networks for Glaucoma Detection During the COVID-19 Pandemic: A Systematic Review and Bibliometric Analysis. J Med Internet Res 2021; 23:e27414. [PMID: 34236992 PMCID: PMC8493455 DOI: 10.2196/27414] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2021] [Revised: 05/11/2021] [Accepted: 07/05/2021] [Indexed: 01/19/2023] Open
Abstract
Background Glaucoma leads to irreversible blindness. Globally, it is the second most common retinal disease that leads to blindness, slightly less common than cataracts. Therefore, there is a great need to avoid the silent growth of this disease using recently developed generative adversarial networks (GANs). Objective This paper aims to introduce a GAN technology for the diagnosis of eye disorders, particularly glaucoma. This paper illustrates deep adversarial learning as a potential diagnostic tool and the challenges involved in its implementation. This study describes and analyzes many of the pitfalls and problems that researchers will need to overcome to implement this kind of technology. Methods To organize this review comprehensively, articles and reviews were collected using the following keywords: (“Glaucoma,” “optic disc,” “blood vessels”) and (“receptive field,” “loss function,” “GAN,” “Generative Adversarial Network,” “Deep learning,” “CNN,” “convolutional neural network” OR encoder). The records were identified from 5 highly reputed databases: IEEE Xplore, Web of Science, Scopus, ScienceDirect, and PubMed. These libraries broadly cover the technical and medical literature. Publications within the last 5 years, specifically 2015-2020, were included because the target GAN technique was invented only in 2014 and the publishing date of the collected papers was not earlier than 2016. Duplicate records were removed, and irrelevant titles and abstracts were excluded. In addition, we excluded papers that used optical coherence tomography and visual field images, except for those with 2D images. A large-scale systematic analysis was performed, and then a summarized taxonomy was generated. Furthermore, the results of the collected articles were summarized and a visual representation of the results was presented on a T-shaped matrix diagram. This study was conducted between March 2020 and November 2020. Results We found 59 articles after conducting a comprehensive survey of the literature. Among the 59 articles, 30 present actual attempts to synthesize images and provide accurate segmentation/classification using single/multiple landmarks or share certain experiences. The other 29 articles discuss the recent advances in GANs, do practical experiments, and contain analytical studies of retinal disease. Conclusions Recent deep learning techniques, namely GANs, have shown encouraging performance in retinal disease detection. Although this methodology involves an extensive computing budget and optimization process, it saturates the greedy nature of deep learning techniques by synthesizing images and solves major medical issues. This paper contributes to this research field by offering a thorough analysis of existing works, highlighting current limitations, and suggesting alternatives to support other researchers and participants in further improving and strengthening future work. Finally, new directions for this research have been identified.
Collapse
Affiliation(s)
- Ali Q Saeed
- Faculty of Information Science & Technology (FTSM), Universiti Kebangsaan Malaysia (UKM), UKM, 43600 Bangi, Selangor, Malaysia, Selangor, MY.,Computer Center, Northern Technical University, Ninevah, IQ
| | - Siti Norul Huda Sheikh Abdullah
- Faculty of Information Science & Technology (FTSM), Universiti Kebangsaan Malaysia (UKM), UKM, 43600 Bangi, Selangor, Malaysia, Selangor, MY
| | - Jemaima Che-Hamzah
- Department of Ophthalmology, Faculty of Medicine, Universiti Kebangsaan Malaysia (UKM), Cheras, Kuala Lumpur, MY
| | - Ahmad Tarmizi Abdul Ghani
- Faculty of Information Science & Technology (FTSM), Universiti Kebangsaan Malaysia (UKM), UKM, 43600 Bangi, Selangor, Malaysia, Selangor, MY
| |
Collapse
|
22
|
Oh S, Park Y, Cho KJ, Kim SJ. Explainable Machine Learning Model for Glaucoma Diagnosis and Its Interpretation. Diagnostics (Basel) 2021; 11:510. [PMID: 33805685 PMCID: PMC8001225 DOI: 10.3390/diagnostics11030510] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2021] [Revised: 03/08/2021] [Accepted: 03/10/2021] [Indexed: 12/14/2022] Open
Abstract
The aim is to develop a machine learning prediction model for the diagnosis of glaucoma and an explanation system for a specific prediction. Clinical data of the patients based on a visual field test, a retinal nerve fiber layer optical coherence tomography (RNFL OCT) test, a general examination including an intraocular pressure (IOP) measurement, and fundus photography were provided for the feature selection process. Five selected features (variables) were used to develop a machine learning prediction model. The support vector machine, C5.0, random forest, and XGboost algorithms were tested for the prediction model. The performance of the prediction models was tested with 10-fold cross-validation. Statistical charts, such as gauge, radar, and Shapley Additive Explanations (SHAP), were used to explain the prediction case. All four models achieved similarly high diagnostic performance, with accuracy values ranging from 0.903 to 0.947. The XGboost model is the best model with an accuracy of 0.947, sensitivity of 0.941, specificity of 0.950, and AUC of 0.945. Three statistical charts were established to explain the prediction based on the characteristics of the XGboost model. Higher diagnostic performance was achieved with the XGboost model. These three statistical charts can help us understand why the machine learning model produces a specific prediction result. This may be the first attempt to apply "explainable artificial intelligence" to eye disease diagnosis.
Collapse
Affiliation(s)
- Sejong Oh
- Software Science, College of Software Convergence, Jukjeon Campus, Dankook University, Yongin 16890, Korea;
| | - Yuli Park
- Department of Ophthalmology, College of Medicine, Dankook University, 119, Dandae-ro, Dongnam-gu, Cheonan-si, Chungnam 31116, Korea; (Y.P.); (K.J.C.)
| | - Kyong Jin Cho
- Department of Ophthalmology, College of Medicine, Dankook University, 119, Dandae-ro, Dongnam-gu, Cheonan-si, Chungnam 31116, Korea; (Y.P.); (K.J.C.)
| | - Seong Jae Kim
- Department of Ophthalmology, Institute of Health Sciences, Gyeongsang National University School of Medicine and Gyeongsang National University Hospital, Jinju 52727, Korea
| |
Collapse
|
23
|
Chai Y, Bian Y, Liu H, Li J, Xu J. Glaucoma diagnosis in the Chinese context: An uncertainty information-centric Bayesian deep learning model. Inf Process Manag 2021. [DOI: 10.1016/j.ipm.2020.102454] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|