1
|
Farahat Z, Zrira N, Souissi N, Bennani Y, Bencherif S, Benamar S, Belmekki M, Ngote MN, Megdiche K. Diabetic retinopathy screening through artificial intelligence algorithms: A systematic review. Surv Ophthalmol 2024:S0039-6257(24)00051-1. [PMID: 38885761 DOI: 10.1016/j.survophthal.2024.05.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2023] [Revised: 05/20/2024] [Accepted: 05/20/2024] [Indexed: 06/20/2024]
Abstract
Diabetic retinopathy (DR) poses a significant challenge in diabetes management, with its progression often asymptomatic until advanced stages. This underscores the urgent need for cost-effective and reliable screening methods. Consequently, the integration of artificial intelligence (AI) tools presents a promising avenue to address this need effectively. We provide an overview of the current state of the art results and techniques in DR screening using AI, while also identifying gaps in research for future exploration. By synthesizing existing database and pinpointing areas requiring further investigation, this paper seeks to guide the direction of future research in the field of automatic diabetic retinopathy screening. There has been a continuous rise in the number of articles detailing deep learning (DL) methods designed for the automatic screening of diabetic retinopathy especially by the year 2021. Researchers utilized various databases, with a primary focus on the IDRiD dataset. This dataset consists of color fundus images captured at an ophthalmological clinic situated in India. It comprises 516 images that depict various stages of DR and diabetic macular edema. Each of the chosen papers concentrates on various DR signs. Nevertheless, a significant portion primarily focused on detecting exudates, which remains insufficient to assess the overall presence of this disease. Various AI methods have been employed to identify DR signs. Among the chosen papers, 4.7 % utilized detection methods, 46.5 % employed classification techniques, 41.9 % relied on segmentation, and 7 % opted for a combination of classification and segmentation. Metrics calculated from 80 % of the articles employing preprocessing techniques demonstrated the significant benefits of this approach in enhancing results quality. In addition, multiple DL techniques, starting by classification, detection then segmentation. Researchers used mostly YOLO for detection, ViT for classification, and U-Net for segmentation. Another perspective on the evolving landscape of AI models for diabetic retinopathy screening lies in the increasing adoption of Convolutional Neural Networks for classification tasks and U-Net architectures for segmentation purposes; however, there is a growing realization within the research community that these techniques, while powerful individually, can be even more effective when integrated. This integration holds promise for not only diagnosing DR, but also accurately classifying its different stages, thereby enabling more tailored treatment strategies. Despite this potential, the development of AI models for DR screening is fraught with challenges. Chief among these is the difficulty in obtaining the high-quality, labeled data necessary for training models to perform effectively. This scarcity of data poses significant barriers to achieving robust performance and can hinder progress in developing accurate screening systems. Moreover, managing the complexity of these models, particularly deep neural networks, presents its own set of challenges. Additionally, interpreting the outputs of these models and ensuring their reliability in real-world clinical settings remain ongoing concerns. Furthermore, the iterative process of training and adapting these models to specific datasets can be time-consuming and resource-intensive. These challenges underscore the multifaceted nature of developing effective AI models for DR screening. Addressing these obstacles requires concerted efforts from researchers, clinicians, and technologists to develop new approaches and overcome existing limitations. By doing so, a full potential of AI may transform DR screening and improve patient outcomes.
Collapse
Affiliation(s)
- Zineb Farahat
- LISTD Laboratory, Mines School of Rabat, Rabat 10000, Morocco; Cheikh Zaïd Foundation Medical Simulation Center, Rabat 10000, Morocco.
| | - Nabila Zrira
- LISTD Laboratory, Mines School of Rabat, Rabat 10000, Morocco
| | | | - Yasmine Bennani
- Cheikh Zaïd Ophthalmic Center, Cheikh Zaïd International University Hospital, Rabat 10000, Morocco; Institut Supérieur d'Ingénierie et Technologies de Santé/Faculté de Médecine Abulcasis, Université Internationale Abulcasis des Sciences de la Santé, Rabat 10000, Morocco
| | - Soufiane Bencherif
- Cheikh Zaïd Ophthalmic Center, Cheikh Zaïd International University Hospital, Rabat 10000, Morocco; Institut Supérieur d'Ingénierie et Technologies de Santé/Faculté de Médecine Abulcasis, Université Internationale Abulcasis des Sciences de la Santé, Rabat 10000, Morocco
| | - Safia Benamar
- Cheikh Zaïd Ophthalmic Center, Cheikh Zaïd International University Hospital, Rabat 10000, Morocco; Institut Supérieur d'Ingénierie et Technologies de Santé/Faculté de Médecine Abulcasis, Université Internationale Abulcasis des Sciences de la Santé, Rabat 10000, Morocco
| | - Mohammed Belmekki
- Cheikh Zaïd Ophthalmic Center, Cheikh Zaïd International University Hospital, Rabat 10000, Morocco; Institut Supérieur d'Ingénierie et Technologies de Santé/Faculté de Médecine Abulcasis, Université Internationale Abulcasis des Sciences de la Santé, Rabat 10000, Morocco
| | - Mohamed Nabil Ngote
- LISTD Laboratory, Mines School of Rabat, Rabat 10000, Morocco; Institut Supérieur d'Ingénierie et Technologies de Santé/Faculté de Médecine Abulcasis, Université Internationale Abulcasis des Sciences de la Santé, Rabat 10000, Morocco
| | - Kawtar Megdiche
- Cheikh Zaïd Foundation Medical Simulation Center, Rabat 10000, Morocco
| |
Collapse
|
2
|
Bhulakshmi D, Rajput DS. A systematic review on diabetic retinopathy detection and classification based on deep learning techniques using fundus images. PeerJ Comput Sci 2024; 10:e1947. [PMID: 38699206 PMCID: PMC11065411 DOI: 10.7717/peerj-cs.1947] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Accepted: 02/28/2024] [Indexed: 05/05/2024]
Abstract
Diabetic retinopathy (DR) is the leading cause of visual impairment globally. It occurs due to long-term diabetes with fluctuating blood glucose levels. It has become a significant concern for people in the working age group as it can lead to vision loss in the future. Manual examination of fundus images is time-consuming and requires much effort and expertise to determine the severity of the retinopathy. To diagnose and evaluate the disease, deep learning-based technologies have been used, which analyze blood vessels, microaneurysms, exudates, macula, optic discs, and hemorrhages also used for initial detection and grading of DR. This study examines the fundamentals of diabetes, its prevalence, complications, and treatment strategies that use artificial intelligence methods such as machine learning (ML), deep learning (DL), and federated learning (FL). The research covers future studies, performance assessments, biomarkers, screening methods, and current datasets. Various neural network designs, including recurrent neural networks (RNNs), generative adversarial networks (GANs), and applications of ML, DL, and FL in the processing of fundus images, such as convolutional neural networks (CNNs) and their variations, are thoroughly examined. The potential research methods, such as developing DL models and incorporating heterogeneous data sources, are also outlined. Finally, the challenges and future directions of this research are discussed.
Collapse
Affiliation(s)
- Dasari Bhulakshmi
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| | - Dharmendra Singh Rajput
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, India
| |
Collapse
|
3
|
Hemanth SV, Alagarsamy S, Rajkumar TD. A novel deep learning model for diabetic retinopathy detection in retinal fundus images using pre-trained CNN and HWBLSTM. J Biomol Struct Dyn 2024:1-19. [PMID: 38373067 DOI: 10.1080/07391102.2024.2314269] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Accepted: 01/30/2024] [Indexed: 02/21/2024]
Abstract
Diabetic retinopathy (DR) is a global visual indicator of diabetes that leads to blindness and loss of vision. Manual testing presents a more difficult task when attempting to detect DR due to the complexity and variances of DR. Early detection and treatment prevent the diabetic patients from visual loss. Also classifying the intensity and levels of DR is crucial to provide necessary treatment. This study develops a novel deep learning (DL) approach called He Weighted Bi-directional Long Short-term Memory (HWBLSTM) with an effective transfer learning technique for detecting DR from the RFI. The collected fundus images initially undergo preprocessing to improve their quality, which includes noise removal and contrast enhancement using a Hybrid Gaussian Filter and probability density Function-based Gamma Correction (HGFPDFGC) technique. The segmentation procedure divides the image into subgroups and is crucial for accurate detection and classification. The segmentation of the study initially removes the optical disk (OD) and blood vessels (BVs) from the preprocessed images using mathematical morphological operations. Next, it segments the retinal lesions from the OD and BV removed images using the Enhanced Grasshopper Optimization-based Region Growing Algorithm (EGORGA). Then, the features from the segmented retinal lesions are learned using a Squeeze Net (SQN), and the dimensionality reduction of the extracted features is done using the Modified Singular Value Decomposition (MSVD) approach. Finally, the classification is performed by employing the HWBLSTM approach, which classifies the DR abnormalities in datasets as non-DR (NDR), non-proliferative DR (NPDR), moderate NPDR (MDNPDR), and severe DR, also known as proliferative DR (PDR). The proposed approach is implemented on APTOS as well as MESSIDOR datasets. The outcomes proved that the proposed technique accurately identifies the DR with minimal computation overhead compared to the existing approaches.Communicated by Ramaswamy H. Sarma.
Collapse
Affiliation(s)
- S V Hemanth
- Department of Computer Science and Engineering, Hyderabad Institute of Technology and Management, Hyderabad, India
| | - Saravanan Alagarsamy
- Department of Information Technology, Sri Sivasubramaniya Nadar College of Engineering, Rajiv Gandhi Salai (OMR), Kalavakkam, India
| | - T Dhiliphan Rajkumar
- Department of Computer Science and Engineering, Kalasalingam Academy of Research and Education, Krishnankoil, India
| |
Collapse
|
4
|
Choi JY, Ryu IH, Kim JK, Lee IS, Yoo TK. Development of a generative deep learning model to improve epiretinal membrane detection in fundus photography. BMC Med Inform Decis Mak 2024; 24:25. [PMID: 38273286 PMCID: PMC10811871 DOI: 10.1186/s12911-024-02431-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2023] [Accepted: 01/17/2024] [Indexed: 01/27/2024] Open
Abstract
BACKGROUND The epiretinal membrane (ERM) is a common retinal disorder characterized by abnormal fibrocellular tissue at the vitreomacular interface. Most patients with ERM are asymptomatic at early stages. Therefore, screening for ERM will become increasingly important. Despite the high prevalence of ERM, few deep learning studies have investigated ERM detection in the color fundus photography (CFP) domain. In this study, we built a generative model to enhance ERM detection performance in the CFP. METHODS This deep learning study retrospectively collected 302 ERM and 1,250 healthy CFP data points from a healthcare center. The generative model using StyleGAN2 was trained using single-center data. EfficientNetB0 with StyleGAN2-based augmentation was validated using independent internal single-center data and external datasets. We randomly assigned healthcare center data to the development (80%) and internal validation (20%) datasets. Data from two publicly accessible sources were used as external validation datasets. RESULTS StyleGAN2 facilitated realistic CFP synthesis with the characteristic cellophane reflex features of the ERM. The proposed method with StyleGAN2-based augmentation outperformed the typical transfer learning without a generative adversarial network. The proposed model achieved an area under the receiver operating characteristic (AUC) curve of 0.926 for internal validation. AUCs of 0.951 and 0.914 were obtained for the two external validation datasets. Compared with the deep learning model without augmentation, StyleGAN2-based augmentation improved the detection performance and contributed to the focus on the location of the ERM. CONCLUSIONS We proposed an ERM detection model by synthesizing realistic CFP images with the pathological features of ERM through generative deep learning. We believe that our deep learning framework will help achieve a more accurate detection of ERM in a limited data setting.
Collapse
Affiliation(s)
- Joon Yul Choi
- Department of Biomedical Engineering, Yonsei University, Wonju, South Korea
| | - Ik Hee Ryu
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea
- Research and development department, VISUWORKS, Seoul, South Korea
| | - Jin Kuk Kim
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea
- Research and development department, VISUWORKS, Seoul, South Korea
| | - In Sik Lee
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea
| | - Tae Keun Yoo
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea.
- Research and development department, VISUWORKS, Seoul, South Korea.
| |
Collapse
|
5
|
Tian M, Wang H, Sun Y, Wu S, Tang Q, Zhang M. Fine-grained attention & knowledge-based collaborative network for diabetic retinopathy grading. Heliyon 2023; 9:e17217. [PMID: 37449186 PMCID: PMC10336422 DOI: 10.1016/j.heliyon.2023.e17217] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 06/09/2023] [Accepted: 06/10/2023] [Indexed: 07/18/2023] Open
Abstract
Accurate diabetic retinopathy (DR) grading is crucial for making the proper treatment plan to reduce the damage caused by vision loss. This task is challenging due to the fact that the DR related lesions are often small and subtle in visual differences and intra-class variations. Moreover, relationships between the lesions and the DR levels are complicated. Although many deep learning (DL) DR grading systems have been developed with some success, there are still rooms for grading accuracy improvement. A common issue is that not much medical knowledge was used in these DL DR grading systems. As a result, the grading results are not properly interpreted by ophthalmologists, thus hinder the potential for practical applications. This paper proposes a novel fine-grained attention & knowledge-based collaborative network (FA+KC-Net) to address this concern. The fine-grained attention network dynamically divides the extracted feature maps into smaller patches and effectively captures small image features that are meaningful in the sense of its training from large amount of retinopathy fundus images. The knowledge-based collaborative network extracts a-priori medical knowledge features, i.e., lesions such as the microaneurysms (MAs), soft exudates (SEs), hard exudates (EXs), and hemorrhages (HEs). Finally, decision rules are developed to fuse the DR grading results from the fine-grained network and the knowledge-based collaborative network to make the final grading. Extensive experiments are carried out on four widely-used datasets, the DDR, Messidor, APTOS, and EyePACS to evaluate the efficacy of our method and compare with other state-of-the-art (SOTA) DL models. Simulation results show that proposed FA+KC-Net is accurate and stable, achieves the best performances on the DDR, Messidor, and APTOS datasets.
Collapse
Affiliation(s)
- Miao Tian
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Hongqiu Wang
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Yingxue Sun
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Shaozhi Wu
- School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Qingqing Tang
- Department of Ophthalmology, West China Hospital, Sichuan University, Chengdu, 610041, China
| | - Meixia Zhang
- Department of Ophthalmology, West China Hospital, Sichuan University, Chengdu, 610041, China
| |
Collapse
|
6
|
Khan IU, Raiaan MAK, Fatema K, Azam S, Rashid RU, Mukta SH, Jonkman M, De Boer F. A Computer-Aided Diagnostic System to Identify Diabetic Retinopathy, Utilizing a Modified Compact Convolutional Transformer and Low-Resolution Images to Reduce Computation Time. Biomedicines 2023; 11:1566. [PMID: 37371661 DOI: 10.3390/biomedicines11061566] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2023] [Revised: 05/23/2023] [Accepted: 05/26/2023] [Indexed: 06/29/2023] Open
Abstract
Diabetic retinopathy (DR) is the foremost cause of blindness in people with diabetes worldwide, and early diagnosis is essential for effective treatment. Unfortunately, the present DR screening method requires the skill of ophthalmologists and is time-consuming. In this study, we present an automated system for DR severity classification employing the fine-tuned Compact Convolutional Transformer (CCT) model to overcome these issues. We assembled five datasets to generate a more extensive dataset containing 53,185 raw images. Various image pre-processing techniques and 12 types of augmentation procedures were applied to improve image quality and create a massive dataset. A new DR-CCTNet model is proposed. It is a modification of the original CCT model to address training time concerns and work with a large amount of data. Our proposed model delivers excellent accuracy even with low-pixel images and still has strong performance with fewer images, indicating that the model is robust. We compare our model's performance with transfer learning models such as VGG19, VGG16, MobileNetV2, and ResNet50. The test accuracy of the VGG19, ResNet50, VGG16, and MobileNetV2 were, respectively, 72.88%, 76.67%, 73.22%, and 71.98%. Our proposed DR-CCTNet model to classify DR outperformed all of these with a 90.17% test accuracy. This approach provides a novel and efficient method for the detection of DR, which may lower the burden on ophthalmologists and expedite treatment for patients.
Collapse
Affiliation(s)
- Inam Ullah Khan
- Health Informatics Research Lab, Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh
| | | | - Kaniz Fatema
- Health Informatics Research Lab, Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh
| | - Sami Azam
- Faculty of Science and Technology, Charles Darwin University, Darwin, NT 0909, Australia
| | - Rafi Ur Rashid
- Department of Computer Science and Engineering, Penn State University, State College, PA 16801, USA
| | - Saddam Hossain Mukta
- Department of Computer Science and Engineering, United International University, Dhaka 1212, Bangladesh
| | - Mirjam Jonkman
- Faculty of Science and Technology, Charles Darwin University, Darwin, NT 0909, Australia
| | - Friso De Boer
- Faculty of Science and Technology, Charles Darwin University, Darwin, NT 0909, Australia
| |
Collapse
|
7
|
Muchuchuti S, Viriri S. Retinal Disease Detection Using Deep Learning Techniques: A Comprehensive Review. J Imaging 2023; 9:84. [PMID: 37103235 PMCID: PMC10145952 DOI: 10.3390/jimaging9040084] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Revised: 04/02/2023] [Accepted: 04/07/2023] [Indexed: 04/28/2023] Open
Abstract
Millions of people are affected by retinal abnormalities worldwide. Early detection and treatment of these abnormalities could arrest further progression, saving multitudes from avoidable blindness. Manual disease detection is time-consuming, tedious and lacks repeatability. There have been efforts to automate ocular disease detection, riding on the successes of the application of Deep Convolutional Neural Networks (DCNNs) and vision transformers (ViTs) for Computer-Aided Diagnosis (CAD). These models have performed well, however, there remain challenges owing to the complex nature of retinal lesions. This work reviews the most common retinal pathologies, provides an overview of prevalent imaging modalities and presents a critical evaluation of current deep-learning research for the detection and grading of glaucoma, diabetic retinopathy, Age-Related Macular Degeneration and multiple retinal diseases. The work concluded that CAD, through deep learning, will increasingly be vital as an assistive technology. As future work, there is a need to explore the potential impact of using ensemble CNN architectures in multiclass, multilabel tasks. Efforts should also be expended on the improvement of model explainability to win the trust of clinicians and patients.
Collapse
Affiliation(s)
| | - Serestina Viriri
- School of Mathematics, Statistics and Computer Science, University of KwaZulu-Natal, Durban 4001, South Africa
| |
Collapse
|