1
|
Lisha LB, Helen Sulochana C. DEC-DRR: deep ensemble of classification model for diabetic retinopathy recognition. Med Biol Eng Comput 2024; 62:2911-2938. [PMID: 38713340 DOI: 10.1007/s11517-024-03076-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2023] [Accepted: 03/16/2024] [Indexed: 05/08/2024]
Abstract
Most diabetes patients are liable to have diabetic retinopathy (DR); however, the majority of them might not be even aware of the ailment. Therefore, early detection and treatment of DR are necessary to prevent vision loss. But, avoiding DR is not a simple process. An ophthalmologist can typically identify DR through an optical evaluation of the fundus and through the evaluation of color pictures. However, due to the increased count of DR patients, this could not be possible as it consumes more time. To rectify this problem, a novel deep ensemble-based DR classification technique is developed in this work. Initially, a Wiener filter (WF) is applied for preprocessing the image. Then, the enhanced U-Net-based segmentation process is done. Subsequent to the segmentation process, features are extracted that include statistical features, inferior superior nasal temporal (ISNT), cup to disc ratio (CDR), and improved LGBP as well. Further, deep ensemble classifiers (DEC) like CNN, Bi-GRU, and DMN are used to recognize the disease. The outcomes from DMN, CNN, and Bi-GRU are then subjected to improved SLF. Additionally, the weights of DMN, CNN, and Bi-GRU are adjusted via pelican updated Tasmanian devil optimization (PU-TDO). Finally, outputs on DR (microaneurysms, hemorrhages, hard exudates, and soft exudates) are obtained. The performance of DEC + PU-TDO for diabetic retinopathy is computed over extant models with regard to different measures for four datasets. The results on accuracy using the DEC + PU-TDO scheme for the IDRID dataset is maximum around 0.975 at 90th LP while other models have less accuracy. The FPR of DEC + PU-TDO is less around 0.039 at the 90th LP for the SUSTech-SYSU dataset, while other extant models have maximum FPR.
Collapse
Affiliation(s)
- L B Lisha
- Department of Computer Science and Engineering, Marthandam College of Engineering and Technology, Kuttakuzhi, Veeyannoor, Kanyakumari, Tamil Nadu, India.
| | - C Helen Sulochana
- Department of Electronics and Communication Engineering, St. Xavier's Catholic College of Engineering, Chunkankadai, Kanyakumari, Tamil Nadu, India
| |
Collapse
|
2
|
Pavithra S, Jaladi D, Tamilarasi K. Optical imaging for diabetic retinopathy diagnosis and detection using ensemble models. Photodiagnosis Photodyn Ther 2024; 48:104259. [PMID: 38944405 DOI: 10.1016/j.pdpdt.2024.104259] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2024] [Revised: 06/16/2024] [Accepted: 06/20/2024] [Indexed: 07/01/2024]
Abstract
Diabetes, characterized by heightened blood sugar levels, can lead to a condition called Diabetic Retinopathy (DR), which adversely impacts the eyes due to elevated blood sugar affecting the retinal blood vessels. The most common cause of blindness in diabetics is thought to be Diabetic Retinopathy (DR), particularly in working-age individuals living in poor nations. People with type 1 or type 2 diabetes may develop this illness, and the risk rises with the length of diabetes and inadequate blood sugar management. There are limits to traditional approaches for the early identification of diabetic retinopathy (DR). In order to diagnose diabetic retinopathy, a model based on Convolutional neural network (CNN) is used in a unique way in this research. The suggested model uses a number of deep learning (DL) models, such as VGG19, Resnet50, and InceptionV3, to extract features. After concatenation, these characteristics are sent through the CNN algorithm for classification. By combining the advantages of several models, ensemble approaches can be effective tools for detecting diabetic retinopathy and increase overall performance and resilience. Classification and image recognition are just a few of the tasks that may be accomplished with ensemble approaches like combination of VGG19,Inception V3 and Resnet 50 to achieve high accuracy. The proposed model is evaluated using a publicly accessible collection of fundus images.VGG19, ResNet50, and InceptionV3 differ in their neural network architectures, feature extraction capabilities, object detection methods, and approaches to retinal delineation. VGG19 may excel in capturing fine details, ResNet50 in recognizing complex patterns, and InceptionV3 in efficiently capturing multi-scale features. Their combined use in an ensemble approach can provide a comprehensive analysis of retinal images, aiding in the delineation of retinal regions and identification of abnormalities associated with diabetic retinopathy. For instance, micro aneurysms, the earliest signs of DR, often require precise detection of subtle vascular abnormalities. VGG19's proficiency in capturing fine details allows for the identification of these minute changes in retinal morphology. On the other hand, ResNet50's strength lies in recognizing intricate patterns, making it effective in detecting neoneovascularization and complex haemorrhagic lesions. Meanwhile, InceptionV3's multi-scale feature extraction enables comprehensive analysis, crucial for assessing macular oedema and ischaemic changes across different retinal layers.
Collapse
Affiliation(s)
- S Pavithra
- School of Computer Science and Engineering, VIT University, Chennai, Tamil Nadu, India.
| | - Deepika Jaladi
- School of Computer Science and Engineering, VIT University, Chennai, Tamil Nadu, India.
| | - K Tamilarasi
- School of Computer Science and Engineering, VIT University, Chennai, Tamil Nadu, India.
| |
Collapse
|
3
|
Zou K, Lin T, Han Z, Wang M, Yuan X, Chen H, Zhang C, Shen X, Fu H. Confidence-aware multi-modality learning for eye disease screening. Med Image Anal 2024; 96:103214. [PMID: 38815358 DOI: 10.1016/j.media.2024.103214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2023] [Revised: 05/06/2024] [Accepted: 05/17/2024] [Indexed: 06/01/2024]
Abstract
Multi-modal ophthalmic image classification plays a key role in diagnosing eye diseases, as it integrates information from different sources to complement their respective performances. However, recent improvements have mainly focused on accuracy, often neglecting the importance of confidence and robustness in predictions for diverse modalities. In this study, we propose a novel multi-modality evidential fusion pipeline for eye disease screening. It provides a measure of confidence for each modality and elegantly integrates the multi-modality information using a multi-distribution fusion perspective. Specifically, our method first utilizes normal inverse gamma prior distributions over pre-trained models to learn both aleatoric and epistemic uncertainty for uni-modality. Then, the normal inverse gamma distribution is analyzed as the Student's t distribution. Furthermore, within a confidence-aware fusion framework, we propose a mixture of Student's t distributions to effectively integrate different modalities, imparting the model with heavy-tailed properties and enhancing its robustness and reliability. More importantly, the confidence-aware multi-modality ranking regularization term induces the model to more reasonably rank the noisy single-modal and fused-modal confidence, leading to improved reliability and accuracy. Experimental results on both public and internal datasets demonstrate that our model excels in robustness, particularly in challenging scenarios involving Gaussian noise and modality missing conditions. Moreover, our model exhibits strong generalization capabilities to out-of-distribution data, underscoring its potential as a promising solution for multimodal eye disease screening.
Collapse
Affiliation(s)
- Ke Zou
- National Key Laboratory of Fundamental Science on Synthetic Vision, Sichuan University, Chengdu, 610065, China; College of Computer Science, Sichuan University, Chengdu, 610065, China
| | - Tian Lin
- Joint Shantou International Eye Center, Shantou University and the Chinese University of Hong Kong, Shantou 515041, China; Medical College, Shantou University, Shantou 515041, China
| | - Zongbo Han
- College of Intelligence and Computing, Tianjin University, Tianjin 300350, China
| | - Meng Wang
- Institute of High Performance Computing, Agency for Science, Technology and Research, 138632, Singapore
| | - Xuedong Yuan
- National Key Laboratory of Fundamental Science on Synthetic Vision, Sichuan University, Chengdu, 610065, China; College of Computer Science, Sichuan University, Chengdu, 610065, China.
| | - Haoyu Chen
- Joint Shantou International Eye Center, Shantou University and the Chinese University of Hong Kong, Shantou 515041, China; Medical College, Shantou University, Shantou 515041, China.
| | - Changqing Zhang
- College of Intelligence and Computing, Tianjin University, Tianjin 300350, China
| | - Xiaojing Shen
- National Key Laboratory of Fundamental Science on Synthetic Vision, Sichuan University, Chengdu, 610065, China; College of Mathematics, Sichuan University, Chengdu, 610065, China
| | - Huazhu Fu
- Institute of High Performance Computing, Agency for Science, Technology and Research, 138632, Singapore.
| |
Collapse
|
4
|
Yao J, Lim J, Lim GYS, Ong JCL, Ke Y, Tan TF, Tan TE, Vujosevic S, Ting DSW. Novel artificial intelligence algorithms for diabetic retinopathy and diabetic macular edema. EYE AND VISION (LONDON, ENGLAND) 2024; 11:23. [PMID: 38880890 PMCID: PMC11181581 DOI: 10.1186/s40662-024-00389-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/03/2024] [Accepted: 05/09/2024] [Indexed: 06/18/2024]
Abstract
BACKGROUND Diabetic retinopathy (DR) and diabetic macular edema (DME) are major causes of visual impairment that challenge global vision health. New strategies are needed to tackle these growing global health problems, and the integration of artificial intelligence (AI) into ophthalmology has the potential to revolutionize DR and DME management to meet these challenges. MAIN TEXT This review discusses the latest AI-driven methodologies in the context of DR and DME in terms of disease identification, patient-specific disease profiling, and short-term and long-term management. This includes current screening and diagnostic systems and their real-world implementation, lesion detection and analysis, disease progression prediction, and treatment response models. It also highlights the technical advancements that have been made in these areas. Despite these advancements, there are obstacles to the widespread adoption of these technologies in clinical settings, including regulatory and privacy concerns, the need for extensive validation, and integration with existing healthcare systems. We also explore the disparity between the potential of AI models and their actual effectiveness in real-world applications. CONCLUSION AI has the potential to revolutionize the management of DR and DME, offering more efficient and precise tools for healthcare professionals. However, overcoming challenges in deployment, regulatory compliance, and patient privacy is essential for these technologies to realize their full potential. Future research should aim to bridge the gap between technological innovation and clinical application, ensuring AI tools integrate seamlessly into healthcare workflows to enhance patient outcomes.
Collapse
Affiliation(s)
- Jie Yao
- Singapore National Eye Centre, Singapore Eye Research Institute, 11 Third Hospital Avenue, Singapore, 168751, Singapore
- Duke-NUS Medical School, Singapore, Singapore
| | - Joshua Lim
- Singapore National Eye Centre, Singapore Eye Research Institute, 11 Third Hospital Avenue, Singapore, 168751, Singapore
| | - Gilbert Yong San Lim
- Duke-NUS Medical School, Singapore, Singapore
- SingHealth AI Health Program, Singapore, Singapore
| | - Jasmine Chiat Ling Ong
- Duke-NUS Medical School, Singapore, Singapore
- Division of Pharmacy, Singapore General Hospital, Singapore, Singapore
| | - Yuhe Ke
- Department of Anesthesiology and Perioperative Science, Singapore General Hospital, Singapore, Singapore
| | - Ting Fang Tan
- Singapore National Eye Centre, Singapore Eye Research Institute, 11 Third Hospital Avenue, Singapore, 168751, Singapore
| | - Tien-En Tan
- Singapore National Eye Centre, Singapore Eye Research Institute, 11 Third Hospital Avenue, Singapore, 168751, Singapore
- Duke-NUS Medical School, Singapore, Singapore
| | - Stela Vujosevic
- Department of Biomedical, Surgical and Dental Sciences, University of Milan, Milan, Italy
- Eye Clinic, IRCCS MultiMedica, Milan, Italy
| | - Daniel Shu Wei Ting
- Singapore National Eye Centre, Singapore Eye Research Institute, 11 Third Hospital Avenue, Singapore, 168751, Singapore.
- Duke-NUS Medical School, Singapore, Singapore.
- SingHealth AI Health Program, Singapore, Singapore.
| |
Collapse
|
5
|
Liu Z, Hu Y, Qiu Z, Niu Y, Zhou D, Li X, Shen J, Jiang H, Li H, Liu J. Cross-modal attention network for retinal disease classification based on multi-modal images. BIOMEDICAL OPTICS EXPRESS 2024; 15:3699-3714. [PMID: 38867787 PMCID: PMC11166426 DOI: 10.1364/boe.516764] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/03/2024] [Revised: 04/15/2024] [Accepted: 05/02/2024] [Indexed: 06/14/2024]
Abstract
Multi-modal eye disease screening improves diagnostic accuracy by providing lesion information from different sources. However, existing multi-modal automatic diagnosis methods tend to focus on the specificity of modalities and ignore the spatial correlation of images. This paper proposes a novel cross-modal retinal disease diagnosis network (CRD-Net) that digs out the relevant features from modal images aided for multiple retinal disease diagnosis. Specifically, our model introduces a cross-modal attention (CMA) module to query and adaptively pay attention to the relevant features of the lesion in the different modal images. In addition, we also propose multiple loss functions to fuse features with modality correlation and train a multi-modal retinal image classification network to achieve a more accurate diagnosis. Experimental evaluation on three publicly available datasets shows that our CRD-Net outperforms existing single-modal and multi-modal methods, demonstrating its superior performance.
Collapse
Affiliation(s)
- Zirong Liu
- School of Ophthalmology and Optometry and Eye Hospital, Wenzhou Medical University, Wenzhou 325027, China
| | - Yan Hu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
| | - Zhongxi Qiu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
| | - Yanyan Niu
- School of Ophthalmology and Optometry and Eye Hospital, Wenzhou Medical University, Wenzhou 325027, China
| | - Dan Zhou
- School of Ophthalmology and Optometry and Eye Hospital, Wenzhou Medical University, Wenzhou 325027, China
| | - Xiaoling Li
- School of Ophthalmology and Optometry and Eye Hospital, Wenzhou Medical University, Wenzhou 325027, China
| | - Junyong Shen
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
| | - Hongyang Jiang
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
| | - Heng Li
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
| | - Jiang Liu
- School of Ophthalmology and Optometry and Eye Hospital, Wenzhou Medical University, Wenzhou 325027, China
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen 518055, China
| |
Collapse
|
6
|
Zhang H, Liu J, Liu W, Chen H, Yu Z, Yuan Y, Wang P, Qin J. MHD-Net: Memory-Aware Hetero-Modal Distillation Network for Thymic Epithelial Tumor Typing With Missing Pathology Modality. IEEE J Biomed Health Inform 2024; 28:3003-3014. [PMID: 38470599 DOI: 10.1109/jbhi.2024.3376462] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/14/2024]
Abstract
Fusing multi-modal radiology and pathology data with complementary information can improve the accuracy of tumor typing. However, collecting pathology data is difficult since it is high-cost and sometimes only obtainable after the surgery, which limits the application of multi-modal methods in diagnosis. To address this problem, we propose comprehensively learning multi-modal radiology-pathology data in training, and only using uni-modal radiology data in testing. Concretely, a Memory-aware Hetero-modal Distillation Network (MHD-Net) is proposed, which can distill well-learned multi-modal knowledge with the assistance of memory from the teacher to the student. In the teacher, to tackle the challenge in hetero-modal feature fusion, we propose a novel spatial-differentiated hetero-modal fusion module (SHFM) that models spatial-specific tumor information correlations across modalities. As only radiology data is accessible to the student, we store pathology features in the proposed contrast-boosted typing memory module (CTMM) that achieves type-wise memory updating and stage-wise contrastive memory boosting to ensure the effectiveness and generalization of memory items. In the student, to improve the cross-modal distillation, we propose a multi-stage memory-aware distillation (MMD) scheme that reads memory-aware pathology features from CTMM to remedy missing modal-specific information. Furthermore, we construct a Radiology-Pathology Thymic Epithelial Tumor (RPTET) dataset containing paired CT and WSI images with annotations. Experiments on the RPTET and CPTAC-LUAD datasets demonstrate that MHD-Net significantly improves tumor typing and outperforms existing multi-modal methods on missing modality situations.
Collapse
|
7
|
Uppamma P, Bhattacharya S. A multidomain bio-inspired feature extraction and selection model for diabetic retinopathy severity classification: an ensemble learning approach. Sci Rep 2023; 13:18572. [PMID: 37903967 PMCID: PMC10616283 DOI: 10.1038/s41598-023-45886-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Accepted: 10/25/2023] [Indexed: 11/01/2023] Open
Abstract
Diabetes retinopathy (DR) is one of the leading causes of blindness globally. Early detection of this condition is essential for preventing patients' loss of eyesight caused by diabetes mellitus being untreated for an extended period. This paper proposes the design of an augmented bioinspired multidomain feature extraction and selection model for diabetic retinopathy severity estimation using an ensemble learning process. The proposed approach initiates by identifying DR severity levels from retinal images that segment the optical disc, macula, blood vessels, exudates, and hemorrhages using an adaptive thresholding process. Once the images are segmented, multidomain features are extracted from the retinal images, including frequency, entropy, cosine, gabor, and wavelet components. These data were fed into a novel Modified Moth Flame Optimization-based feature selection method that assisted in optimal feature selection. Finally, an ensemble model using various ML (machine learning) algorithms, which included Naive Bayes, K-Nearest Neighbours, Support Vector Machine, Multilayer Perceptron, Random Forests, and Logistic Regression were used to identify the various severity complications of DR. The experiments on different openly accessible data sources have shown that the proposed method outperformed conventional methods and achieved an Accuracy of 96.5% in identifying DR severity levels.
Collapse
Affiliation(s)
- Posham Uppamma
- School of Information Technology and Engineering, Vellore Institute of Technology, Vellore, Tamilnadu, 632014, India
| | - Sweta Bhattacharya
- School of Information Technology and Engineering, Vellore Institute of Technology, Vellore, Tamilnadu, 632014, India.
| |
Collapse
|
8
|
Carrera-Escalé L, Benali A, Rathert AC, Martín-Pinardel R, Bernal-Morales C, Alé-Chilet A, Barraso M, Marín-Martinez S, Feu-Basilio S, Rosinés-Fonoll J, Hernandez T, Vilá I, Castro-Dominguez R, Oliva C, Vinagre I, Ortega E, Gimenez M, Vellido A, Romero E, Zarranz-Ventura J. Radiomics-Based Assessment of OCT Angiography Images for Diabetic Retinopathy Diagnosis. OPHTHALMOLOGY SCIENCE 2022; 3:100259. [PMID: 36578904 PMCID: PMC9791596 DOI: 10.1016/j.xops.2022.100259] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/30/2022] [Revised: 10/25/2022] [Accepted: 11/14/2022] [Indexed: 11/23/2022]
Abstract
Purpose To evaluate the diagnostic accuracy of machine learning (ML) techniques applied to radiomic features extracted from OCT and OCT angiography (OCTA) images for diabetes mellitus (DM), diabetic retinopathy (DR), and referable DR (R-DR) diagnosis. Design Cross-sectional analysis of a retinal image dataset from a previous prospective OCTA study (ClinicalTrials.govNCT03422965). Participants Patients with type 1 DM and controls included in the progenitor study. Methods Radiomic features were extracted from fundus retinographies, OCT, and OCTA images in each study eye. Logistic regression, linear discriminant analysis, support vector classifier (SVC)-linear, SVC-radial basis function, and random forest models were created to evaluate their diagnostic accuracy for DM, DR, and R-DR diagnosis in all image types. Main Outcome Measures Area under the receiver operating characteristic curve (AUC) mean and standard deviation for each ML model and each individual and combined image types. Results A dataset of 726 eyes (439 individuals) were included. For DM diagnosis, the greatest AUC was observed for OCT (0.82, 0.03). For DR detection, the greatest AUC was observed for OCTA (0.77, 0.03), especially in the 3 × 3 mm superficial capillary plexus OCTA scan (0.76, 0.04). For R-DR diagnosis, the greatest AUC was observed for OCTA (0.87, 0.12) and the deep capillary plexus OCTA scan (0.86, 0.08). The addition of clinical variables (age, sex, etc.) improved most models AUC for DM, DR and R-DR diagnosis. The performance of the models was similar in unilateral and bilateral eyes image datasets. Conclusions Radiomics extracted from OCT and OCTA images allow identification of patients with DM, DR, and R-DR using standard ML classifiers. OCT was the best test for DM diagnosis, OCTA for DR and R-DR diagnosis and the addition of clinical variables improved most models. This pioneer study demonstrates that radiomics-based ML techniques applied to OCT and OCTA images may be an option for DR screening in patients with type 1 DM. Financial Disclosures Proprietary or commercial disclosure may be found after the references.
Collapse
Key Words
- AI, artificial intelligence
- AUC, area under the curve
- Artificial intelligence
- DCP, deep capillary plexus
- DM, diabetes mellitus
- DR, diabetic retinopathy
- Diabetic retinopathy
- FR, fundus retinographies
- LDA, linear discriminant analysis
- LR, logistic regression
- ML, machine learning
- Machine learning
- OCT angiography
- OCTA, OCT angiography
- R-DR, referable DR
- RF, random forest
- Radiomics
- SCP, superficial capillary plexus
- SVC, support vector classifier
- rbf, radial basis function
Collapse
Affiliation(s)
- Laura Carrera-Escalé
- Intelligent Data Science and Artificial Intelligence (IDEAI) Research Center,Department of Computer Science, Facultat d’Informàtica de Barcelona (FIB), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain
| | - Anass Benali
- Intelligent Data Science and Artificial Intelligence (IDEAI) Research Center,Department of Computer Science, Facultat d’Informàtica de Barcelona (FIB), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain
| | - Ann-Christin Rathert
- Intelligent Data Science and Artificial Intelligence (IDEAI) Research Center,Department of Computer Science, Facultat d’Informàtica de Barcelona (FIB), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain
| | - Ruben Martín-Pinardel
- Intelligent Data Science and Artificial Intelligence (IDEAI) Research Center,Department of Computer Science, Facultat d’Informàtica de Barcelona (FIB), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain,August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain
| | | | - Anibal Alé-Chilet
- Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Marina Barraso
- Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Sara Marín-Martinez
- Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Silvia Feu-Basilio
- Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Josep Rosinés-Fonoll
- Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Teresa Hernandez
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Irene Vilá
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | | | - Cristian Oliva
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain
| | - Irene Vinagre
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Diabetes Unit, Hospital Clínic de Barcelona, Spain,Institut Clínic de Malalties Digestives i Metaboliques (ICMDM), Hospital Clínic de Barcelona, Spain
| | - Emilio Ortega
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Diabetes Unit, Hospital Clínic de Barcelona, Spain,Institut Clínic de Malalties Digestives i Metaboliques (ICMDM), Hospital Clínic de Barcelona, Spain
| | - Marga Gimenez
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Diabetes Unit, Hospital Clínic de Barcelona, Spain,Institut Clínic de Malalties Digestives i Metaboliques (ICMDM), Hospital Clínic de Barcelona, Spain
| | - Alfredo Vellido
- Intelligent Data Science and Artificial Intelligence (IDEAI) Research Center,Department of Computer Science, Facultat d’Informàtica de Barcelona (FIB), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain
| | - Enrique Romero
- Intelligent Data Science and Artificial Intelligence (IDEAI) Research Center,Department of Computer Science, Facultat d’Informàtica de Barcelona (FIB), Universitat Politècnica de Catalunya (UPC), Barcelona, Spain
| | - Javier Zarranz-Ventura
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain,Institut Clínic d´Oftalmología (ICOF), Hospital Clínic de Barcelona, Barcelona, Spain,Diabetes Unit, Hospital Clínic de Barcelona, Spain,School of Medicine, Universitat de Barcelona, Spain,Correspondence: Javier Zarranz-Ventura, MD, PhD, C/ Sabino Arana 1, Barcelona 08028, Spain.
| |
Collapse
|
9
|
Deng Z, Cai Y, Chen L, Gong Z, Bao Q, Yao X, Fang D, Yang W, Zhang S, Ma L. RFormer: Transformer-Based Generative Adversarial Network for Real Fundus Image Restoration on a New Clinical Benchmark. IEEE J Biomed Health Inform 2022; 26:4645-4655. [PMID: 35767498 DOI: 10.1109/jbhi.2022.3187103] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Ophthalmologists have used fundus images to screen and diagnose eye diseases. However, different equipments and ophthalmologists pose large variations to the quality of fundus images. Low-quality (LQ) degraded fundus images easily lead to uncertainty in clinical screening and generally increase the risk of misdiagnosis. Thus, real fundus image restoration is worth studying. Unfortunately, real clinical benchmark has not been explored for this task so far. In this paper, we investigate the real clinical fundus image restoration problem. Firstly, We establish a clinical dataset, Real Fundus (RF), including 120 low- and high-quality (HQ) image pairs. Then we propose a novel Transformer-based Generative Adversarial Network (RFormer) to restore the real degradation of clinical fundus images. The key component in our network is the Window-based Self-Attention Block (WSAB) which captures non-local self-similarity and long-range dependencies. To produce more visually pleasant results, a Transformer-based discriminator is introduced. Extensive experiments on our clinical benchmark show that the proposed RFormer significantly outperforms the state-of-the-art (SOTA) methods. In addition, experiments of downstream tasks such as vessel segmentation and optic disc/cup detection demonstrate that our proposed RFormer benefits clinical fundus image analysis and applications. The dataset, code, and models will be made publicly available at https://github.com/dengzhuo-AI/Real-Fundus.
Collapse
|