1
|
Daudé P, Ramasawmy R, Javed A, Lederman RJ, Chow K, Campbell-Washburn AE. Inline automatic quality control of 2D phase-contrast flow MRI for subject-specific scan time adaptation. Magn Reson Med 2024; 92:751-760. [PMID: 38469944 PMCID: PMC11142871 DOI: 10.1002/mrm.30083] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2023] [Revised: 02/01/2024] [Accepted: 02/24/2024] [Indexed: 03/13/2024]
Abstract
PURPOSE To develop an inline automatic quality control to achieve consistent diagnostic image quality with subject-specific scan time, and to demonstrate this method for 2D phase-contrast flow MRI to reach a predetermined SNR. METHODS We designed a closed-loop feedback framework between image reconstruction and data acquisition to intermittently check SNR (every 20 s) and automatically stop the acquisition when a target SNR is achieved. A free-breathing 2D pseudo-golden-angle spiral phase-contrast sequence was modified to listen for image-quality messages from the reconstructions. Ten healthy volunteers and 1 patient were imaged at 0.55 T. Target SNR was selected based on retrospective analysis of cardiac output error, and performance of the automatic SNR-driven "stop" was assessed inline. RESULTS SNR calculation and automated segmentation was feasible within 20 s with inline deployment. The SNR-driven acquisition time was 2 min 39 s ± 67 s (aorta) and 3 min ± 80 s (main pulmonary artery) with a min/max acquisition time of 1 min 43 s/4 min 52 s (aorta) and 1 min 43 s/5 min 50 s (main pulmonary artery) across 6 healthy volunteers, while ensuring a diagnostic measurement with relative absolute error in quantitative flow measurement lower than 2.1% (aorta) and 6.3% (main pulmonary artery). CONCLUSION The inline quality control enables subject-specific optimized scan times while ensuring consistent diagnostic image quality. The distribution of automated stopping times across the population revealed the value of a subject-specific scan time.
Collapse
Affiliation(s)
- Pierre Daudé
- Laboratory of Imaging Technology, Cardiovascular Branch, Division of Intramural Research, National Heart, Lung & Blood Institute, NIH, USA
| | - Rajiv Ramasawmy
- Laboratory of Imaging Technology, Cardiovascular Branch, Division of Intramural Research, National Heart, Lung & Blood Institute, NIH, USA
| | - Ahsan Javed
- Laboratory of Imaging Technology, Cardiovascular Branch, Division of Intramural Research, National Heart, Lung & Blood Institute, NIH, USA
| | - Robert J Lederman
- Cardiovascular Branch, Division of Intramural Research, National Heart, Lung & Blood Institute, NIH, USA
| | - Kelvin Chow
- Siemens Healthcare Ltd., Calgary, AB, Canada
| | - Adrienne E. Campbell-Washburn
- Laboratory of Imaging Technology, Cardiovascular Branch, Division of Intramural Research, National Heart, Lung & Blood Institute, NIH, USA
| |
Collapse
|
2
|
Loizillon S, Bottani S, Maire A, Ströer S, Dormont D, Colliot O, Burgos N. Automatic motion artefact detection in brain T1-weighted magnetic resonance images from a clinical data warehouse using synthetic data. Med Image Anal 2024; 93:103073. [PMID: 38176355 DOI: 10.1016/j.media.2023.103073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Revised: 12/12/2023] [Accepted: 12/21/2023] [Indexed: 01/06/2024]
Abstract
Containing the medical data of millions of patients, clinical data warehouses (CDWs) represent a great opportunity to develop computational tools. Magnetic resonance images (MRIs) are particularly sensitive to patient movements during image acquisition, which will result in artefacts (blurring, ghosting and ringing) in the reconstructed image. As a result, a significant number of MRIs in CDWs are corrupted by these artefacts and may be unusable. Since their manual detection is impossible due to the large number of scans, it is necessary to develop tools to automatically exclude (or at least identify) images with motion in order to fully exploit CDWs. In this paper, we propose a novel transfer learning method from research to clinical data for the automatic detection of motion in 3D T1-weighted brain MRI. The method consists of two steps: a pre-training on research data using synthetic motion, followed by a fine-tuning step to generalise our pre-trained model to clinical data, relying on the labelling of 4045 images. The objectives were both (1) to be able to exclude images with severe motion, (2) to detect mild motion artefacts. Our approach achieved excellent accuracy for the first objective with a balanced accuracy nearly similar to that of the annotators (balanced accuracy>80 %). However, for the second objective, the performance was weaker and substantially lower than that of human raters. Overall, our framework will be useful to take advantage of CDWs in medical imaging and highlight the importance of a clinical validation of models trained on research data.
Collapse
Affiliation(s)
- Sophie Loizillon
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute - ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié-Salpêtrière, Paris 75013, France
| | - Simona Bottani
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute - ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié-Salpêtrière, Paris 75013, France
| | - Aurélien Maire
- AP-HP, Innovation & Données - Département des Services Numériques, Paris 75012, France
| | - Sebastian Ströer
- AP-HP, Hôpital de la Pitié Salpêtrière, Department of Neuroradiology, Paris 75013, France
| | - Didier Dormont
- AP-HP, Hôpital de la Pitié Salpêtrière, Department of Neuroradiology, Paris 75013, France; Sorbonne Université, Institut du Cerveau - Paris Brain Institute - ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié-Salpêtrière, DMU DIAMENT, Paris 75013, France
| | - Olivier Colliot
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute - ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié-Salpêtrière, Paris 75013, France
| | - Ninon Burgos
- Sorbonne Université, Institut du Cerveau - Paris Brain Institute - ICM, CNRS, Inria, Inserm, AP-HP, Hôpital de la Pitié-Salpêtrière, Paris 75013, France.
| |
Collapse
|
3
|
Torfeh T, Aouadi S, Yoganathan SA, Paloor S, Hammoud R, Al-Hammadi N. Deep Learning Approaches for Automatic Quality Assurance of Magnetic Resonance Images Using ACR Phantom. BMC Med Imaging 2023; 23:197. [PMID: 38031032 PMCID: PMC10685462 DOI: 10.1186/s12880-023-01157-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2023] [Accepted: 11/17/2023] [Indexed: 12/01/2023] Open
Abstract
BACKGROUND In recent years, there has been a growing trend towards utilizing Artificial Intelligence (AI) and machine learning techniques in medical imaging, including for the purpose of automating quality assurance. In this research, we aimed to develop and evaluate various deep learning-based approaches for automatic quality assurance of Magnetic Resonance (MR) images using the American College of Radiology (ACR) standards. METHODS The study involved the development, optimization, and testing of custom convolutional neural network (CNN) models. Additionally, popular pre-trained models such as VGG16, VGG19, ResNet50, InceptionV3, EfficientNetB0, and EfficientNetB5 were trained and tested. The use of pre-trained models, particularly those trained on the ImageNet dataset, for transfer learning was also explored. Two-class classification models were employed for assessing spatial resolution and geometric distortion, while an approach classifying the image into 10 classes representing the number of visible spokes was used for the low contrast. RESULTS Our results showed that deep learning-based methods can be effectively used for MR image quality assurance and can improve the performance of these models. The low contrast test was one of the most challenging tests within the ACR phantom. CONCLUSIONS Overall, for geometric distortion and spatial resolution, all of the deep learning models tested produced prediction accuracy of 80% or higher. The study also revealed that training the models from scratch performed slightly better compared to transfer learning. For the low contrast, our investigation emphasized the adaptability and potential of deep learning models. The custom CNN models excelled in predicting the number of visible spokes, achieving commendable accuracy, recall, precision, and F1 scores.
Collapse
Affiliation(s)
- Tarraf Torfeh
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar.
| | - Souha Aouadi
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - S A Yoganathan
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - Satheesh Paloor
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - Rabih Hammoud
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - Noora Al-Hammadi
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| |
Collapse
|
4
|
Fu Y, Jiang L, Pan S, Chen P, Wang X, Dai N, Chen X, Xu M. Deep multi-task learning for nephropathy diagnosis on immunofluorescence images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 241:107747. [PMID: 37619430 DOI: 10.1016/j.cmpb.2023.107747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/23/2023] [Revised: 06/14/2023] [Accepted: 08/03/2023] [Indexed: 08/26/2023]
Abstract
BACKGROUND AND OBJECTIVE As an advanced technique, immunofluorescence (IF) is one of the most widely-used medical image for nephropathy diagnosis, due to its ease of acquisition with low cost. In practice, the clinically collected IF images are commonly corrupted by blurs at different degrees, mainly because of the inaccurate focus at the acquisition stage. Although deep neural network (DNN) methods achieve the great success in nephropathy diagnosis, their performance dramatically drops over the blurred IF images. This significantly limits the potential of leveraging the advanced DNN techniques in real-world nephropathy diagnosis scenarios. METHODS This paper first establishes two IF databases with synthetic blurs (IFVB) and real-world blurs (Real-IF) for nephropathy diagnosis, respectively, including 1,659 patients and 6,521 IF images with various degrees of blurs. According to the analysis on these two databases, we propose a deep hierarchical multi-task learning based nephropathy diagnosis (DeepMT-ND) method to bridge the gap between the low-level vision and high-level medical tasks. Specifically, DeepMT-ND simultaneously handles the main task of automatic nephropathy diagnosis, as well as the auxiliary tasks of image quality assessment (IQA) and de-blurring. RESULTS Extensive experiments show the superiority of our DeepMT-ND in terms of diagnosis accuracy and generalization ability. For instance, our method performs better than nephrologists with at least 15.4% and 6.5% accuracy improvements in IFVB and Real-IF, respectively. Meanwhile, our method also achieves comparable performance in two auxiliary tasks of IQA and de-blurring on blurred IF images. CONCLUSIONS In this paper, we propose a new DeepMT-ND method for nephropathy diagnosis on blurred IF images. The proposed hierarchical multi-task learning framework provides the new scope to narrow the gap between the low-level vision and high-level medical tasks, and will contribute to nephropathy diagnosis in clinical scenarios. The diagnosis accuracy and generalization ability of DeepMT-ND are experimentally verified to be effective over both synthetic and real-world databases.
Collapse
Affiliation(s)
- Yibing Fu
- School of Electronic and Information Engineering, Beihang University, Beijing, China
| | - Lai Jiang
- School of Electronic and Information Engineering, Beihang University, Beijing, China
| | - Sai Pan
- Department of Nephrology, Chinese People's Liberation Army General Hospital, Beijing, China
| | - Pu Chen
- Department of Nephrology, Chinese People's Liberation Army General Hospital, Beijing, China
| | - Xiaofei Wang
- Department of Clinical Neurosciences, University of Cambridge, Cambridge, United Kingdom
| | - Ning Dai
- School of Electronic and Information Engineering, Beihang University, Beijing, China
| | - Xiangmei Chen
- Department of Nephrology, Chinese People's Liberation Army General Hospital, Beijing, China.
| | - Mai Xu
- School of Electronic and Information Engineering, Beihang University, Beijing, China.
| |
Collapse
|
5
|
Kim H, Kang SW, Kim JH, Nagar H, Sabuncu M, Margolis DJA, Kim CK. The role of AI in prostate MRI quality and interpretation: Opportunities and challenges. Eur J Radiol 2023; 165:110887. [PMID: 37245342 DOI: 10.1016/j.ejrad.2023.110887] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2023] [Revised: 05/06/2023] [Accepted: 05/20/2023] [Indexed: 05/30/2023]
Abstract
Prostate MRI plays an important role in imaging the prostate gland and surrounding tissues, particularly in the diagnosis and management of prostate cancer. With the widespread adoption of multiparametric magnetic resonance imaging in recent years, the concerns surrounding the variability of imaging quality have garnered increased attention. Several factors contribute to the inconsistency of image quality, such as acquisition parameters, scanner differences and interobserver variabilities. While efforts have been made to standardize image acquisition and interpretation via the development of systems, such as PI-RADS and PI-QUAL, the scoring systems still depend on the subjective experience and acumen of humans. Artificial intelligence (AI) has been increasingly used in many applications, including medical imaging, due to its ability to automate tasks and lower human error rates. These advantages have the potential to standardize the tasks of image interpretation and quality control of prostate MRI. Despite its potential, thorough validation is required before the implementation of AI in clinical practice. In this article, we explore the opportunities and challenges of AI, with a focus on the interpretation and quality of prostate MRI.
Collapse
Affiliation(s)
- Heejong Kim
- Department of Radiology, Weill Cornell Medical College, 525 E 68th St Box 141, New York, NY 10021, United States
| | - Shin Won Kang
- Research Institute for Future Medicine, Samsung Medical Center, Republic of Korea
| | - Jae-Hun Kim
- Department of Radiology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Republic of Korea
| | - Himanshu Nagar
- Department of Radiation Oncology, Weill Cornell Medical College, 525 E 68th St, New York, NY 10021, United States
| | - Mert Sabuncu
- Department of Radiology, Weill Cornell Medical College, 525 E 68th St Box 141, New York, NY 10021, United States
| | - Daniel J A Margolis
- Department of Radiology, Weill Cornell Medical College, 525 E 68th St Box 141, New York, NY 10021, United States.
| | - Chan Kyo Kim
- Department of Radiology and Center for Imaging Science, Samsung Medical Center, Sungkyunkwan University School of Medicine, Republic of Korea
| |
Collapse
|
6
|
Pollak C, Kügler D, Breteler MMB, Reuter M. Quantifying MR Head Motion in the Rhineland Study - A Robust Method for Population Cohorts. Neuroimage 2023; 275:120176. [PMID: 37209757 DOI: 10.1016/j.neuroimage.2023.120176] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Revised: 04/22/2023] [Accepted: 05/15/2023] [Indexed: 05/22/2023] Open
Abstract
Head motion during MR acquisition reduces image quality and has been shown to bias neuromorphometric analysis. The quantification of head motion, therefore, has both neuroscientific as well as clinical applications, for example, to control for motion in statistical analyses of brain morphology, or as a variable of interest in neurological studies. The accuracy of markerless optical head tracking, however, is largely unexplored. Furthermore, no quantitative analysis of head motion in a general, mostly healthy population cohort exists thus far. In this work, we present a robust registration method for the alignment of depth camera data that sensitively estimates even small head movements of compliant participants. Our method outperforms the vendor-supplied method in three validation experiments: 1. similarity to fMRI motion traces as a low-frequency reference, 2. recovery of the independently acquired breathing signal as a high-frequency reference, and 3. correlation with image-based quality metrics in structural T1-weighted MRI. In addition to the core algorithm, we establish an analysis pipeline that computes average motion scores per time interval or per sequence for inclusion in downstream analyses. We apply the pipeline in the Rhineland Study, a large population cohort study, where we replicate age and body mass index (BMI) as motion correlates and show that head motion significantly increases over the duration of the scan session. We observe weak, yet significant interactions between this within-session increase and age, BMI, and sex. High correlations between fMRI and camera-based motion scores of proceeding sequences further suggest that fMRI motion estimates can be used as a surrogate score in the absence of better measures to control for motion in statistical analyses.
Collapse
Affiliation(s)
- Clemens Pollak
- AI in Medical Imaging, German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany
| | - David Kügler
- AI in Medical Imaging, German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany
| | - Monique M B Breteler
- Population Health Sciences, German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany; Institute for Medical Biometry, Informatics and Epidemiology (IMBIE), Faculty of Medicine, University of Bonn, Bonn, Germany
| | - Martin Reuter
- AI in Medical Imaging, German Center for Neurodegenerative Diseases (DZNE), Bonn, Germany; A.A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Boston, MA, USA; Department of Radiology, Harvard Medical School, Boston, MA, USA.
| |
Collapse
|
7
|
Qiao Z, Li L, Zhao X, Liu L, Zhang Q, Hechmi S, Atri M, Li X. An enhanced Runge Kutta boosted machine learning framework for medical diagnosis. Comput Biol Med 2023; 160:106949. [PMID: 37159961 DOI: 10.1016/j.compbiomed.2023.106949] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Revised: 03/27/2023] [Accepted: 04/15/2023] [Indexed: 05/11/2023]
Abstract
With the development and maturity of machine learning methods, medical diagnosis aided with machine learning methods has become a popular method to assist doctors in diagnosing and treating patients. However, machine learning methods are greatly affected by their hyperparameters, for instance, the kernel parameter in kernel extreme learning machine (KELM) and the learning rate in residual neural networks (ResNet). If the hyperparameters are appropriately set, the performance of the classifier can be significantly improved. To boost the performance of the machine learning methods, this paper proposes to improve the Runge Kutta optimizer (RUN) to adaptively adjust the hyperparameters of the machine learning methods for medical diagnosis purposes. Although RUN has a solid mathematical theoretical foundation, there are still some performance defects when dealing with complex optimization problems. To remedy these defects, this paper proposes a new enhanced RUN method with a grey wolf mechanism and an orthogonal learning mechanism called GORUN. The superior performance of the GORUN was validated against other well-established optimizers on IEEE CEC 2017 benchmark functions. Then, the proposed GORUN is employed to optimize the machine learning models, including the KELM and ResNet, to construct robust models for medical diagnosis. The performance of the proposed machine learning framework was validated on several medical data sets, and the experimental results have demonstrated its superiority.
Collapse
Affiliation(s)
- Zenglin Qiao
- School of Science, Beijing University of Posts and Telecommunications, Beijing, 100876, China.
| | - Lynn Li
- China Telecom Stocks Co.,Ltd., Hangzhou Branch, Hangzhou, 310000, China.
| | - Xinchao Zhao
- School of Science, Beijing University of Posts and Telecommunications, Beijing, 100876, China.
| | - Lei Liu
- College of Computer Science, Sichuan University, Chengdu, Sichuan, 610065, China.
| | - Qian Zhang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, Zhejiang, 325035, China.
| | - Shili Hechmi
- Dept. Computer Sciences, Tabuk University, Tabuk, Saudi Arabia.
| | - Mohamed Atri
- College of Computer Science, King Khalid University, Abha, Saudi Arabia.
| | - Xiaohua Li
- Library, Wenzhou University, Wenzhou, Zhejiang, 325035, China.
| |
Collapse
|
8
|
End to End Multitask Joint Learning Model for Osteoporosis Classification in CT Images. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2023; 2023:3018320. [PMID: 36970245 PMCID: PMC10036193 DOI: 10.1155/2023/3018320] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/10/2022] [Revised: 02/23/2023] [Accepted: 03/01/2023] [Indexed: 03/18/2023]
Abstract
Osteoporosis is a significant global health concern that can be difficult to detect early due to a lack of symptoms. At present, the examination of osteoporosis depends mainly on methods containing dual-energy X-ray, quantitative CT, etc., which are high costs in terms of equipment and human time. Therefore, a more efficient and economical method is urgently needed for diagnosing osteoporosis. With the development of deep learning, automatic diagnosis models for various diseases have been proposed. However, the establishment of these models generally requires images with only lesion areas, and annotating the lesion areas is time-consuming. To address this challenge, we propose a joint learning framework for osteoporosis diagnosis that combines localization, segmentation, and classification to enhance diagnostic accuracy. Our method includes a boundary heat map regression branch for thinning segmentation and a gated convolution module for adjusting context features in the classification module. We also integrate segmentation and classification features and propose a feature fusion module to adjust the weight of different levels of vertebrae. We trained our model on a self-built dataset and achieved an overall accuracy rate of 93.3% for the three label categories (normal, osteopenia, and osteoporosis) in the testing datasets. The area under the curve for the normal category is 0.973; for the osteopenia category, it is 0.965; and for the osteoporosis category, it is 0.985. Our method provides a promising alternative for the diagnosis of osteoporosis at present.
Collapse
|
9
|
Automated MRI Field of View Prescription from Region of Interest Prediction by Intra-Stack Attention Neural Network. BIOENGINEERING (BASEL, SWITZERLAND) 2023; 10:bioengineering10010092. [PMID: 36671663 PMCID: PMC9854842 DOI: 10.3390/bioengineering10010092] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Revised: 01/07/2023] [Accepted: 01/09/2023] [Indexed: 01/13/2023]
Abstract
Manual prescription of the field of view (FOV) by MRI technologists is variable and prolongs the scanning process. Often, the FOV is too large or crops critical anatomy. We propose a deep learning framework, trained by radiologists' supervision, for automating FOV prescription. An intra-stack shared feature extraction network and an attention network are used to process a stack of 2D image inputs to generate scalars defining the location of a rectangular region of interest (ROI). The attention mechanism is used to make the model focus on a small number of informative slices in a stack. Then, the smallest FOV that makes the neural network predicted ROI free of aliasing is calculated by an algebraic operation derived from MR sampling theory. The framework's performance is examined quantitatively with intersection over union (IoU) and pixel error on position and qualitatively with a reader study. The proposed model achieves an average IoU of 0.867 and an average ROI position error of 9.06 out of 512 pixels on 80 test cases, significantly better than two baseline models and not significantly different from a radiologist. Finally, the FOV given by the proposed framework achieves an acceptance rate of 92% from an experienced radiologist.
Collapse
|
10
|
Classification of Diabetic Retinopathy Severity in Fundus Images Using the Vision Transformer and Residual Attention. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2023; 2023:1305583. [PMID: 36636467 PMCID: PMC9831706 DOI: 10.1155/2023/1305583] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/26/2022] [Revised: 12/04/2022] [Accepted: 12/05/2022] [Indexed: 01/05/2023]
Abstract
Diabetic retinopathy (DR) is a common retinal vascular disease, which can cause severe visual impairment. It is of great clinical significance to use fundus images for intelligent diagnosis of DR. In this paper, an intelligent DR classification model of fundus images is proposed. This method can detect all the five stages of DR, including of no DR, mild, moderate, severe, and proliferative. This model is composed of two key modules. FEB, feature extraction block, is mainly used for feature extraction of fundus images, and GPB, grading prediction block, is used to classify the five stages of DR. The transformer in the FEB has more fine-grained attention that can pay more attention to retinal hemorrhage and exudate areas. The residual attention in the GPB can effectively capture different spatial regions occupied by different classes of objects. Comprehensive experiments on DDR datasets well demonstrate the superiority of our method, and compared with the benchmark method, our method has achieved competitive performance.
Collapse
|