1
|
Al-Ani S, Guo H, Fyfe S, Long Z, Donnaz S, Kim Y. Effect of training sample size, image resolution and epochs on filamentous and floc-forming bacteria classification using machine learning. JOURNAL OF ENVIRONMENTAL MANAGEMENT 2025; 379:124803. [PMID: 40056595 DOI: 10.1016/j.jenvman.2025.124803] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/13/2024] [Revised: 01/15/2025] [Accepted: 03/01/2025] [Indexed: 03/10/2025]
Abstract
Computer vision techniques can expedite the detection of bacterial growth in wastewater treatment plants and alleviate some of the shortcomings associated with traditional detection methods. In recent years, researchers capitalized on this potential by developing segmentation algorithms that were specifically tailored to identify the overgrowth of filamentous bacteria and the risk of sludge bulking. This study investigated the optimization of an artificial intelligence (AI) segmentation model in terms of accuracy metrics and computational requirements. Specifically, three model variables were tested, including training sample size, image resolution, and number of training epochs. The results indicated that larger sample sizes resulted in higher output accuracy up to a certain limit (300 images), beyond which no significant improvements were observed. High image resolution (788 × 530) provided more details for the deep learning model to detect the fine edges between bacteria albeit with significant additional computational requirements. The addition of more training epochs resulted in a minor increase in segmentation accuracy, particularly for thin interconnected filamentous bacteria. Overall, high resolution and epochs did not have a major effect when the sample size was large (300 and 500 images). The findings highlight the optimal balance between model accuracy and computational demands, emphasizing the importance of prioritizing diverse training samples with sufficient sample size. This approach is critical for large-scale implementation, as it enhances the potential of AI to deliver timely and accurate predictions, leading to early warnings of wastewater treatment issues such as sludge bulking.
Collapse
Affiliation(s)
- Sama Al-Ani
- Department of Civil Engineering, McMaster University, Hamilton, ON, L8S 4L8, Canada
| | - Hui Guo
- Veolia Water Technologies & Solutions, Oakville, ON, L6M 4J4, Canada
| | - Sheila Fyfe
- Veolia Water Technologies & Solutions, Oakville, ON, L6M 4J4, Canada
| | - Zebo Long
- Veolia Water Technologies & Solutions, Oakville, ON, L6M 4J4, Canada
| | - Sylvain Donnaz
- Veolia Water Technologies & Solutions, Oakville, ON, L6M 4J4, Canada
| | - Younggy Kim
- Department of Civil Engineering, McMaster University, Hamilton, ON, L8S 4L8, Canada.
| |
Collapse
|
2
|
Chowa SS, Bhuiyan MRI, Payel IJ, Karim A, Khan IU, Montaha S, Hasan MZ, Jonkman M, Azam S. A Low Complexity Efficient Deep Learning Model for Automated Retinal Disease Diagnosis. JOURNAL OF HEALTHCARE INFORMATICS RESEARCH 2025; 9:1-40. [PMID: 39897099 PMCID: PMC11782765 DOI: 10.1007/s41666-024-00182-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 08/08/2024] [Accepted: 12/17/2024] [Indexed: 02/04/2025]
Abstract
The identification and early treatment of retinal disease can help to prevent loss of vision. Early diagnosis allows a greater range of treatment options and results in better outcomes. Optical coherence tomography (OCT) is a technology used by ophthalmologists to detect and diagnose certain eye conditions. In this paper, human retinal OCT images are classified into four classes using deep learning. Several image preprocessing techniques are employed to enhance the image quality. An augmentation technique, called generative adversarial network (GAN), is utilized in the Drusen and DME classes to address data imbalance issues, resulting in a total of 130,649 images. A lightweight optimized compact convolutional transformers (OCCT) model is developed by conducting an ablation study on the initial CCT model for categorizing retinal conditions. The proposed OCCT model is compared with two transformer-based models: vision Transformer (ViT) and Swin Transformer. The models are trained and evaluated with 32 × 32 sized images of the GAN-generated enhanced dataset. Additionally, eight transfer learning models are presented with the same input images to compare their performance with the OCCT model. The proposed model's stability is assessed by decreasing the number of training images and evaluating the performance. The OCCT model's accuracy is 97.09%, and it outperforms the two transformer models. The result further indicates that the OCCT model sustains its performance, even if the number of images is reduced.
Collapse
Affiliation(s)
- Sadia Sultana Chowa
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka-1341, Bangladesh
| | - Md. Rahad Islam Bhuiyan
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka-1341, Bangladesh
| | - Israt Jahan Payel
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka-1341, Bangladesh
| | - Asif Karim
- Faculty of Science and Technology, Charles Darwin University, Casuarina, NT 0909 Australia
| | - Inam Ullah Khan
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka-1341, Bangladesh
| | - Sidratul Montaha
- Department of Computer Science, University of Calgary, Calgary, AB T2N 1N4 Canada
| | - Md. Zahid Hasan
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka-1341, Bangladesh
| | - Mirjam Jonkman
- Faculty of Science and Technology, Charles Darwin University, Casuarina, NT 0909 Australia
| | - Sami Azam
- Faculty of Science and Technology, Charles Darwin University, Casuarina, NT 0909 Australia
| |
Collapse
|
3
|
Gao Y, Vali M. Combination of Deep and Statistical Features of the Tissue of Pathology Images to Classify and Diagnose the Degree of Malignancy of Prostate Cancer. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024:10.1007/s10278-024-01363-9. [PMID: 39663318 DOI: 10.1007/s10278-024-01363-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2024] [Revised: 11/19/2024] [Accepted: 11/29/2024] [Indexed: 12/13/2024]
Abstract
Prostate cancer is one of the most prevalent male-specific diseases, where early and accurate diagnosis is essential for effective treatment and preventing disease progression. Assessing disease severity involves analyzing histological tissue samples, which are graded from 1 (healthy) to 5 (severely malignant) based on pathological features. However, traditional manual grading is labor-intensive and prone to variability. This study addresses the challenge of automating prostate cancer classification by proposing a novel histological grade analysis approach. The method integrates the gray-level co-occurrence matrix (GLCM) for extracting texture features with Haar wavelet modification to enhance feature quality. A convolutional neural network (CNN) is then employed for robust classification. The proposed method was evaluated using statistical and performance metrics, achieving an average accuracy of 97.3%, a precision of 98%, and an AUC of 0.95. These results underscore the effectiveness of the approach in accurately categorizing prostate tissue grades. This study demonstrates the potential of automated classification methods to support pathologists, enhance diagnostic precision, and improve clinical outcomes in prostate cancer care.
Collapse
Affiliation(s)
- Yan Gao
- School of Electrical and Mechanical Engineering, Xuchang University, Xuchang, 461000, Henan, China.
| | - Mahsa Vali
- Department of Electrical and Computer Engineering, Isfahan University of Technology, Isfahan, 84156-83111, Iran
| |
Collapse
|
4
|
Kang C, Lo JE, Zhang H, Ng SM, Lin JC, Scott IU, Kalpathy-Cramer J, Liu SHA, Greenberg PB. Artificial intelligence for diagnosing exudative age-related macular degeneration. Cochrane Database Syst Rev 2024; 10:CD015522. [PMID: 39417312 PMCID: PMC11483348 DOI: 10.1002/14651858.cd015522.pub2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/19/2024]
Abstract
BACKGROUND Age-related macular degeneration (AMD) is a retinal disorder characterized by central retinal (macular) damage. Approximately 10% to 20% of non-exudative AMD cases progress to the exudative form, which may result in rapid deterioration of central vision. Individuals with exudative AMD (eAMD) need prompt consultation with retinal specialists to minimize the risk and extent of vision loss. Traditional methods of diagnosing ophthalmic disease rely on clinical evaluation and multiple imaging techniques, which can be resource-consuming. Tests leveraging artificial intelligence (AI) hold the promise of automatically identifying and categorizing pathological features, enabling the timely diagnosis and treatment of eAMD. OBJECTIVES To determine the diagnostic accuracy of artificial intelligence (AI) as a triaging tool for exudative age-related macular degeneration (eAMD). SEARCH METHODS We searched CENTRAL, MEDLINE, Embase, three clinical trials registries, and Data Archiving and Networked Services (DANS) for gray literature. We did not restrict searches by language or publication date. The date of the last search was April 2024. SELECTION CRITERIA Included studies compared the test performance of algorithms with that of human readers to detect eAMD on retinal images collected from people with AMD who were evaluated at eye clinics in community or academic medical centers, and who were not receiving treatment for eAMD when the images were taken. We included algorithms that were either internally or externally validated or both. DATA COLLECTION AND ANALYSIS Pairs of review authors independently extracted data and assessed study quality using the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool with revised signaling questions. For studies that reported more than one set of performance results, we extracted only one set of diagnostic accuracy data per study based on the last development stage or the optimal algorithm as indicated by the study authors. For two-class algorithms, we collected data from the 2x2 table whenever feasible. For multi-class algorithms, we first consolidated data from all classes other than eAMD before constructing the corresponding 2x2 tables. Assuming a common positivity threshold applied by the included studies, we chose random-effects, bivariate logistic models to estimate summary sensitivity and specificity as the primary performance metrics. MAIN RESULTS We identified 36 eligible studies that reported 40 sets of algorithm performance data, encompassing over 16,000 participants and 62,000 images. We included 28 studies (78%) that reported 31 algorithms with performance data in the meta-analysis. The remaining nine studies (25%) reported eight algorithms that lacked usable performance data; we reported them in the qualitative synthesis. Study characteristics and risk of bias Most studies were conducted in Asia, followed by Europe, the USA, and collaborative efforts spanning multiple countries. Most studies identified study participants from the hospital setting, while others used retinal images from public repositories; a few studies did not specify image sources. Based on four of the 36 studies reporting demographic information, the age of the study participants ranged from 62 to 82 years. The included algorithms used various retinal image types as model input, such as optical coherence tomography (OCT) images (N = 15), fundus images (N = 6), and multi-modal imaging (N = 7). The predominant core method used was deep neural networks. All studies that reported externally validated algorithms were at high risk of bias mainly due to potential selection bias from either a two-gate design or the inappropriate exclusion of potentially eligible retinal images (or participants). Findings Only three of the 40 included algorithms were externally validated (7.5%, 3/40). The summary sensitivity and specificity were 0.94 (95% confidence interval (CI) 0.90 to 0.97) and 0.99 (95% CI 0.76 to 1.00), respectively, when compared to human graders (3 studies; 27,872 images; low-certainty evidence). The prevalence of images with eAMD ranged from 0.3% to 49%. Twenty-eight algorithms were reportedly either internally validated (20%, 8/40) or tested on a development set (50%, 20/40); the pooled sensitivity and specificity were 0.93 (95% CI 0.89 to 0.96) and 0.96 (95% CI 0.94 to 0.98), respectively, when compared to human graders (28 studies; 33,409 images; low-certainty evidence). We did not identify significant sources of heterogeneity among these 28 algorithms. Although algorithms using OCT images appeared more homogeneous and had the highest summary specificity (0.97, 95% CI 0.93 to 0.98), they were not superior to algorithms using fundus images alone (0.94, 95% CI 0.89 to 0.97) or multimodal imaging (0.96, 95% CI 0.88 to 0.99; P for meta-regression = 0.239). The median prevalence of images with eAMD was 30% (interquartile range [IQR] 22% to 39%). We did not include eight studies that described nine algorithms (one study reported two sets of algorithm results) to distinguish eAMD from normal images, images of other AMD, or other non-AMD retinal lesions in the meta-analysis. Five of these algorithms were generally based on smaller datasets (range 21 to 218 participants per study) yet with a higher prevalence of eAMD images (range 33% to 66%). Relative to human graders, the reported sensitivity in these studies ranged from 0.95 and 0.97, while the specificity ranged from 0.94 to 0.99. Similarly, using small datasets (range 46 to 106), an additional four algorithms for detecting eAMD from other retinal lesions showed high sensitivity (range 0.96 to 1.00) and specificity (range 0.77 to 1.00). AUTHORS' CONCLUSIONS Low- to very low-certainty evidence suggests that an algorithm-based test may correctly identify most individuals with eAMD without increasing unnecessary referrals (false positives) in either the primary or the specialty care settings. There were significant concerns for applying the review findings due to variations in the eAMD prevalence in the included studies. In addition, among the included algorithm-based tests, diagnostic accuracy estimates were at risk of bias due to study participants not reflecting real-world characteristics, inadequate model validation, and the likelihood of selective results reporting. Limited quality and quantity of externally validated algorithms highlighted the need for high-certainty evidence. This evidence will require a standardized definition for eAMD on different imaging modalities and external validation of the algorithm to assess generalizability.
Collapse
Affiliation(s)
- Chaerim Kang
- Division of Ophthalmology, Brown University, Providence, RI, USA
| | - Jui-En Lo
- Department of Internal Medicine, MetroHealth Medical Center/Case Western Reserve University, Cleveland, USA
| | - Helen Zhang
- Program in Liberal Medical Education, Brown University, Providence, RI, USA
| | - Sueko M Ng
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
| | - John C Lin
- Department of Medicine, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA
| | - Ingrid U Scott
- Department of Ophthalmology and Public Health Sciences, Penn State College of Medicine, Hershey, PA, USA
| | | | - Su-Hsun Alison Liu
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
- Department of Epidemiology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
| | - Paul B Greenberg
- Division of Ophthalmology, Brown University, Providence, RI, USA
- Section of Ophthalmology, VA Providence Healthcare System, Providence, RI, USA
| |
Collapse
|
5
|
Chen Z, Zhang Y, Zhou Z, Wang L, Zhang H, Wang P, Xu J. An efficient ANN SoC for detecting Alzheimer's disease based on recurrent computing. Comput Biol Med 2024; 181:108993. [PMID: 39173486 DOI: 10.1016/j.compbiomed.2024.108993] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2024] [Revised: 05/22/2024] [Accepted: 08/02/2024] [Indexed: 08/24/2024]
Abstract
Alzheimer's Disease (AD) is an irreversible, degenerative condition that, while incurable, can have its progression slowed or impeded. While there are numerous methods utilizing neural networks for AD detection, there is a scarcity of High-performance AD detection chips. Moreover, excessively complex neural networks are not conducive to on-chip implementation and clinical applications. This study addresses the challenges of high misdiagnosis rates and significant hardware costs inherent in traditional AD detection techniques. A novel and efficient AD detection framework based on a recurrent computational strategy is proposed. The framework harnesses an Artificial Neural Network (ANN) embedded within a System on Chip (SoC) to perform sophisticated Electroencephalogram (EEG) analysis. The approach began by employing a reduced IEEE754 single-precision encoding method to hardware-encode the preprocessed EEG data, thereby minimizing the memory storage area. Next, data remapping techniques were utilized to ensure the continuity of the input data read addresses and reduce the memory access pressure during neural network computations. Subsequently, hierarchical and Processing Element (PE) reuse technologies were leveraged to perform the multiply-accumulate operations of the ANN. Finally, a step function was chosen to establish binary classification circuits dedicated to AD detection. Experimental results indicate that the optimized SoC achieves a 70 % reduction in area and a 50 % reduction in power consumption compared to traditional designs. For various neural network models, the detection model proposed in this paper incurs less overhead, with a training speed 3 to 4 times faster than other traditional models, and a high accuracy rate of 98.53 %.
Collapse
Affiliation(s)
- Zhikang Chen
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, 315211, Zhejiang, China.
| | - Yuejun Zhang
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, 315211, Zhejiang, China.
| | - Ziyu Zhou
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, 315211, Zhejiang, China.
| | - Lixun Wang
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, 315211, Zhejiang, China.
| | - Huihong Zhang
- Faculty of Electrical Engineering and Computer Science, Ningbo University, Ningbo, 315211, Zhejiang, China.
| | - Pengjun Wang
- Electrical and Electronic Engineering, Wenzhou University, Wenzhou, 325035, Zhejiang, China.
| | - Jinyan Xu
- Department of Neurology, The First Affiliated Hospital of Ningbo University, Ningbo, 315020, China.
| |
Collapse
|
6
|
Hossain MI, Zamzmi G, Mouton P, Sun Y, Goldgof D. Explainable AI (XAI) for Neonatal Pain Assessment via Influence Function Modification. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-5. [PMID: 40039767 DOI: 10.1109/embc53108.2024.10782892] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
As machine learning increasingly plays a crucial role in various medical applications, the need for improved explainability of these complex, often opaque models becomes more urgent. Influence functions have emerged as a critical method for explaining these black-box models. Influence functions measure how much each training sample affects a model's predictions on test data, with previous research indicating that the most influential training samples usually exhibit a high degree of semantic similarity to the test point. Building on this concept, we propose a novel approach that modifies the influence function for more precise influence estimations. This involves adding a new weighting factor to the influence function based on the similarity of the test and training data. We employ cosine similarity, Euclidean distance, and the structural similarity index to calculate this weight. The modified influence method is evaluated on a neonatal pain assessment model to explain the decision, revealing excellent performance in identifying influential training points compared to the baseline method. These results show the effectiveness of the proposed approach in elucidating the decision-making process.
Collapse
|
7
|
Zhao L, Zhang Z. A improved pooling method for convolutional neural networks. Sci Rep 2024; 14:1589. [PMID: 38238357 PMCID: PMC10796389 DOI: 10.1038/s41598-024-51258-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2023] [Accepted: 01/02/2024] [Indexed: 01/22/2024] Open
Abstract
The pooling layer in convolutional neural networks plays a crucial role in reducing spatial dimensions, and improving computational efficiency. However, standard pooling operations such as max pooling or average pooling are not suitable for all applications and data types. Therefore, developing custom pooling layers that can adaptively learn and extract relevant features from specific datasets is of great significance. In this paper, we propose a novel approach to design and implement customizable pooling layers to enhance feature extraction capabilities in CNNs. The proposed T-Max-Avg pooling layer incorporates a threshold parameter T, which selects the K highest interacting pixels as specified, allowing it to control whether the output features of the input data are based on the maximum values or weighted averages. By learning the optimal pooling strategy during training, our custom pooling layer can effectively capture and represent discriminative information in the input data, thereby improving classification performance. Experimental results show that the proposed T-Max-Avg pooling layer achieves good performance on three different datasets. When compared to LeNet-5 model with average pooling, max pooling, and Avg-TopK methods, the T-Max-Avg pooling method achieves the highest accuracy on CIFAR-10, CIFAR-100, and MNIST datasets.
Collapse
Affiliation(s)
- Lei Zhao
- School of Electronics and Information Engineering, Lanzhou Jiaotong University, Lanzhou, 730070, China
| | - Zhonglin Zhang
- School of Electronics and Information Engineering, Lanzhou Jiaotong University, Lanzhou, 730070, China.
| |
Collapse
|
8
|
Vishnu Priyan S, Vinod Kumar R, Moorthy C, Nishok VS. A fusion of deep neural networks and game theory for retinal disease diagnosis with OCT images. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:1011-1039. [PMID: 38759091 DOI: 10.3233/xst-240027] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/19/2024]
Abstract
Retinal disorders pose a serious threat to world healthcare because they frequently result in visual loss or impairment. For retinal disorders to be diagnosed precisely, treated individually, and detected early, deep learning is a necessary subset of artificial intelligence. This paper provides a complete approach to improve the accuracy and reliability of retinal disease identification using images from OCT (Retinal Optical Coherence Tomography). The Hybrid Model GIGT, which combines Generative Adversarial Networks (GANs), Inception, and Game Theory, is a novel method for diagnosing retinal diseases using OCT pictures. This technique, which is carried out in Python, includes preprocessing images, feature extraction, GAN classification, and a game-theoretic examination. Resizing, grayscale conversion, noise reduction using Gaussian filters, contrast enhancement using Contrast Limiting Adaptive Histogram Equalization (CLAHE), and edge recognition via the Canny technique are all part of the picture preparation step. These procedures set up the OCT pictures for efficient analysis. The Inception model is used for feature extraction, which enables the extraction of discriminative characteristics from the previously processed pictures. GANs are used for classification, which improves accuracy and resilience by adding a strategic and dynamic aspect to the diagnostic process. Additionally, a game-theoretic analysis is utilized to evaluate the security and dependability of the model in the face of hostile attacks. Strategic analysis and deep learning work together to provide a potent diagnostic tool. This suggested model's remarkable 98.2% accuracy rate shows how this method has the potential to improve the detection of retinal diseases, improve patient outcomes, and address the worldwide issue of visual impairment.
Collapse
Affiliation(s)
- S Vishnu Priyan
- Department of Biomedical Engineering, Kings Engineering College, Chennai, India
| | - R Vinod Kumar
- Department of Electronics and Communication Engineering, Saveetha Engineering College, Chennai, India
| | - C Moorthy
- Dr. Mahalingam College of Engineering and Technology, Pollachi, India
| | - V S Nishok
- Department of Electronics and Communication Engineering, Akshaya College of Engineering and Technology, Coimbatore, India
| |
Collapse
|
9
|
Kourounis G, Elmahmudi AA, Thomson B, Hunter J, Ugail H, Wilson C. Computer image analysis with artificial intelligence: a practical introduction to convolutional neural networks for medical professionals. Postgrad Med J 2023; 99:1287-1294. [PMID: 37794609 PMCID: PMC10658730 DOI: 10.1093/postmj/qgad095] [Citation(s) in RCA: 10] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2023] [Revised: 09/06/2023] [Accepted: 09/13/2023] [Indexed: 10/06/2023]
Abstract
Artificial intelligence tools, particularly convolutional neural networks (CNNs), are transforming healthcare by enhancing predictive, diagnostic, and decision-making capabilities. This review provides an accessible and practical explanation of CNNs for clinicians and highlights their relevance in medical image analysis. CNNs have shown themselves to be exceptionally useful in computer vision, a field that enables machines to 'see' and interpret visual data. Understanding how these models work can help clinicians leverage their full potential, especially as artificial intelligence continues to evolve and integrate into healthcare. CNNs have already demonstrated their efficacy in diverse medical fields, including radiology, histopathology, and medical photography. In radiology, CNNs have been used to automate the assessment of conditions such as pneumonia, pulmonary embolism, and rectal cancer. In histopathology, CNNs have been used to assess and classify colorectal polyps, gastric epithelial tumours, as well as assist in the assessment of multiple malignancies. In medical photography, CNNs have been used to assess retinal diseases and skin conditions, and to detect gastric and colorectal polyps during endoscopic procedures. In surgical laparoscopy, they may provide intraoperative assistance to surgeons, helping interpret surgical anatomy and demonstrate safe dissection zones. The integration of CNNs into medical image analysis promises to enhance diagnostic accuracy, streamline workflow efficiency, and expand access to expert-level image analysis, contributing to the ultimate goal of delivering further improvements in patient and healthcare outcomes.
Collapse
Affiliation(s)
- Georgios Kourounis
- NIHR Blood and Transplant Research Unit, Newcastle University and Cambridge University, Newcastle upon Tyne, NE1 7RU, United Kingdom
- Institute of Transplantation, The Freeman Hospital, Newcastle upon Tyne, NE7 7DN, United Kingdom
| | - Ali Ahmed Elmahmudi
- Faculty of Engineering and Informatics, Bradford University, Bradford, BD7 1DP, United Kingdom
| | - Brian Thomson
- Faculty of Engineering and Informatics, Bradford University, Bradford, BD7 1DP, United Kingdom
| | - James Hunter
- Nuffield Department of Surgical Sciences, University of Oxford, Oxford, OX3 9DU, United Kingdom
| | - Hassan Ugail
- Faculty of Engineering and Informatics, Bradford University, Bradford, BD7 1DP, United Kingdom
| | - Colin Wilson
- NIHR Blood and Transplant Research Unit, Newcastle University and Cambridge University, Newcastle upon Tyne, NE1 7RU, United Kingdom
- Institute of Transplantation, The Freeman Hospital, Newcastle upon Tyne, NE7 7DN, United Kingdom
| |
Collapse
|
10
|
Abbas Q, Albathan M, Altameem A, Almakki RS, Hussain A. Deep-Ocular: Improved Transfer Learning Architecture Using Self-Attention and Dense Layers for Recognition of Ocular Diseases. Diagnostics (Basel) 2023; 13:3165. [PMID: 37891986 PMCID: PMC10605427 DOI: 10.3390/diagnostics13203165] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Revised: 10/07/2023] [Accepted: 10/09/2023] [Indexed: 10/29/2023] Open
Abstract
It is difficult for clinicians or less-experienced ophthalmologists to detect early eye-related diseases. By hand, eye disease diagnosis is labor-intensive, prone to mistakes, and challenging because of the variety of ocular diseases such as glaucoma (GA), diabetic retinopathy (DR), cataract (CT), and normal eye-related diseases (NL). An automated ocular disease detection system with computer-aided diagnosis (CAD) tools is required to recognize eye-related diseases. Nowadays, deep learning (DL) algorithms enhance the classification results of retinograph images. To address these issues, we developed an intelligent detection system based on retinal fundus images. To create this system, we used ODIR and RFMiD datasets, which included various retinographics of distinct classes of the fundus, using cutting-edge image classification algorithms like ensemble-based transfer learning. In this paper, we suggest a three-step hybrid ensemble model that combines a classifier, a feature extractor, and a feature selector. The original image features are first extracted using a pre-trained AlexNet model with an enhanced structure. The improved AlexNet (iAlexNet) architecture with attention and dense layers offers enhanced feature extraction, task adaptability, interpretability, and potential accuracy benefits compared to other transfer learning architectures, making it particularly suited for tasks like retinograph classification. The extracted features are then selected using the ReliefF method, and then the most crucial elements are chosen to minimize the feature dimension. Finally, an XgBoost classifier offers classification outcomes based on the desired features. These classifications represent different ocular illnesses. We utilized data augmentation techniques to control class imbalance issues. The deep-ocular model, based mainly on the AlexNet-ReliefF-XgBoost model, achieves an accuracy of 95.13%. The results indicate the proposed ensemble model can assist dermatologists in making early decisions for the diagnosing and screening of eye-related diseases.
Collapse
Affiliation(s)
- Qaisar Abbas
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (M.A.); (A.A.); (R.S.A.)
| | - Mubarak Albathan
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (M.A.); (A.A.); (R.S.A.)
| | - Abdullah Altameem
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (M.A.); (A.A.); (R.S.A.)
| | - Riyad Saleh Almakki
- College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia; (M.A.); (A.A.); (R.S.A.)
| | - Ayyaz Hussain
- Department of Computer Science, Quaid-i-Azam University, Islamabad 44000, Pakistan;
| |
Collapse
|
11
|
Azeem M, Javaid S, Khalil RA, Fahim H, Althobaiti T, Alsharif N, Saeed N. Neural Networks for the Detection of COVID-19 and Other Diseases: Prospects and Challenges. Bioengineering (Basel) 2023; 10:850. [PMID: 37508877 PMCID: PMC10416184 DOI: 10.3390/bioengineering10070850] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2023] [Revised: 07/09/2023] [Accepted: 07/12/2023] [Indexed: 07/30/2023] Open
Abstract
Artificial neural networks (ANNs) ability to learn, correct errors, and transform a large amount of raw data into beneficial medical decisions for treatment and care has increased in popularity for enhanced patient safety and quality of care. Therefore, this paper reviews the critical role of ANNs in providing valuable insights for patients' healthcare decisions and efficient disease diagnosis. We study different types of ANNs in the existing literature that advance ANNs' adaptation for complex applications. Specifically, we investigate ANNs' advances for predicting viral, cancer, skin, and COVID-19 diseases. Furthermore, we propose a deep convolutional neural network (CNN) model called ConXNet, based on chest radiography images, to improve the detection accuracy of COVID-19 disease. ConXNet is trained and tested using a chest radiography image dataset obtained from Kaggle, achieving more than 97% accuracy and 98% precision, which is better than other existing state-of-the-art models, such as DeTraC, U-Net, COVID MTNet, and COVID-Net, having 93.1%, 94.10%, 84.76%, and 90% accuracy and 94%, 95%, 85%, and 92% precision, respectively. The results show that the ConXNet model performed significantly well for a relatively large dataset compared with the aforementioned models. Moreover, the ConXNet model reduces the time complexity by using dropout layers and batch normalization techniques. Finally, we highlight future research directions and challenges, such as the complexity of the algorithms, insufficient available data, privacy and security, and integration of biosensing with ANNs. These research directions require considerable attention for improving the scope of ANNs for medical diagnostic and treatment applications.
Collapse
Affiliation(s)
- Muhammad Azeem
- School of Science, Engineering & Environment, University of Salford, Manchester M5 4WT, UK;
| | - Shumaila Javaid
- Department of Control Science and Engineering, College of Electronics and Information Engineering, Tongji University, Shanghai 201804, China; (S.J.); (H.F.)
| | - Ruhul Amin Khalil
- Department of Electrical Engineering, University of Engineering and Technology, Peshawar 25120, Pakistan;
- Department of Electrical and Communication Engineering, United Arab Emirates University (UAEU), Al-Ain 15551, United Arab Emirates
| | - Hamza Fahim
- Department of Control Science and Engineering, College of Electronics and Information Engineering, Tongji University, Shanghai 201804, China; (S.J.); (H.F.)
| | - Turke Althobaiti
- Department of Computer Science, Faculty of Science, Northern Border University, Arar 73222, Saudi Arabia;
| | - Nasser Alsharif
- Department of Administrative and Financial Sciences, Ranyah University Collage, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia;
| | - Nasir Saeed
- Department of Electrical and Communication Engineering, United Arab Emirates University (UAEU), Al-Ain 15551, United Arab Emirates
| |
Collapse
|
12
|
Feng H, Chen J, Zhang Z, Lou Y, Zhang S, Yang W. A bibliometric analysis of artificial intelligence applications in macular edema: exploring research hotspots and Frontiers. Front Cell Dev Biol 2023; 11:1174936. [PMID: 37255600 PMCID: PMC10225517 DOI: 10.3389/fcell.2023.1174936] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2023] [Accepted: 05/02/2023] [Indexed: 06/01/2023] Open
Abstract
Background: Artificial intelligence (AI) is used in ophthalmological disease screening and diagnostics, medical image diagnostics, and predicting late-disease progression rates. We reviewed all AI publications associated with macular edema (ME) research Between 2011 and 2022 and performed modeling, quantitative, and qualitative investigations. Methods: On 1st February 2023, we screened the Web of Science Core Collection for AI applications related to ME, from which 297 studies were identified and analyzed (2011-2022). We collected information on: publications, institutions, country/region, keywords, journal name, references, and research hotspots. Literature clustering networks and Frontier knowledge bases were investigated using bibliometrix-BiblioShiny, VOSviewer, and CiteSpace bibliometric platforms. We used the R "bibliometrix" package to synopsize our observations, enumerate keywords, visualize collaboration networks between countries/regions, and generate a topic trends plot. VOSviewer was used to examine cooperation between institutions and identify citation relationships between journals. We used CiteSpace to identify clustering keywords over the timeline and identify keywords with the strongest citation bursts. Results: In total, 47 countries published AI studies related to ME; the United States had the highest H-index, thus the greatest influence. China and the United States cooperated most closely between all countries. Also, 613 institutions generated publications - the Medical University of Vienna had the highest number of studies. This publication record and H-index meant the university was the most influential in the ME field. Reference clusters were also categorized into 10 headings: retinal Optical Coherence Tomography (OCT) fluid detection, convolutional network models, deep learning (DL)-based single-shot predictions, retinal vascular disease, diabetic retinopathy (DR), convolutional neural networks (CNNs), automated macular pathology diagnosis, dry age-related macular degeneration (DARMD), class weight, and advanced DL architecture systems. Frontier keywords were represented by diabetic macular edema (DME) (2021-2022). Conclusion: Our review of the AI-related ME literature was comprehensive, systematic, and objective, and identified future trends and current hotspots. With increased DL outputs, the ME research focus has gradually shifted from manual ME examinations to automatic ME detection and associated symptoms. In this review, we present a comprehensive and dynamic overview of AI in ME and identify future research areas.
Collapse
Affiliation(s)
- Haiwen Feng
- Department of Software Engineering, School of Software, Shenyang University of Technology, Shenyang, Liaoning, China
| | - Jiaqi Chen
- Department of Software Engineering, School of Software, Shenyang University of Technology, Shenyang, Liaoning, China
| | - Zhichang Zhang
- Department of Computer, School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| | - Yan Lou
- Department of Computer, School of Intelligent Medicine, China Medical University, Shenyang, Liaoning, China
| | - Shaochong Zhang
- Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China
| | - Weihua Yang
- Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China
| |
Collapse
|
13
|
Moradi M, Chen Y, Du X, Seddon JM. Deep ensemble learning for automated non-advanced AMD classification using optimized retinal layer segmentation and SD-OCT scans. Comput Biol Med 2023; 154:106512. [PMID: 36701964 DOI: 10.1016/j.compbiomed.2022.106512] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Revised: 11/30/2022] [Accepted: 12/31/2022] [Indexed: 01/11/2023]
Abstract
BACKGROUND Accurate retinal layer segmentation in optical coherence tomography (OCT) images is crucial for quantitatively analyzing age-related macular degeneration (AMD) and monitoring its progression. However, previous retinal segmentation models depend on experienced experts and manually annotating retinal layers is time-consuming. On the other hand, accuracy of AMD diagnosis is directly related to the segmentation model's performance. To address these issues, we aimed to improve AMD detection using optimized retinal layer segmentation and deep ensemble learning. METHOD We integrated a graph-cut algorithm with a cubic spline to automatically annotate 11 retinal boundaries. The refined images were fed into a deep ensemble mechanism that combined a Bagged Tree and end-to-end deep learning classifiers. We tested the developed deep ensemble model on internal and external datasets. RESULTS The total error rates for our segmentation model using the boundary refinement approach was significantly lower than OCT Explorer segmentations (1.7% vs. 7.8%, p-value = 0.03). We utilized the refinement approach to quantify 169 imaging features using Zeiss SD-OCT volume scans. The presence of drusen and thickness of total retina, neurosensory retina, and ellipsoid zone to inner-outer segment (EZ-ISOS) thickness had higher contributions to AMD classification compared to other features. The developed ensemble learning model obtained a higher diagnostic accuracy in a shorter time compared with two human graders. The area under the curve (AUC) for normal vs. early AMD was 99.4%. CONCLUSION Testing results showed that the developed framework is repeatable and effective as a potentially valuable tool in retinal imaging research.
Collapse
Affiliation(s)
- Mousa Moradi
- Department of Biomedical Engineering, University of Massachusetts, Amherst, MA, United States
| | - Yu Chen
- Department of Biomedical Engineering, University of Massachusetts, Amherst, MA, United States.
| | - Xian Du
- Department of Mechanical and Industrial Engineering, University of Massachusetts, Amherst, MA, United States.
| | - Johanna M Seddon
- Department of Ophthalmology & Visual Sciences, University of Massachusetts Chan Medical School, Worcester, MA, United States.
| |
Collapse
|
14
|
Susheel Kumar K, Pratap Singh N. Identification of retinal diseases based on retinal blood vessel segmentation using Dagum PDF and feature-based machine learning. THE IMAGING SCIENCE JOURNAL 2023. [DOI: 10.1080/13682199.2023.2183319] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2023]
Affiliation(s)
- K. Susheel Kumar
- Department of Computer science and Engineering, National Institute of Technology, Hamirpur, India
- Department of Computer Science and Engineering, Gandhi Institute of Technology and Management, Bengaluru, India
| | - Nagendra Pratap Singh
- Department of Computer science and Engineering, National Institute of Technology, Hamirpur, India
| |
Collapse
|
15
|
Classification of Retinal Diseases in Optical Coherence Tomography Images Using Artificial Intelligence and Firefly Algorithm. Diagnostics (Basel) 2023; 13:diagnostics13030433. [PMID: 36766537 PMCID: PMC9914873 DOI: 10.3390/diagnostics13030433] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2022] [Revised: 01/20/2023] [Accepted: 01/24/2023] [Indexed: 01/27/2023] Open
Abstract
In recent years, the number of studies for the automatic diagnosis of biomedical diseases has increased. Many of these studies have used Deep Learning, which gives extremely good results but requires a vast amount of data and computing load. If the processor is of insufficient quality, this takes time and places an excessive load on the processor. On the other hand, Machine Learning is faster than Deep Learning and does not have a much-needed computing load, but it does not provide as high an accuracy value as Deep Learning. Therefore, our goal is to develop a hybrid system that provides a high accuracy value, while requiring a smaller computing load and less time to diagnose biomedical diseases such as the retinal diseases we chose for this study. For this purpose, first, retinal layer extraction was conducted through image preprocessing. Then, traditional feature extractors were combined with pre-trained Deep Learning feature extractors. To select the best features, we used the Firefly algorithm. In the end, multiple binary classifications were conducted instead of multiclass classification with Machine Learning classifiers. Two public datasets were used in this study. The first dataset had a mean accuracy of 0.957, and the second dataset had a mean accuracy of 0.954.
Collapse
|
16
|
Asif S, Amjad K. Deep Residual Network for Diagnosis of Retinal Diseases Using Optical Coherence Tomography Images. Interdiscip Sci 2022; 14:906-916. [PMID: 35767116 DOI: 10.1007/s12539-022-00533-z] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2022] [Revised: 06/02/2022] [Accepted: 06/06/2022] [Indexed: 06/15/2023]
Abstract
Diabetic retinopathy occurs due to damage to the blood vessels in the retina, and it is a major health problem in recent years that progresses slowly without recognizable symptoms. Optical coherence tomography (OCT) is a popular and widely used noninvasive imaging modality for the diagnosis of diabetic retinopathy. Accurate and early diagnosis of this disease using OCT images is crucial for the prevention of blindness. In recent years, several deep learning methods have been very successful in automating the process of detecting retinal diseases from OCT images. However, most methods face reliability and interpretability issues. In this study, we propose a deep residual network for the classification of four classes of retinal diseases, namely diabetic macular edema (DME), choroidal neovascularization (CNV), DRUSEN and NORMAL in OCT images. The proposed model is based on the popular architecture called ResNet50, which eliminates the vanishing gradient problem and is pre-trained on large dataset such as ImageNet and trained end-to-end on the publicly available OCT image dataset. We removed the fully connected layer of ResNet50 and placed our new fully connected block on top to improve the classification accuracy and avoid overfitting in the proposed model. The proposed model was trained and evaluated using different performance metrics, including receiver operating characteristic (ROC) curve on a dataset of 84,452 OCT images with expert disease grading as DRUSEN, CNV, DME and NORMAL. The proposed model provides an improved overall classification accuracy of 99.48% with only 5 misclassifications out of 968 test samples and outperforms existing methods on the same dataset. The results show that the proposed model is well suited for the diagnosis of retinal diseases in ophthalmology clinics.
Collapse
Affiliation(s)
- Sohaib Asif
- School of Computer Science and Engineering, Central South University, Changsha, Hunan, China.
| | - Kamran Amjad
- School of Automation, Central South University, Changsha, Hunan, China
| |
Collapse
|
17
|
S J, Kandaswami JA. Localization and Semantic Segmentation of Polyp in an Effort of Early Diagnosis of Colorectal Cancer from Wireless Capsule Endoscopy Images. 2022 SEVENTH INTERNATIONAL CONFERENCE ON PARALLEL, DISTRIBUTED AND GRID COMPUTING (PDGC) 2022; 52:749-754. [DOI: 10.1109/pdgc56933.2022.10053299] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/04/2025]
Affiliation(s)
- Jothiraj S
- SRM Institute Science and Technology,Department of Biomedical Engineering,Chengalpattu,Kattankulathur,India
| | - Jayanthy Anavai Kandaswami
- SRM Institute Science and Technology,Department of Biomedical Engineering,Chengalpattu,Kattankulathur,India
| |
Collapse
|
18
|
Yao Z, Jin T, Mao B, Lu B, Zhang Y, Li S, Chen W. Construction and Multicenter Diagnostic Verification of Intelligent Recognition System for Endoscopic Images From Early Gastric Cancer Based on YOLO-V3 Algorithm. Front Oncol 2022; 12:815951. [PMID: 35145918 PMCID: PMC8822233 DOI: 10.3389/fonc.2022.815951] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2021] [Accepted: 01/05/2022] [Indexed: 12/12/2022] Open
Abstract
Introduction Endoscopy is an important tool for the diagnosis of early gastric cancer. Therefore, a combination of artificial intelligence and endoscopy has the ability to increase the speed and efficiency of early gastric cancer diagnosis. YOU ONLY LOOK ONCE (YOLO) is an advanced object detection depth neural network algorithm that has not been widely used in gastrointestinal image recognition. Objective We developed an artificial intelligence system herein referred to as “EGC-YOLO” for the rapid and accurate diagnosis of endoscopic images from early gastric cancer. Methods More than 40000 gastroscopic images from 1653 patients in Yixing people’s Hospital were used as the training set for the system, while endoscopic images from the other two hospitals were used as external validation test sets. The sensitivity, specificity, positive predictive value, Youden index and ROC curve were analyzed to evaluate detection efficiencies for EGC-YOLO. Results EGC-YOLO was able to diagnose early gastric cancer in the two test sets with a high superiority and efficiency. The accuracy, sensitivity, specificity and positive predictive value for Test Sets 1 and 2 were 85.15% and 86.02%, 85.36% and 83.02%, 84.41% and 92.21%, and 95.22% and 95.65%, respectively. In Test Sets 1 and 2, the corresponding Threshold-values were 0.02, 0.16 and 0.17 at the maximum of the Youden index. An increase in Threshold-values was associated with a downward trend in sensitivity and accuracy, while specificity remained relatively stable at more than 80%. Conclusions The EGC-YOLO system is superior for the efficient, accurate and rapid detection of early gastric cancer lesions. For different data sets, it is important to select the appropriate threshold-value in advance to achieve the best performance of the EGC-YOLO system.
Collapse
Affiliation(s)
- Zhendong Yao
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, Suzhou, China
- Department of Gastroenterology, Yixing People’s Hospital, Yixing, China
| | - Tao Jin
- Department of Gastroenterology, Yixing People’s Hospital, Yixing, China
| | - Boneng Mao
- Department of Gastroenterology, Yixing People’s Hospital, Yixing, China
| | - Bo Lu
- Microsoft Teams Calling Meeting Device of Sharepoint Onedrive eXperience (Teams CMD SOX), Microsoft Ltd Co., Suzhou, China
| | - Yefei Zhang
- Department of Gastroenterology, The Second Affiliated Hospital of Soochow University, Suzhou, China
| | - Sisi Li
- Department of Gastroenterology, Civil Aviation Hospital of Shanghai, Shanghai, China
| | - Weichang Chen
- Department of Gastroenterology, The First Affiliated Hospital of Soochow University, Suzhou, China
- *Correspondence: Weichang Chen,
| |
Collapse
|
19
|
Barua PD, Chan WY, Dogan S, Baygin M, Tuncer T, Ciaccio EJ, Islam N, Cheong KH, Shahid ZS, Acharya UR. Multilevel Deep Feature Generation Framework for Automated Detection of Retinal Abnormalities Using OCT Images. ENTROPY (BASEL, SWITZERLAND) 2021; 23:1651. [PMID: 34945957 PMCID: PMC8700736 DOI: 10.3390/e23121651] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Revised: 11/22/2021] [Accepted: 11/25/2021] [Indexed: 01/04/2023]
Abstract
Optical coherence tomography (OCT) images coupled with many learning techniques have been developed to diagnose retinal disorders. This work aims to develop a novel framework for extracting deep features from 18 pre-trained convolutional neural networks (CNN) and to attain high performance using OCT images. In this work, we have developed a new framework for automated detection of retinal disorders using transfer learning. This model consists of three phases: deep fused and multilevel feature extraction, using 18 pre-trained networks and tent maximal pooling, feature selection with ReliefF, and classification using the optimized classifier. The novelty of this proposed framework is the feature generation using widely used CNNs and to select the most suitable features for classification. The extracted features using our proposed intelligent feature extractor are fed to iterative ReliefF (IRF) to automatically select the best feature vector. The quadratic support vector machine (QSVM) is utilized as a classifier in this work. We have developed our model using two public OCT image datasets, and they are named database 1 (DB1) and database 2 (DB2). The proposed framework can attain 97.40% and 100% classification accuracies using the two OCT datasets, DB1 and DB2, respectively. These results illustrate the success of our model.
Collapse
Affiliation(s)
- Prabal Datta Barua
- School of Management & Enterprise, University of Southern Queensland, Toowoomba, QLD 4350, Australia;
- Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia
- Cogninet Brain Team, Cogninet Australia, Sydney, NSW 2010, Australia
| | - Wai Yee Chan
- University Malaya Research Imaging Centre, Department of Biomedical Imaging, Faculty of Medicine, University of Malaya, Kuala Lumpur 59100, Malaysia;
| | - Sengul Dogan
- Department of Digital Forensics Engineering, College of Technology, Firat University, Elazig 23002, Turkey; (S.D.); (T.T.)
| | - Mehmet Baygin
- Department of Computer Engineering, College of Engineering, Ardahan University, Ardahan 75000, Turkey;
| | - Turker Tuncer
- Department of Digital Forensics Engineering, College of Technology, Firat University, Elazig 23002, Turkey; (S.D.); (T.T.)
| | - Edward J. Ciaccio
- Department of Medicine, Columbia University Irving Medical Center, New York, NY 10032-3784, USA;
| | - Nazrul Islam
- Glaucoma Faculty, Bangladesh Eye Hospital & Institute, Dhaka 1206, Bangladesh;
| | - Kang Hao Cheong
- Science, Mathematics and Technology Cluster, Singapore University of Technology and Design, Singapore 487372, Singapore
| | - Zakia Sultana Shahid
- Department of Ophthalmology, Anwer Khan Modern Medical College, Dhaka 1205, Bangladesh;
| | - U. Rajendra Acharya
- Department of Electronics and Computer Engineering, Ngee Ann Polytechnic, Singapore 599489, Singapore
- Department of Biomedical Engineering, School of Science and Technology, SUSS University, Singapore 129799, Singapore
- Department of Biomedical Informatics and Medical Engineering, Asia University, Taichung 41354, Taiwan
| |
Collapse
|
20
|
Qureshi KN, Alhudhaif A, Ali M, Qureshi MA, Jeon G. Self-assessment and deep learning-based coronavirus detection and medical diagnosis systems for healthcare. MULTIMEDIA SYSTEMS 2021; 28:1439-1448. [PMID: 34511733 PMCID: PMC8421458 DOI: 10.1007/s00530-021-00839-w] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/15/2020] [Accepted: 08/12/2021] [Indexed: 05/17/2023]
Abstract
Coronavirus is one of the serious threat and challenge for existing healthcare systems. Several prevention methods and precautions have been proposed by medical specialists to treat the virus and secure infected patients. Deep learning methods have been adopted for disease detection, especially for medical image classification. In this paper, we proposed a deep learning-based medical image classification for COVID-19 patients namely deep learning model for coronavirus (DLM-COVID-19). The proposed model improves the medical image classification and optimization for better disease diagnosis. This paper also proposes a mobile application for COVID-19 patient detection using a self-assessment test combined with medical expertise and diagnose and prevent the virus using the online system. The proposed deep learning model is evaluated with existing algorithms where it shows better performance in terms of sensitivity, specificity, and accuracy. Whereas the proposed application also helps to overcome the virus risk and spread.
Collapse
Affiliation(s)
| | - Adi Alhudhaif
- Department of Computer Science, College of Computer Engineering and Sciences in Al-kharj, Prince Sattam bin Abdulaziz University, P.O. Box 151, Al-Kharj 11942, Saudi Arabia
| | - Moazam Ali
- Department of Computer Science, Bahria University, Islamabad, Pakistan
| | | | | |
Collapse
|
21
|
A Comparative Analysis of Machine Learning Algorithms to Predict Alzheimer's Disease. JOURNAL OF HEALTHCARE ENGINEERING 2021; 2021:9917919. [PMID: 34336171 PMCID: PMC8289609 DOI: 10.1155/2021/9917919] [Citation(s) in RCA: 20] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/17/2021] [Revised: 05/29/2021] [Accepted: 06/23/2021] [Indexed: 02/06/2023]
Abstract
Alzheimer's disease has been one of the major concerns recently. Around 45 million people are suffering from this disease. Alzheimer's is a degenerative brain disease with an unspecified cause and pathogenesis which primarily affects older people. The main cause of Alzheimer's disease is Dementia, which progressively damages the brain cells. People lost their thinking ability, reading ability, and many more from this disease. A machine learning system can reduce this problem by predicting the disease. The main aim is to recognize Dementia among various patients. This paper represents the result and analysis regarding detecting Dementia from various machine learning models. The Open Access Series of Imaging Studies (OASIS) dataset has been used for the development of the system. The dataset is small, but it has some significant values. The dataset has been analyzed and applied in several machine learning models. Support vector machine, logistic regression, decision tree, and random forest have been used for prediction. First, the system has been run without fine-tuning and then with fine-tuning. Comparing the results, it is found that the support vector machine provides the best results among the models. It has the best accuracy in detecting Dementia among numerous patients. The system is simple and can easily help people by detecting Dementia among them.
Collapse
|