1
|
Rashid PQ, Türker İ. Lung Disease Detection Using U-Net Feature Extractor Cascaded by Graph Convolutional Network. Diagnostics (Basel) 2024; 14:1313. [PMID: 38928728 PMCID: PMC11202625 DOI: 10.3390/diagnostics14121313] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/20/2024] [Revised: 06/17/2024] [Accepted: 06/18/2024] [Indexed: 06/28/2024] Open
Abstract
Computed tomography (CT) scans have recently emerged as a major technique for the fast diagnosis of lung diseases via image classification techniques. In this study, we propose a method for the diagnosis of COVID-19 disease with improved accuracy by utilizing graph convolutional networks (GCN) at various layer formations and kernel sizes to extract features from CT scan images. We apply a U-Net model to aid in segmentation and feature extraction. In contrast with previous research retrieving deep features from convolutional filters and pooling layers, which fail to fully consider the spatial connectivity of the nodes, we employ GCNs for classification and prediction to capture spatial connectivity patterns, which provides a significant association benefit. We handle the extracted deep features to form an adjacency matrix that contains a graph structure and pass it to a GCN along with the original image graph and the largest kernel graph. We combine these graphs to form one block of the graph input and then pass it through a GCN with an additional dropout layer to avoid overfitting. Our findings show that the suggested framework, called the feature-extracted graph convolutional network (FGCN), performs better in identifying lung diseases compared to recently proposed deep learning architectures that are not based on graph representations. The proposed model also outperforms a variety of transfer learning models commonly used for medical diagnosis tasks, highlighting the abstraction potential of the graph representation over traditional methods.
Collapse
Affiliation(s)
| | - İlker Türker
- Department of Computer Engineering, Karabuk University, 78050 Karabuk, Turkey;
| |
Collapse
|
2
|
Huang L, Yao C, Zhang L, Luo S, Ying F, Ying W. Enhancing computer image recognition with improved image algorithms. Sci Rep 2024; 14:13709. [PMID: 38877063 PMCID: PMC11178774 DOI: 10.1038/s41598-024-64193-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Accepted: 06/06/2024] [Indexed: 06/16/2024] Open
Abstract
Advances in computer image recognition have significantly impacted many industries, including healthcare, security and autonomous systems. This paper aims to explore the potential of improving image algorithms to enhance computer image recognition. Specifically, we will focus on regression methods as a means to improve the accuracy and efficiency of identifying images. In this study, we will analyze various regression techniques and their applications in computer image recognition, as well as the resulting performance improvements through detailed examples and data analysis. This paper deals with the problems related to visual image processing in outdoor unstructured environment. Finally, the heterogeneous patterns are converted into the same pattern, and the heterogeneous patterns are extracted from the fusion features of data modes. The simulation results show that the perception ability and recognition ability of outdoor image recognition in complex environment are improved.
Collapse
Affiliation(s)
| | - Cheng Yao
- Zhejiang University, Hangzhou, 310027, China
| | | | - Shijian Luo
- Zhejiang University, Hangzhou, 310027, China
| | - Fangtian Ying
- Macau University of Science and Technology, Macau, 519020, China
| | | |
Collapse
|
3
|
Ji J, Lao Y, Huo L. Convolutional neural network application for supply-demand matching in Zhuang ethnic clothing image classification. Sci Rep 2024; 14:13348. [PMID: 38858436 PMCID: PMC11164973 DOI: 10.1038/s41598-024-64082-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 06/05/2024] [Indexed: 06/12/2024] Open
Abstract
This study aims to design a classification technique suitable for Zhuang ethnic clothing images by integrating the concept of supply-demand matching and convolutional neural networks. Firstly, addressing the complex structure and unique visual style of Zhuang ethnic clothing, this study proposes an image resolution model based on supply-demand matching and convolutional networks. By integrating visual style and label constraints, this model accurately extracts local features. Secondly, the model's effectiveness and resolution performance are analyzed through various performance metrics in experiments. The results indicate a significant improvement in detection accuracy at different annotation points. The model outperforms other comparative methods in pixel accuracy (90.5%), average precision (83.7%), average recall (80.1%), and average F1 score (81.2%). Next, this study introduces a clothing image classification algorithm based on key points and channel attention. Through key point detection and channel attention mechanisms, image features are optimized, enabling accurate classification and attribute prediction of Zhuang ethnic clothing. Experimental results demonstrate a notable enhancement in category classification and attribute prediction, with classification accuracy and recall exceeding 90% in top-k tasks, showcasing outstanding performance. In conclusion, this study provides innovative approaches and effective solutions for deep learning classification of Zhuang ethnic clothing images.
Collapse
Affiliation(s)
- Jingyi Ji
- School of Art, South China University of Technology, Guangzhou, 510641, People's Republic of China
| | - Yonghua Lao
- National Engineering Research Center for Tissue Restoration and Reconstruction, School of Materials Science and Engineering, South China University of Technology, Guangzhou, 510641, People's Republic of China
| | - Lei Huo
- School of Art, South China University of Technology, Guangzhou, 510641, People's Republic of China.
| |
Collapse
|
4
|
Cao Z, Wang K, Wen J, Li C, Wu Y, Wang X, Yu W. Fine-grained image classification on bats using VGG16-CBAM: a practical example with 7 horseshoe bats taxa (CHIROPTERA: Rhinolophidae: Rhinolophus) from Southern China. Front Zool 2024; 21:10. [PMID: 38561769 PMCID: PMC10983684 DOI: 10.1186/s12983-024-00531-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Accepted: 03/18/2024] [Indexed: 04/04/2024] Open
Abstract
BACKGROUND Rapid identification and classification of bats are critical for practical applications. However, species identification of bats is a typically detrimental and time-consuming manual task that depends on taxonomists and well-trained experts. Deep Convolutional Neural Networks (DCNNs) provide a practical approach for the extraction of the visual features and classification of objects, with potential application for bat classification. RESULTS In this study, we investigated the capability of deep learning models to classify 7 horseshoe bat taxa (CHIROPTERA: Rhinolophus) from Southern China. We constructed an image dataset of 879 front, oblique, and lateral targeted facial images of live individuals collected during surveys between 2012 and 2021. All images were taken using a standard photograph protocol and setting aimed at enhancing the effectiveness of the DCNNs classification. The results demonstrated that our customized VGG16-CBAM model achieved up to 92.15% classification accuracy with better performance than other mainstream models. Furthermore, the Grad-CAM visualization reveals that the model pays more attention to the taxonomic key regions in the decision-making process, and these regions are often preferred by bat taxonomists for the classification of horseshoe bats, corroborating the validity of our methods. CONCLUSION Our finding will inspire further research on image-based automatic classification of chiropteran species for early detection and potential application in taxonomy.
Collapse
Affiliation(s)
- Zhong Cao
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou, 510006, China
| | - Kunhui Wang
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou, 510006, China
| | - Jiawei Wen
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou, 510006, China
| | - Chuxian Li
- School of Electronics and Communication Engineering, Guangzhou University, Guangzhou, 510006, China
| | - Yi Wu
- School of Life Sciences, Guangzhou University, Guangzhou, 510006, China
| | - Xiaoyun Wang
- School of Life Sciences, Guangzhou University, Guangzhou, 510006, China.
| | - Wenhua Yu
- School of Life Sciences, Guangzhou University, Guangzhou, 510006, China.
| |
Collapse
|
5
|
Mao M, Va H, Hong M. Video Classification of Cloth Simulations: Deep Learning and Position-Based Dynamics for Stiffness Prediction. SENSORS (BASEL, SWITZERLAND) 2024; 24:549. [PMID: 38257643 PMCID: PMC10820360 DOI: 10.3390/s24020549] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/04/2023] [Revised: 01/04/2024] [Accepted: 01/12/2024] [Indexed: 01/24/2024]
Abstract
In virtual reality, augmented reality, or animation, the goal is to represent the movement of deformable objects in the real world as similar as possible in the virtual world. Therefore, this paper proposed a method to automatically extract cloth stiffness values from video scenes, and then they are applied as material properties for virtual cloth simulation. We propose the use of deep learning (DL) models to tackle this issue. The Transformer model, in combination with pre-trained architectures like DenseNet121, ResNet50, VGG16, and VGG19, stands as a leading choice for video classification tasks. Position-Based Dynamics (PBD) is a computational framework widely used in computer graphics and physics-based simulations for deformable entities, notably cloth. It provides an inherently stable and efficient way to replicate complex dynamic behaviors, such as folding, stretching, and collision interactions. Our proposed model characterizes virtual cloth based on softness-to-stiffness labels and accurately categorizes videos using this labeling. The cloth movement dataset utilized in this research is derived from a meticulously designed stiffness-oriented cloth simulation. Our experimental assessment encompasses an extensive dataset of 3840 videos, contributing to a multi-label video classification dataset. Our results demonstrate that our proposed model achieves an impressive average accuracy of 99.50%. These accuracies significantly outperform alternative models such as RNN, GRU, LSTM, and Transformer.
Collapse
Affiliation(s)
- Makara Mao
- Department of Software Convergence, Soonchunhyang University, Asan 31538, Republic of Korea; (M.M.); (H.V.)
| | - Hongly Va
- Department of Software Convergence, Soonchunhyang University, Asan 31538, Republic of Korea; (M.M.); (H.V.)
| | - Min Hong
- Department of Computer Software Engineering, Soonchunhyang University, Asan 31538, Republic of Korea
| |
Collapse
|
6
|
Bakasa W, Viriri S. Stacked ensemble deep learning for pancreas cancer classification using extreme gradient boosting. Front Artif Intell 2023; 6:1232640. [PMID: 37876961 PMCID: PMC10591225 DOI: 10.3389/frai.2023.1232640] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Accepted: 09/04/2023] [Indexed: 10/26/2023] Open
Abstract
Ensemble learning aims to improve prediction performance by combining several models or forecasts. However, how much and which ensemble learning techniques are useful in deep learning-based pipelines for pancreas computed tomography (CT) image classification is a challenge. Ensemble approaches are the most advanced solution to many machine learning problems. These techniques entail training multiple models and combining their predictions to improve the predictive performance of a single model. This article introduces the idea of Stacked Ensemble Deep Learning (SEDL), a pipeline for classifying pancreas CT medical images. The weak learners are Inception V3, VGG16, and ResNet34, and we employed a stacking ensemble. By combining the first-level predictions, an input train set for XGBoost, the ensemble model at the second level of prediction, is created. Extreme Gradient Boosting (XGBoost), employed as a strong learner, will make the final classification. Our findings showed that SEDL performed better, with a 98.8% ensemble accuracy, after some adjustments to the hyperparameters. The Cancer Imaging Archive (TCIA) public access dataset consists of 80 pancreas CT scans with a resolution of 512 * 512 pixels, from 53 male and 27 female subjects. A sample of two hundred and twenty-two images was used for training and testing data. We concluded that implementing the SEDL technique is an effective way to strengthen the robustness and increase the performance of the pipeline for classifying pancreas CT medical images. Interestingly, grouping like-minded or talented learners does not make a difference.
Collapse
Affiliation(s)
| | - Serestina Viriri
- School of Mathematics Statistics & Computer Science, College of Agriculture, Engineering and Science, University of KwaZulu-Natal, Durban, South Africa
| |
Collapse
|
7
|
Leandro I, Lorenzo B, Aleksandar M, Dario M, Rosa G, Agostino A, Daniele T. OCT-based deep-learning models for the identification of retinal key signs. Sci Rep 2023; 13:14628. [PMID: 37670066 PMCID: PMC10480174 DOI: 10.1038/s41598-023-41362-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 08/25/2023] [Indexed: 09/07/2023] Open
Abstract
A new system based on binary Deep Learning (DL) convolutional neural networks has been developed to recognize specific retinal abnormality signs on Optical Coherence Tomography (OCT) images useful for clinical practice. Images from the local hospital database were retrospectively selected from 2017 to 2022. Images were labeled by two retinal specialists and included central fovea cross-section OCTs. Nine models were developed using the Visual Geometry Group 16 architecture to distinguish healthy versus abnormal retinas and to identify eight different retinal abnormality signs. A total of 21,500 OCT images were screened, and 10,770 central fovea cross-section OCTs were included in the study. The system achieved high accuracy in identifying healthy retinas and specific pathological signs, ranging from 93 to 99%. Accurately detecting abnormal retinal signs from OCT images is crucial for patient care. This study aimed to identify specific signs related to retinal pathologies, aiding ophthalmologists in diagnosis. The high-accuracy system identified healthy retinas and pathological signs, making it a useful diagnostic aid. Labelled OCT images remain a challenge, but our approach reduces dataset creation time and shows DL models' potential to improve ocular pathology diagnosis and clinical decision-making.
Collapse
Affiliation(s)
- Inferrera Leandro
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy.
| | - Borsatti Lorenzo
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| | | | - Marangoni Dario
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| | - Giglio Rosa
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| | - Accardo Agostino
- Department of Engineering and Architecture, University of Trieste, Trieste, Italy
| | - Tognetto Daniele
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| |
Collapse
|
8
|
Kuanr M, Mohapatra P, Mittal S, Maindarkar M, Fouda MM, Saba L, Saxena S, Suri JS. Recommender System for the Efficient Treatment of COVID-19 Using a Convolutional Neural Network Model and Image Similarity. Diagnostics (Basel) 2022; 12:2700. [PMID: 36359545 PMCID: PMC9689970 DOI: 10.3390/diagnostics12112700] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2022] [Revised: 10/30/2022] [Accepted: 11/03/2022] [Indexed: 09/09/2023] Open
Abstract
Background: Hospitals face a significant problem meeting patients' medical needs during epidemics, especially when the number of patients increases rapidly, as seen during the recent COVID-19 pandemic. This study designs a treatment recommender system (RS) for the efficient management of human capital and resources such as doctors, medicines, and resources in hospitals. We hypothesize that a deep learning framework, when combined with search paradigms in an image framework, can make the RS very efficient. Methodology: This study uses a Convolutional neural network (CNN) model for the feature extraction of the images and discovers the most similar patients. The input queries patients from the hospital database with similar chest X-ray images. It uses a similarity metric for the similarity computation of the images. Results: This methodology recommends the doctors, medicines, and resources associated with similar patients to a COVID-19 patients being admitted to the hospital. The performance of the proposed RS is verified with five different feature extraction CNN models and four similarity measures. The proposed RS with a ResNet-50 CNN feature extraction model and Maxwell-Boltzmann similarity is found to be a proper framework for treatment recommendation with a mean average precision of more than 0.90 for threshold similarities in the range of 0.7 to 0.9 and an average highest cosine similarity of more than 0.95. Conclusions: Overall, an RS with a CNN model and image similarity is proven as an efficient tool for the proper management of resources during the peak period of pandemics and can be adopted in clinical settings.
Collapse
Affiliation(s)
- Madhusree Kuanr
- Department of Computer Science and Engineering, IIIT, Bhubaneswar 751003, India
| | | | - Sanchi Mittal
- Department of Computer Science and Engineering, IIIT, Bhubaneswar 751003, India
| | - Mahesh Maindarkar
- Stroke Monitoring and Diagnostic Division, AtheroPointTM, Roseville, CA 95661, USA
| | - Mostafa M. Fouda
- Department of Electrical and Computer Engineering, Idaho State University, Pocatello, ID 83209, USA
| | - Luca Saba
- Department of Radiology, University of Cagliari, 09123 Cagliari, Italy
| | - Sanjay Saxena
- Department of Computer Science and Engineering, IIIT, Bhubaneswar 751003, India
| | - Jasjit S. Suri
- Stroke Monitoring and Diagnostic Division, AtheroPointTM, Roseville, CA 95661, USA
- Knowledge Engineering Center, Global Biomedical Technologies, Inc., Roseville, CA 95661, USA
| |
Collapse
|
9
|
Hsieh YM, Wang TJ, Lin CY, Tsai YF, Cheng FT. Convolutional Autoencoder and Transfer Learning for Automatic Virtual Metrology. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3187617] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Affiliation(s)
- Yu-Ming Hsieh
- Institute of Manufacturing Information and Systems, National Cheng Kung University, Tainan, Taiwan
| | - Tan-Ju Wang
- Institute of Manufacturing Information and Systems, National Cheng Kung University, Tainan, Taiwan
| | - Chin-Yi Lin
- Institute of Manufacturing Information and Systems, National Cheng Kung University, Tainan, Taiwan
| | - Yueh-Feng Tsai
- Institute of Manufacturing Information and Systems, National Cheng Kung University, Tainan, Taiwan
| | - Fan-Tien Cheng
- Institute of Manufacturing Information and Systems, National Cheng Kung University, Tainan, Taiwan
| |
Collapse
|