1
|
Song Z, Li H, Zhang Y, Zhu C, Jiang M, Song L, Wang Y, Ouyang M, Hu F, Zheng Q. s 2MRI-ADNet: an interpretable deep learning framework integrating Euclidean-graph representations of Alzheimer's disease solely from structural MRI. MAGMA (NEW YORK, N.Y.) 2024; 37:845-857. [PMID: 38869733 DOI: 10.1007/s10334-024-01178-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Revised: 05/19/2024] [Accepted: 06/04/2024] [Indexed: 06/14/2024]
Abstract
OBJECTIVE To establish a multi-dimensional representation solely on structural MRI (sMRI) for early diagnosis of AD. METHODS A total of 3377 participants' sMRI from four independent databases were retrospectively identified to construct an interpretable deep learning model that integrated multi-dimensional representations of AD solely on sMRI (called s2MRI-ADNet) by a dual-channel learning strategy of gray matter volume (GMV) from Euclidean space and the regional radiomics similarity network (R2SN) from graph space. Specifically, the GMV feature map learning channel (called GMV-Channel) was to take into consideration spatial information of both long-range spatial relations and detailed localization information, while the node feature and connectivity strength learning channel (called NFCS-Channel) was to characterize the graph-structured R2SN network by a separable learning strategy. RESULTS The s2MRI-ADNet achieved a superior classification accuracy of 92.1% and 91.4% under intra-database and inter-database cross-validation. The GMV-Channel and NFCS-Channel captured complementary group-discriminative brain regions, revealing a complementary interpretation of the multi-dimensional representation of brain structure in Euclidean and graph spaces respectively. Besides, the generalizable and reproducible interpretation of the multi-dimensional representation in capturing complementary group-discriminative brain regions revealed a significant correlation between the four independent databases (p < 0.05). Significant associations (p < 0.05) between attention scores and brain abnormality, between classification scores and clinical measure of cognitive ability, CSF biomarker, metabolism, and genetic risk score also provided solid neurobiological interpretation. CONCLUSION The s2MRI-ADNet solely on sMRI could leverage the complementary multi-dimensional representations of AD in Euclidean and graph spaces, and achieved superior performance in the early diagnosis of AD, facilitating its potential in both clinical translation and popularization.
Collapse
Affiliation(s)
- Zhiwei Song
- School of Computer and Control Engineering, Yantai University, Yantai, 264005, China
| | - Honglun Li
- Department of Radiology, Yantai Yuhuangding Hospital Affiliated with Qingdao University Medical College, Yantai, 264099, China
| | - Yiyu Zhang
- School of Computer and Control Engineering, Yantai University, Yantai, 264005, China
| | - Chuanzhen Zhu
- School of Computer and Control Engineering, Yantai University, Yantai, 264005, China
| | - Minbo Jiang
- School of Computer and Control Engineering, Yantai University, Yantai, 264005, China
| | - Limei Song
- School of Medical Imaging, Weifang Medical University, Weifang, 261000, China
| | - Yi Wang
- School of Computer and Control Engineering, Yantai University, Yantai, 264005, China
- Key Laboratory of Medical Imaging and Artificial Intelligence of Hunan Province, Xiangnan University, Chenzhou, 423000, Hunan, China
| | - Minhui Ouyang
- Department of Radiology, Children's Hospital of Philadelphia, Philadelphia, PA, 19104, USA
- Department of Radiology, Perelman School of Medicine, University of Pennsylvania, Philadelphia, PA, 19104, USA
| | - Fang Hu
- Key Laboratory of Medical Imaging and Artificial Intelligence of Hunan Province, Xiangnan University, Chenzhou, 423000, Hunan, China
| | - Qiang Zheng
- School of Computer and Control Engineering, Yantai University, Yantai, 264005, China.
| |
Collapse
|
2
|
Zia-Ur-Rehman, Awang MK, Ali G, Faheem M. Deep learning techniques for Alzheimer's disease detection in 3D imaging: A systematic review. Health Sci Rep 2024; 7:e70025. [PMID: 39296636 PMCID: PMC11409051 DOI: 10.1002/hsr2.70025] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2024] [Revised: 07/08/2024] [Accepted: 08/13/2024] [Indexed: 09/21/2024] Open
Abstract
Background and Aims Alzheimer's disease (AD) is a degenerative neurological condition that worsens over time and leads to deterioration in cognitive abilities, reduced memory, and, eventually, a decrease in overall functioning. Timely and correct identification of Alzheimer's is essential for effective treatment. The systematic study specifically examines the application of deep learning (DL) algorithms in identifying AD using three-dimensional (3D) imaging methods. The main goal is to evaluate these methods' current state, efficiency, and potential enhancements, offering valuable insights into how DL could improve AD's rapid and precise diagnosis. Methods We searched different online repositories, such as IEEE Xplore, Elsevier, MDPI, PubMed Central, Science Direct, ACM, Springer, and others, to thoroughly summarize current research on DL methods to diagnose AD by analyzing 3D imaging data published between 2020 and 2024. We use PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines to ensure the organization and understandability of the information collection process. We thoroughly analyzed the literature to determine the primary techniques used in these investigations and their findings. Results and Conclusion The ability of convolutional neural networks (CNNs) and their variations, including 3D CNNs and recurrent neural networks, to detect both temporal and spatial characteristics in volumetric data has led to their widespread use. Methods such as transfer learning, combining multimodal data, and using attention procedures have improved models' precision and reliability. We selected 87 articles for evaluation. Out of these, 31 papers included various concepts, explanations, and elucidations of models and theories, while the other 56 papers primarily concentrated on issues related to practical implementation. This article introduces popular imaging types, 3D imaging for Alzheimer's detection, discusses the benefits and restrictions of the DL-based approach to AD assessment, and gives a view toward future developments resulting from critical evaluation.
Collapse
Affiliation(s)
- Zia-Ur-Rehman
- Faculty of Informatics and Computing Universiti Sultan Zainal Abidin (UniSZA) Terengganu Malaysia
| | - Mohd Khalid Awang
- Faculty of Informatics and Computing Universiti Sultan Zainal Abidin (UniSZA) Terengganu Malaysia
| | - Ghulam Ali
- Department of Computer Science University of Okara Okara Pakistan
| | - Muhammad Faheem
- School of Technology and Innovations University of Vaasa Vaasa Finland
| |
Collapse
|
3
|
Odusami M, Damaševičius R, Milieškaitė-Belousovienė E, Maskeliūnas R. Alzheimer's disease stage recognition from MRI and PET imaging data using Pareto-optimal quantum dynamic optimization. Heliyon 2024; 10:e34402. [PMID: 39145034 PMCID: PMC11320145 DOI: 10.1016/j.heliyon.2024.e34402] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2023] [Revised: 07/06/2024] [Accepted: 07/09/2024] [Indexed: 08/16/2024] Open
Abstract
The threat posed by Alzheimer's disease (AD) to human health has grown significantly. However, the precise diagnosis and classification of AD stages remain a challenge. Neuroimaging methods such as structural magnetic resonance imaging (sMRI) and fluorodeoxyglucose positron emission tomography (FDG-PET) have been used to diagnose and categorize AD. However, feature selection approaches that are frequently used to extract additional data from multimodal imaging are prone to errors. This paper suggests using a static pulse-coupled neural network and a Laplacian pyramid to combine sMRI and FDG-PET data. After that, the fused images are used to train the Mobile Vision Transformer (MViT), optimized with Pareto-Optimal Quantum Dynamic Optimization for Neural Architecture Search, while the fused images are augmented to avoid overfitting and then classify unfused MRI and FDG-PET images obtained from the AD Neuroimaging Initiative (ADNI) and Open Access Series of Imaging Studies (OASIS) datasets into various stages of AD. The architectural hyperparameters of MViT are optimized using Quantum Dynamic Optimization, which ensures a Pareto-optimal solution. The Peak Signal-to-Noise Ratio (PSNR), the Mean Squared Error (MSE), and the Structured Similarity Indexing Method (SSIM) are used to measure the quality of the fused image. We found that the fused image was consistent in all metrics, having 0.64 SIMM, 35.60 PSNR, and 0.21 MSE for the FDG-PET image. In the classification of AD vs. cognitive normal (CN), AD vs. mild cognitive impairment (MCI), and CN vs. MCI, the precision of the proposed method is 94.73%, 92.98% and 89.36%, respectively. The sensitivity is 90. 70%, 90. 70%, and 90. 91% while the specificity is 100%, 100%, and 85. 71%, respectively, in the ADNI MRI test data.
Collapse
Affiliation(s)
- Modupe Odusami
- Faculty of Informatics, Kaunas University of Technology, Kaunas, Lithuania
| | | | | | - Rytis Maskeliūnas
- Faculty of Informatics, Kaunas University of Technology, Kaunas, Lithuania
| |
Collapse
|
4
|
Li Y, El Habib Daho M, Conze PH, Zeghlache R, Le Boité H, Tadayoni R, Cochener B, Lamard M, Quellec G. A review of deep learning-based information fusion techniques for multimodal medical image classification. Comput Biol Med 2024; 177:108635. [PMID: 38796881 DOI: 10.1016/j.compbiomed.2024.108635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Revised: 03/18/2024] [Accepted: 05/18/2024] [Indexed: 05/29/2024]
Abstract
Multimodal medical imaging plays a pivotal role in clinical diagnosis and research, as it combines information from various imaging modalities to provide a more comprehensive understanding of the underlying pathology. Recently, deep learning-based multimodal fusion techniques have emerged as powerful tools for improving medical image classification. This review offers a thorough analysis of the developments in deep learning-based multimodal fusion for medical classification tasks. We explore the complementary relationships among prevalent clinical modalities and outline three main fusion schemes for multimodal classification networks: input fusion, intermediate fusion (encompassing single-level fusion, hierarchical fusion, and attention-based fusion), and output fusion. By evaluating the performance of these fusion techniques, we provide insight into the suitability of different network architectures for various multimodal fusion scenarios and application domains. Furthermore, we delve into challenges related to network architecture selection, handling incomplete multimodal data management, and the potential limitations of multimodal fusion. Finally, we spotlight the promising future of Transformer-based multimodal fusion techniques and give recommendations for future research in this rapidly evolving field.
Collapse
Affiliation(s)
- Yihao Li
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | - Mostafa El Habib Daho
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France.
| | | | - Rachid Zeghlache
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | - Hugo Le Boité
- Sorbonne University, Paris, France; Ophthalmology Department, Lariboisière Hospital, AP-HP, Paris, France
| | - Ramin Tadayoni
- Ophthalmology Department, Lariboisière Hospital, AP-HP, Paris, France; Paris Cité University, Paris, France
| | - Béatrice Cochener
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France; Ophthalmology Department, CHRU Brest, Brest, France
| | - Mathieu Lamard
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | | |
Collapse
|
5
|
Hassan N, Musa Miah AS, Shin J. Residual-Based Multi-Stage Deep Learning Framework for Computer-Aided Alzheimer's Disease Detection. J Imaging 2024; 10:141. [PMID: 38921618 PMCID: PMC11204904 DOI: 10.3390/jimaging10060141] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2024] [Revised: 06/06/2024] [Accepted: 06/06/2024] [Indexed: 06/27/2024] Open
Abstract
Alzheimer's Disease (AD) poses a significant health risk globally, particularly among the elderly population. Recent studies underscore its prevalence, with over 50% of elderly Japanese facing a lifetime risk of dementia, primarily attributed to AD. As the most prevalent form of dementia, AD gradually erodes brain cells, leading to severe neurological decline. In this scenario, it is important to develop an automatic AD-detection system, and many researchers have been working to develop an AD-detection system by taking advantage of the advancement of deep learning (DL) techniques, which have shown promising results in various domains, including medical image analysis. However, existing approaches for AD detection often suffer from limited performance due to the complexities associated with training hierarchical convolutional neural networks (CNNs). In this paper, we introduce a novel multi-stage deep neural network architecture based on residual functions to address the limitations of existing AD-detection approaches. Inspired by the success of residual networks (ResNets) in image-classification tasks, our proposed system comprises five stages, each explicitly formulated to enhance feature effectiveness while maintaining model depth. Following feature extraction, a deep learning-based feature-selection module is applied to mitigate overfitting, incorporating batch normalization, dropout and fully connected layers. Subsequently, machine learning (ML)-based classification algorithms, including Support Vector Machines (SVM), Random Forest (RF) and SoftMax, are employed for classification tasks. Comprehensive evaluations conducted on three benchmark datasets, namely ADNI1: Complete 1Yr 1.5T, MIRAID and OASIS Kaggle, demonstrate the efficacy of our proposed model. Impressively, our model achieves accuracy rates of 99.47%, 99.10% and 99.70% for ADNI1: Complete 1Yr 1.5T, MIRAID and OASIS datasets, respectively, outperforming existing systems in binary class problems. Our proposed model represents a significant advancement in the AD-analysis domain.
Collapse
Affiliation(s)
- Najmul Hassan
- School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan;
| | | | - Jungpil Shin
- School of Computer Science and Engineering, The University of Aizu, Aizuwakamatsu 965-8580, Japan;
| |
Collapse
|
6
|
Zhang H, Chen J, Liao B, Wu FX, Bi XA. Deep Canonical Correlation Fusion Algorithm Based on Denoising Autoencoder for ASD Diagnosis and Pathogenic Brain Region Identification. Interdiscip Sci 2024; 16:455-468. [PMID: 38573456 DOI: 10.1007/s12539-024-00625-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2023] [Revised: 02/22/2024] [Accepted: 02/25/2024] [Indexed: 04/05/2024]
Abstract
Autism Spectrum Disorder (ASD) is defined as a neurodevelopmental condition distinguished by unconventional neural activities. Early intervention is key to managing the progress of ASD, and current research primarily focuses on the use of structural magnetic resonance imaging (sMRI) or resting-state functional magnetic resonance imaging (rs-fMRI) for diagnosis. Moreover, the use of autoencoders for disease classification has not been sufficiently explored. In this study, we introduce a new framework based on autoencoder, the Deep Canonical Correlation Fusion algorithm based on Denoising Autoencoder (DCCF-DAE), which proves to be effective in handling high-dimensional data. This framework involves efficient feature extraction from different types of data with an advanced autoencoder, followed by the fusion of these features through the DCCF model. Then we utilize the fused features for disease classification. DCCF integrates functional and structural data to help accurately diagnose ASD and identify critical Regions of Interest (ROIs) in disease mechanisms. We compare the proposed framework with other methods by the Autism Brain Imaging Data Exchange (ABIDE) database and the results demonstrate its outstanding performance in ASD diagnosis. The superiority of DCCF-DAE highlights its potential as a crucial tool for early ASD diagnosis and monitoring.
Collapse
Affiliation(s)
- Huilian Zhang
- Key Laboratory of Data Science and Intelligence Education, Ministry of Education, Hainan Normal University, Haikou, 571126, China
- College of Mathematics and Statistics, Hainan Normal University, Haikou, 571126, China
| | - Jie Chen
- Key Laboratory of Data Science and Intelligence Education, Ministry of Education, Hainan Normal University, Haikou, 571126, China
- College of Mathematics and Statistics, Hainan Normal University, Haikou, 571126, China
| | - Bo Liao
- Key Laboratory of Data Science and Intelligence Education, Ministry of Education, Hainan Normal University, Haikou, 571126, China
- College of Mathematics and Statistics, Hainan Normal University, Haikou, 571126, China
| | - Fang-Xiang Wu
- Division of Biomedical Engineering, University of Saskatchewan, Saskatoon, S7N5A9, Canada
| | - Xia-An Bi
- Key Laboratory of Data Science and Intelligence Education, Ministry of Education, Hainan Normal University, Haikou, 571126, China.
- College of Mathematics and Statistics, Hainan Normal University, Haikou, 571126, China.
- College of Information Science and Engineering, Hunan Normal University, Changsha, Hunan, 410081, China.
| |
Collapse
|
7
|
Castellano G, Esposito A, Lella E, Montanaro G, Vessio G. Automated detection of Alzheimer's disease: a multi-modal approach with 3D MRI and amyloid PET. Sci Rep 2024; 14:5210. [PMID: 38433282 PMCID: PMC10909869 DOI: 10.1038/s41598-024-56001-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2022] [Accepted: 02/28/2024] [Indexed: 03/05/2024] Open
Abstract
Recent advances in deep learning and imaging technologies have revolutionized automated medical image analysis, especially in diagnosing Alzheimer's disease through neuroimaging. Despite the availability of various imaging modalities for the same patient, the development of multi-modal models leveraging these modalities remains underexplored. This paper addresses this gap by proposing and evaluating classification models using 2D and 3D MRI images and amyloid PET scans in uni-modal and multi-modal frameworks. Our findings demonstrate that models using volumetric data learn more effective representations than those using only 2D images. Furthermore, integrating multiple modalities enhances model performance over single-modality approaches significantly. We achieved state-of-the-art performance on the OASIS-3 cohort. Additionally, explainability analyses with Grad-CAM indicate that our model focuses on crucial AD-related regions for its predictions, underscoring its potential to aid in understanding the disease's causes.
Collapse
Affiliation(s)
| | - Andrea Esposito
- Department of Computer Science, University of Bari Aldo Moro, Bari, Italy
| | - Eufemia Lella
- Sirio - Research & Innovation, Sidea Group, Bari, Italy
| | | | - Gennaro Vessio
- Department of Computer Science, University of Bari Aldo Moro, Bari, Italy.
| |
Collapse
|
8
|
Ying M, Wang Y, Yang K, Wang H, Liu X. A deep learning knowledge distillation framework using knee MRI and arthroscopy data for meniscus tear detection. Front Bioeng Biotechnol 2024; 11:1326706. [PMID: 38292305 PMCID: PMC10825958 DOI: 10.3389/fbioe.2023.1326706] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2023] [Accepted: 12/22/2023] [Indexed: 02/01/2024] Open
Abstract
Purpose: To construct a deep learning knowledge distillation framework exploring the utilization of MRI alone or combing with distilled Arthroscopy information for meniscus tear detection. Methods: A database of 199 paired knee Arthroscopy-MRI exams was used to develop a multimodal teacher network and an MRI-based student network, which used residual neural networks architectures. A knowledge distillation framework comprising the multimodal teacher network T and the monomodal student network S was proposed. We optimized the loss functions of mean squared error (MSE) and cross-entropy (CE) to enable the student network S to learn arthroscopic information from the teacher network T through our deep learning knowledge distillation framework, ultimately resulting in a distilled student network S T. A coronal proton density (PD)-weighted fat-suppressed MRI sequence was used in this study. Fivefold cross-validation was employed, and the accuracy, sensitivity, specificity, F1-score, receiver operating characteristic (ROC) curves and area under the receiver operating characteristic curve (AUC) were used to evaluate the medial and lateral meniscal tears detection performance of the models, including the undistilled student model S, the distilled student model S T and the teacher model T. Results: The AUCs of the undistilled student model S, the distilled student model S T, the teacher model T for medial meniscus (MM) tear detection and lateral meniscus (LM) tear detection are 0.773/0.672, 0.792/0.751 and 0.834/0.746, respectively. The distilled student model S T had higher AUCs than the undistilled model S. After undergoing knowledge distillation processing, the distilled student model demonstrated promising results, with accuracy (0.764/0.734), sensitivity (0.838/0.661), and F1-score (0.680/0.754) for both medial and lateral tear detection better than the undistilled one with accuracy (0.734/0.648), sensitivity (0.733/0.607), and F1-score (0.620/0.673). Conclusion: Through the knowledge distillation framework, the student model S based on MRI benefited from the multimodal teacher model T and achieved an improved meniscus tear detection performance.
Collapse
Affiliation(s)
- Mengjie Ying
- Department of Orthopedics, Shanghai Sixth People’s Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Yufan Wang
- Engineering Research Center for Digital Medicine of the Ministry of Education, Shanghai, China
- School of Biomedical Engineering and Med-X Research Institute, Shanghai Jiao Tong University, Shanghai, China
| | - Kai Yang
- Department of Radiology, Shanghai Sixth People’s Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Haoyuan Wang
- Department of Orthopedics, Shanghai Sixth People’s Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Xudong Liu
- Department of Orthopedics, Shanghai Sixth People’s Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| |
Collapse
|
9
|
Zhang J, He X, Liu Y, Cai Q, Chen H, Qing L. Multi-modal cross-attention network for Alzheimer's disease diagnosis with multi-modality data. Comput Biol Med 2023; 162:107050. [PMID: 37269680 DOI: 10.1016/j.compbiomed.2023.107050] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2022] [Revised: 04/26/2023] [Accepted: 05/03/2023] [Indexed: 06/05/2023]
Abstract
Alzheimer's disease (AD) is a neurodegenerative disorder, the most common cause of dementia, so the accurate diagnosis of AD and its prodromal stage mild cognitive impairment (MCI) is significant. Recent studies have demonstrated that multiple neuroimaging and biological measures contain complementary information for diagnosis. Many existing multi-modal models based on deep learning simply concatenate each modality's features despite substantial differences in representation spaces. In this paper, we propose a novel multi-modal cross-attention AD diagnosis (MCAD) framework to learn the interaction between modalities for better playing their complementary roles for AD diagnosis with multi-modal data including structural magnetic resonance imaging (sMRI), fluorodeoxyglucose-positron emission tomography (FDG-PET) and cerebrospinal fluid (CSF) biomarkers. Specifically, the imaging and non-imaging representations are learned by the image encoder based on cascaded dilated convolutions and CSF encoder, respectively. Then, a multi-modal interaction module is introduced, which takes advantage of cross-modal attention to integrate imaging and non-imaging information and reinforce relationships between these modalities. Moreover, an extensive objective function is designed to reduce the discrepancy between modalities for effectively fusing the features of multi-modal data, which could further improve the diagnosis performance. We evaluate the effectiveness of our proposed method on the ADNI dataset, and the extensive experiments demonstrate that our MCAD achieves superior performance for multiple AD-related classification tasks, compared to several competing methods. Also, we investigate the importance of cross-attention and the contribution of each modality to the diagnostics performance. The experimental results demonstrate that combining multi-modality data via cross-attention is helpful for accurate AD diagnosis.
Collapse
Affiliation(s)
- Jin Zhang
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, 610065, China
| | - Xiaohai He
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, 610065, China.
| | - Yan Liu
- Department of Neurology, The Affiliated Hospital of Southwest Jiaotong University, The Third People's Hospital of Chengdu, Chengdu, Sichuan, 610031, China
| | - Qingyan Cai
- Department of Geriatric Medicine, The Fourth People's Hospital of Chengdu, Chengdu, Sichuan, 610036, China
| | - Honggang Chen
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, 610065, China
| | - Linbo Qing
- College of Electronics and Information Engineering, Sichuan University, Chengdu, Sichuan, 610065, China
| |
Collapse
|
10
|
Tang C, Wei M, Sun J, Wang S, Zhang Y. CsAGP: Detecting Alzheimer's disease from multimodal images via dual-transformer with cross-attention and graph pooling. JOURNAL OF KING SAUD UNIVERSITY. COMPUTER AND INFORMATION SCIENCES 2023; 35:101618. [PMID: 38559705 PMCID: PMC7615783 DOI: 10.1016/j.jksuci.2023.101618] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Alzheimer's disease (AD) is a terrible and degenerative disease commonly occurring in the elderly. Early detection can prevent patients from further damage, which is crucial in treating AD. Over the past few decades, it has been demonstrated that neuroimaging can be a critical diagnostic tool for AD, and the feature fusion of different neuroimaging modalities can enhance diagnostic performance. Most previous studies in multimodal feature fusion have only concatenated the high-level features extracted by neural networks from various neuroimaging images simply. However, a major problem of these studies is over-looking the low-level feature interactions between modalities in the feature extraction stage, resulting in suboptimal performance in AD diagnosis. In this paper, we develop a dual-branch vision transformer with cross-attention and graph pooling, namely CsAGP, which enables multi-level feature interactions between the inputs to learn a shared feature representation. Specifically, we first construct a brand-new cross-attention fusion module (CAFM), which processes MRI and PET images by two independent branches of differing computational complexity. These features are fused merely by the cross-attention mechanism to enhance each other. After that, a concise graph pooling algorithm-based Reshape-Pooling-Reshape (RPR) framework is developed for token selection to reduce token redundancy in the proposed model. Extensive experiments on the Alzheimer's Disease Neuroimaging Initiative (ADNI) database demonstrated that the suggested method obtains 99.04%, 97.43%, 98.57%, and 98.72% accuracy for the classification of AD vs. CN, AD vs. MCI, CN vs. MCI, and AD vs. CN vs. MCI, respectively.
Collapse
Affiliation(s)
- Chaosheng Tang
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, PR China
| | - Mingyang Wei
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, PR China
| | - Junding Sun
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, PR China
| | - Shuihua Wang
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, PR China
- School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| | - Yudong Zhang
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo, Henan 454000, PR China
- School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| | | |
Collapse
|
11
|
Kang W, Lin L, Sun S, Wu S. Three-round learning strategy based on 3D deep convolutional GANs for Alzheimer's disease staging. Sci Rep 2023; 13:5750. [PMID: 37029214 PMCID: PMC10081988 DOI: 10.1038/s41598-023-33055-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2022] [Accepted: 04/06/2023] [Indexed: 04/09/2023] Open
Abstract
Accurately diagnosing of Alzheimer's disease (AD) and its early stages is critical for prompt treatment or potential intervention to delay the the disease's progression. Convolutional neural networks (CNNs) models have shown promising results in structural MRI (sMRI)-based diagnosis, but their performance, particularly for 3D models, is constrained by the lack of labeled training samples. To address the overfitting problem brought on by the insufficient training sample size, we propose a three-round learning strategy that combines transfer learning with generative adversarial learning. In the first round, a 3D Deep Convolutional Generative Adversarial Networks (DCGAN) model was trained with all available sMRI data to learn the common feature of sMRI through unsupervised generative adversarial learning. The second round involved transferring and fine-tuning, and the pre-trained discriminator (D) of the DCGAN learned more specific features for the classification task between AD and cognitively normal (CN). In the final round, the weights learned in the AD versus CN classification task were transferred to the MCI diagnosis. By highlighting brain regions with high prediction weights using 3D Grad-CAM, we further enhanced the model's interpretability. The proposed model achieved accuracies of 92.8%, 78.1%, and 76.4% in the classifications of AD versus CN, AD versus MCI, and MCI versus CN, respectively. The experimental results show that our proposed model avoids overfitting brought on by a paucity of sMRI data and enables the early detection of AD.
Collapse
Affiliation(s)
- Wenjie Kang
- Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing, 100124, China
| | - Lan Lin
- Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing, 100124, China.
| | - Shen Sun
- Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing, 100124, China
| | - Shuicai Wu
- Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing, 100124, China
| |
Collapse
|
12
|
Xu X, Lin L, Sun S, Wu S. A review of the application of three-dimensional convolutional neural networks for the diagnosis of Alzheimer's disease using neuroimaging. Rev Neurosci 2023:revneuro-2022-0122. [PMID: 36729918 DOI: 10.1515/revneuro-2022-0122] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2022] [Accepted: 01/02/2023] [Indexed: 02/03/2023]
Abstract
Alzheimer's disease (AD) is a degenerative disorder that leads to progressive, irreversible cognitive decline. To obtain an accurate and timely diagnosis and detect AD at an early stage, numerous approaches based on convolutional neural networks (CNNs) using neuroimaging data have been proposed. Because 3D CNNs can extract more spatial discrimination information than 2D CNNs, they have emerged as a promising research direction in the diagnosis of AD. The aim of this article is to present the current state of the art in the diagnosis of AD using 3D CNN models and neuroimaging modalities, focusing on the 3D CNN architectures and classification methods used, and to highlight potential future research topics. To give the reader a better overview of the content mentioned in this review, we briefly introduce the commonly used imaging datasets and the fundamentals of CNN architectures. Then we carefully analyzed the existing studies on AD diagnosis, which are divided into two levels according to their inputs: 3D subject-level CNNs and 3D patch-level CNNs, highlighting their contributions and significance in the field. In addition, this review discusses the key findings and challenges from the studies and highlights the lessons learned as a roadmap for future research. Finally, we summarize the paper by presenting some major findings, identifying open research challenges, and pointing out future research directions.
Collapse
Affiliation(s)
- Xinze Xu
- Intelligent Physiological Measurement and Clinical Translation, Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing 100124, China
| | - Lan Lin
- Intelligent Physiological Measurement and Clinical Translation, Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing 100124, China
| | - Shen Sun
- Intelligent Physiological Measurement and Clinical Translation, Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing 100124, China
| | - Shuicai Wu
- Intelligent Physiological Measurement and Clinical Translation, Beijing International Platform for Scientific and Technological Cooperation, Department of Biomedical Engineering, Faculty of Environment and Life Sciences, Beijing University of Technology, Beijing 100124, China
| |
Collapse
|
13
|
Kong Z, Zhang M, Zhu W, Yi Y, Wang T, Zhang B. Data enhancement based on M2-Unet for liver segmentation in Computed Tomography. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/15/2022]
|
14
|
Baghdadi NA, Malki A, Balaha HM, Badawy M, Elhosseini M. A 3C-TL-GTO: Alzheimer Automatic Accurate Classification Using Transfer Learning and Artificial Gorilla Troops Optimizer. SENSORS (BASEL, SWITZERLAND) 2022; 22:s22114250. [PMID: 35684871 PMCID: PMC9185328 DOI: 10.3390/s22114250] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/10/2022] [Revised: 05/24/2022] [Accepted: 05/28/2022] [Indexed: 05/10/2023]
Abstract
Alzheimer's disease (AD) is a chronic disease that affects the elderly. There are many different types of dementia, but Alzheimer's disease is one of the leading causes of death. AD is a chronic brain disorder that leads to problems with language, disorientation, mood swings, bodily functions, memory loss, cognitive decline, mood or personality changes, and ultimately death due to dementia. Unfortunately, no cure has yet been developed for it, and it has no known causes. Clinically, imaging tools can aid in the diagnosis, and deep learning has recently emerged as an important component of these tools. Deep learning requires little or no image preprocessing and can infer an optimal data representation from raw images without prior feature selection. As a result, they produce a more objective and less biased process. The performance of a convolutional neural network (CNN) is primarily affected by the hyperparameters chosen and the dataset used. A deep learning model for classifying Alzheimer's patients has been developed using transfer learning and optimized by Gorilla Troops for early diagnosis. This study proposes the A3C-TL-GTO framework for MRI image classification and AD detection. The A3C-TL-GTO is an empirical quantitative framework for accurate and automatic AD classification, developed and evaluated with the Alzheimer's Dataset (four classes of images) and the Alzheimer's Disease Neuroimaging Initiative (ADNI). The proposed framework reduces the bias and variability of preprocessing steps and hyperparameters optimization to the classifier model and dataset used. Our strategy, evaluated on MRIs, is easily adaptable to other imaging methods. According to our findings, the proposed framework was an excellent instrument for this task, with a significant potential advantage for patient care. The ADNI dataset, an online dataset on Alzheimer's disease, was used to obtain magnetic resonance imaging (MR) brain images. The experimental results demonstrate that the proposed framework achieves 96.65% accuracy for the Alzheimer's Dataset and 96.25% accuracy for the ADNI dataset. Moreover, a better performance in terms of accuracy is demonstrated over other state-of-the-art approaches.
Collapse
Affiliation(s)
- Nadiah A. Baghdadi
- College of Nursing, Princess Nourah Bint Abdulrahman University, Riyadh 11671, Saudi Arabia;
| | - Amer Malki
- College of Computer Science and Engineering, Taibah University, Yanbu 46421, Saudi Arabia; (A.M.); (M.E.)
| | - Hossam Magdy Balaha
- Computers and Control Systems Engineering Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt;
| | - Mahmoud Badawy
- Computers and Control Systems Engineering Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt;
- Correspondence:
| | - Mostafa Elhosseini
- College of Computer Science and Engineering, Taibah University, Yanbu 46421, Saudi Arabia; (A.M.); (M.E.)
- Computers and Control Systems Engineering Department, Faculty of Engineering, Mansoura University, Mansoura 35516, Egypt;
| |
Collapse
|
15
|
Ji B, Wang H, Zhang M, Mao B, Li X. An Efficient Lightweight Network Based on Magnetic Resonance Images for Predicting Alzheimer's Disease. INT J SEMANT WEB INF 2022. [DOI: 10.4018/ijswis.313715] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
Abstract
Brain magnetic resonance images (MRI) are widely used for the classification of Alzheimer's disease (AD). The size of 3D images is, however, too large. Some of the sliced image features are lost, which results in conflicting network size and classification performance. This article uses key components in the transformer model to propose a new lightweight method, ensuring the lightness of the network and achieving highly accurate classification. First, the transformer model is imitated by using image patch input to enhance feature perception. Second, the Gaussian error linear unit (GELU), commonly used in transformer models, is used to enhance the generalization ability of the network. Finally, the network uses MRI slices as learning data. The depthwise separable convolution makes the network more lightweight. Experiments are carried out on the ADNI public database. The accuracy rate of AD vs. normal control (NC) experiments reaches 98.54%. The amount of network parameters is 1.3% of existing similar networks.
Collapse
|