1
|
Vedaei F, Mashhadi N, Alizadeh M, Zabrecky G, Monti D, Wintering N, Navarreto E, Hriso C, Newberg AB, Mohamed FB. Deep learning-based multimodality classification of chronic mild traumatic brain injury using resting-state functional MRI and PET imaging. Front Neurosci 2024; 17:1333725. [PMID: 38312737 PMCID: PMC10837852 DOI: 10.3389/fnins.2023.1333725] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2023] [Accepted: 12/28/2023] [Indexed: 02/06/2024] Open
Abstract
Mild traumatic brain injury (mTBI) is a public health concern. The present study aimed to develop an automatic classifier to distinguish between patients with chronic mTBI (n = 83) and healthy controls (HCs) (n = 40). Resting-state functional MRI (rs-fMRI) and positron emission tomography (PET) imaging were acquired from the subjects. We proposed a novel deep-learning-based framework, including an autoencoder (AE), to extract high-level latent and rectified linear unit (ReLU) and sigmoid activation functions. Single and multimodality algorithms integrating multiple rs-fMRI metrics and PET data were developed. We hypothesized that combining different imaging modalities provides complementary information and improves classification performance. Additionally, a novel data interpretation approach was utilized to identify top-performing features learned by the AEs. Our method delivered a classification accuracy within the range of 79-91.67% for single neuroimaging modalities. However, the performance of classification improved to 95.83%, thereby employing the multimodality model. The models have identified several brain regions located in the default mode network, sensorimotor network, visual cortex, cerebellum, and limbic system as the most discriminative features. We suggest that this approach could be extended to the objective biomarkers predicting mTBI in clinical settings.
Collapse
Affiliation(s)
- Faezeh Vedaei
- Department of Radiology, Jefferson Integrated Magnetic Resonance Imaging Center, Thomas Jefferson University, Philadelphia, PA, United States
| | - Najmeh Mashhadi
- Department of Computer Science and Engineering, University of California, Santa Cruz, Santa Cruz, CA, United States
| | - Mahdi Alizadeh
- Department of Radiology, Jefferson Integrated Magnetic Resonance Imaging Center, Thomas Jefferson University, Philadelphia, PA, United States
| | - George Zabrecky
- Department of Integrative Medicine and Nutritional Sciences, Marcus Institute of Integrative, Health, Thomas Jefferson University, Philadelphia, PA, United States
| | - Daniel Monti
- Department of Integrative Medicine and Nutritional Sciences, Marcus Institute of Integrative, Health, Thomas Jefferson University, Philadelphia, PA, United States
| | - Nancy Wintering
- Department of Integrative Medicine and Nutritional Sciences, Marcus Institute of Integrative, Health, Thomas Jefferson University, Philadelphia, PA, United States
| | - Emily Navarreto
- Department of Integrative Medicine and Nutritional Sciences, Marcus Institute of Integrative, Health, Thomas Jefferson University, Philadelphia, PA, United States
| | - Chloe Hriso
- Department of Integrative Medicine and Nutritional Sciences, Marcus Institute of Integrative, Health, Thomas Jefferson University, Philadelphia, PA, United States
| | - Andrew B. Newberg
- Department of Radiology, Jefferson Integrated Magnetic Resonance Imaging Center, Thomas Jefferson University, Philadelphia, PA, United States
- Department of Integrative Medicine and Nutritional Sciences, Marcus Institute of Integrative, Health, Thomas Jefferson University, Philadelphia, PA, United States
| | - Feroze B. Mohamed
- Department of Radiology, Jefferson Integrated Magnetic Resonance Imaging Center, Thomas Jefferson University, Philadelphia, PA, United States
| |
Collapse
|
2
|
Choi H, Byeon K, Lee J, Hong S, Park B, Park H. Identifying subgroups of eating behavior traits unrelated to obesity using functional connectivity and feature representation learning. Hum Brain Mapp 2024; 45:e26581. [PMID: 38224537 PMCID: PMC10789215 DOI: 10.1002/hbm.26581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 12/13/2023] [Accepted: 12/20/2023] [Indexed: 01/17/2024] Open
Abstract
Eating behavior is highly heterogeneous across individuals and cannot be fully explained using only the degree of obesity. We utilized unsupervised machine learning and functional connectivity measures to explore the heterogeneity of eating behaviors measured by a self-assessment instrument using 424 healthy adults (mean ± standard deviation [SD] age = 47.07 ± 18.89 years; 67% female). We generated low-dimensional representations of functional connectivity using resting-state functional magnetic resonance imaging and estimated latent features using the feature representation capabilities of an autoencoder by nonlinearly compressing the functional connectivity information. The clustering approaches applied to latent features identified three distinct subgroups. The subgroups exhibited different levels of hunger traits, while their body mass indices were comparable. The results were replicated in an independent dataset consisting of 212 participants (mean ± SD age = 38.97 ± 19.80 years; 35% female). The model interpretation technique of integrated gradients revealed that the between-group differences in the integrated gradient maps were associated with functional reorganization in heteromodal association and limbic cortices and reward-related subcortical structures such as the accumbens, amygdala, and caudate. The cognitive decoding analysis revealed that these systems are associated with reward- and emotion-related systems. Our findings provide insights into the macroscopic brain organization of eating behavior-related subgroups independent of obesity.
Collapse
Affiliation(s)
- Hyoungshin Choi
- Department of Electrical and Computer EngineeringSungkyunkwan UniversitySuwonRepublic of Korea
- Center for Neuroscience Imaging ResearchInstitute for Basic ScienceSuwonRepublic of Korea
| | | | - Jong‐eun Lee
- Department of Electrical and Computer EngineeringSungkyunkwan UniversitySuwonRepublic of Korea
- Center for Neuroscience Imaging ResearchInstitute for Basic ScienceSuwonRepublic of Korea
| | - Seok‐Jun Hong
- Center for Neuroscience Imaging ResearchInstitute for Basic ScienceSuwonRepublic of Korea
- Center for the Developing BrainChild Mind InstituteNew YorkUSA
- Department of Biomedical EngineeringSungkyunkwan UniversitySuwonRepublic of Korea
| | - Bo‐yong Park
- Center for Neuroscience Imaging ResearchInstitute for Basic ScienceSuwonRepublic of Korea
- Department of Data ScienceInha UniversityIncheonRepublic of Korea
- Department of Statistics and Data ScienceInha UniversityIncheonRepublic of Korea
| | - Hyunjin Park
- Center for Neuroscience Imaging ResearchInstitute for Basic ScienceSuwonRepublic of Korea
- School of Electronic and Electrical EngineeringSungkyunkwan UniversitySuwonRepublic of Korea
| |
Collapse
|
3
|
Aamir A, Tamosiunaite M, Wörgötter F. Interpreting the decisions of CNNs via influence functions. Front Comput Neurosci 2023; 17:1172883. [PMID: 37564901 PMCID: PMC10410673 DOI: 10.3389/fncom.2023.1172883] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2023] [Accepted: 06/26/2023] [Indexed: 08/12/2023] Open
Abstract
An understanding of deep neural network decisions is based on the interpretability of model, which provides explanations that are understandable to human beings and helps avoid biases in model predictions. This study investigates and interprets the model output based on images from the training dataset, i.e., to debug the results of a network model in relation to the training dataset. Our objective was to understand the behavior (specifically, class prediction) of deep learning models through the analysis of perturbations of the loss functions. We calculated influence scores for the VGG16 network at different hidden layers across three types of disturbances in the original images of the ImageNet dataset: texture, style, and background elimination. The global and layer-wise influence scores allowed the identification of the most influential training images for the given testing set. We illustrated our findings using influence scores by highlighting the types of disturbances that bias predictions of the network. According to our results, layer-wise influence analysis pairs well with local interpretability methods such as Shapley values to demonstrate significant differences between disturbed image subgroups. Particularly in an image classification task, our layer-wise interpretability approach plays a pivotal role to identify the classification bias in pre-trained convolutional neural networks, thus, providing useful insights to retrain specific hidden layers.
Collapse
Affiliation(s)
- Aisha Aamir
- Third Institute of Physics – Biophysics and Bernstein Center for Computational Neuroscience, University of Göttingen, Göttingen, Germany
| | - Minija Tamosiunaite
- Third Institute of Physics – Biophysics and Bernstein Center for Computational Neuroscience, University of Göttingen, Göttingen, Germany
- Department of Informatics, Vytautas Magnus University, Kaunas, Lithuania
| | - Florentin Wörgötter
- Third Institute of Physics – Biophysics and Bernstein Center for Computational Neuroscience, University of Göttingen, Göttingen, Germany
| |
Collapse
|
4
|
Thomas AW, Ré C, Poldrack RA. Benchmarking explanation methods for mental state decoding with deep learning models. Neuroimage 2023; 273:120109. [PMID: 37059157 DOI: 10.1016/j.neuroimage.2023.120109] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2022] [Revised: 04/04/2023] [Accepted: 04/11/2023] [Indexed: 04/16/2023] Open
Abstract
Deep learning (DL) models find increasing application in mental state decoding, where researchers seek to understand the mapping between mental states (e.g., experiencing anger or joy) and brain activity by identifying those spatial and temporal features of brain activity that allow to accurately identify (i.e., decode) these states. Once a DL model has been trained to accurately decode a set of mental states, neuroimaging researchers often make use of methods from explainable artificial intelligence research to understand the model's learned mappings between mental states and brain activity. Here, we benchmark prominent explanation methods in a mental state decoding analysis of multiple functional Magnetic Resonance Imaging (fMRI) datasets. Our findings demonstrate a gradient between two key characteristics of an explanation in mental state decoding, namely, its faithfulness and its alignment with other empirical evidence on the mapping between brain activity and decoded mental state: explanation methods with high explanation faithfulness, which capture the model's decision process well, generally provide explanations that align less well with other empirical evidence than the explanations of methods with less faithfulness. Based on our findings, we provide guidance for neuroimaging researchers on how to choose an explanation method to gain insight into the mental state decoding decisions of DL models.
Collapse
Affiliation(s)
- Armin W Thomas
- Stanford Data Science, Stanford University, 450 Serra Mall, 94305, Stanford, USA.
| | - Christopher Ré
- Dept. of Computer Science, Stanford University, 450 Serra Mall, 94305, Stanford, USA
| | - Russell A Poldrack
- Dept. of Psychology, Stanford University, 450 Serra Mall, Stanford, 94305, USA
| |
Collapse
|
5
|
Gopinath N. Artificial intelligence and neuroscience: An update on fascinating relationships. Process Biochem 2023. [DOI: 10.1016/j.procbio.2022.12.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 12/23/2022]
|
6
|
Farahani FV, Fiok K, Lahijanian B, Karwowski W, Douglas PK. Explainable AI: A review of applications to neuroimaging data. Front Neurosci 2022; 16:906290. [PMID: 36583102 PMCID: PMC9793854 DOI: 10.3389/fnins.2022.906290] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2022] [Accepted: 11/14/2022] [Indexed: 12/05/2022] Open
Abstract
Deep neural networks (DNNs) have transformed the field of computer vision and currently constitute some of the best models for representations learned via hierarchical processing in the human brain. In medical imaging, these models have shown human-level performance and even higher in the early diagnosis of a wide range of diseases. However, the goal is often not only to accurately predict group membership or diagnose but also to provide explanations that support the model decision in a context that a human can readily interpret. The limited transparency has hindered the adoption of DNN algorithms across many domains. Numerous explainable artificial intelligence (XAI) techniques have been developed to peer inside the "black box" and make sense of DNN models, taking somewhat divergent approaches. Here, we suggest that these methods may be considered in light of the interpretation goal, including functional or mechanistic interpretations, developing archetypal class instances, or assessing the relevance of certain features or mappings on a trained model in a post-hoc capacity. We then focus on reviewing recent applications of post-hoc relevance techniques as applied to neuroimaging data. Moreover, this article suggests a method for comparing the reliability of XAI methods, especially in deep neural networks, along with their advantages and pitfalls.
Collapse
Affiliation(s)
- Farzad V. Farahani
- Department of Biostatistics, Johns Hopkins University, Baltimore, MD, United States
- Department of Industrial Engineering and Management Systems, University of Central Florida, Orlando, FL, United States
| | - Krzysztof Fiok
- Department of Industrial Engineering and Management Systems, University of Central Florida, Orlando, FL, United States
| | - Behshad Lahijanian
- Department of Industrial and Systems Engineering, University of Florida, Gainesville, FL, United States
- H. Milton Stewart School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, GA, United States
| | - Waldemar Karwowski
- Department of Industrial Engineering and Management Systems, University of Central Florida, Orlando, FL, United States
| | - Pamela K. Douglas
- School of Modeling, Simulation, and Training, University of Central Florida, Orlando, FL, United States
| |
Collapse
|
7
|
Bassi PRAS, Attux R. Covid-19 detection using chest X-rays: is lung segmentation important for generalization? RESEARCH ON BIOMEDICAL ENGINEERING 2022. [PMCID: PMC9628459 DOI: 10.1007/s42600-022-00242-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Purpose We evaluated the generalization capability of deep neural networks (DNNs) in the task of classifying chest X-rays as Covid-19, normal or pneumonia, when trained in a relatively small and mixed datasets. Methods We proposed a DNN to perform lung segmentation and classification, stacking a segmentation module (U-Net), an original intermediate module and a classification module (DenseNet201). To evaluate generalization capability, we tested the network with an external dataset (from distinct localities) and used Bayesian inference to estimate the probability distributions of performance metrics. Furthermore, we introduce a novel evaluation technique, which uses layer-wise relevance propagation (LRP) and Brixia scores to compare the DNN grounds for decision with radiologists. Results The proposed DNN achieved 0.917 AUC (area under the ROC curve) on the external test dataset, surpassing a DenseNet without segmentation, which showed 0.906 AUC. Bayesian inference indicated mean accuracy of 76.1% and [0.695, 0.826] 95% HDI (high-density interval, which concentrates 95% of the metric’s probability mass) with segmentation and, without segmentation, 71.7% and [0.646, 0.786]. Conclusion Employing an analysis based on LRP and Brixia scores, we discovered that areas where radiologists found strong Covid-19 symptoms are the most important for the stacked DNN classification. External validation showed smaller accuracies than internal, indicating difficulty in generalization, which is positively affected by lung segmentation. Finally, the performance on the external dataset and the analysis with LRP suggest that DNNs can successfully detect Covid-19 even when trained on small and mixed datasets.
Collapse
Affiliation(s)
- Pedro R. A. S. Bassi
- Department of Computer Engineering and Industrial Automation, School of Electrical and Computer Engineering, University of Campinas - UNICAMP, Campinas, SP 13083-970 Brazil ,Present Address: Alma Mater Studiorum - University of Bologna, 40126 Bologna, BO Italy ,Present Address: Istituto Italiano Di Tecnologia, 16163 Genoa, GE Italy
| | - Romis Attux
- Department of Computer Engineering and Industrial Automation, School of Electrical and Computer Engineering, University of Campinas - UNICAMP, Campinas, SP 13083-970 Brazil
| |
Collapse
|
8
|
Thomas AW, Ré C, Poldrack RA. Interpreting mental state decoding with deep learning models. Trends Cogn Sci 2022; 26:972-986. [PMID: 36223760 DOI: 10.1016/j.tics.2022.07.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2021] [Revised: 07/15/2022] [Accepted: 07/18/2022] [Indexed: 01/12/2023]
Abstract
In mental state decoding, researchers aim to identify the set of mental states (e.g., experiencing happiness or fear) that can be reliably identified from the activity patterns of a brain region (or network). Deep learning (DL) models are highly promising for mental state decoding because of their unmatched ability to learn versatile representations of complex data. However, their widespread application in mental state decoding is hindered by their lack of interpretability, difficulties in applying them to small datasets, and in ensuring their reproducibility and robustness. We recommend approaching these challenges by leveraging recent advances in explainable artificial intelligence (XAI) and transfer learning, and also provide recommendations on how to improve the reproducibility and robustness of DL models in mental state decoding.
Collapse
Affiliation(s)
- Armin W Thomas
- Stanford Data Science, Stanford University, Stanford, CA, USA; Department of Psychology, Stanford University, Stanford, CA, USA.
| | - Christopher Ré
- Department of Computer Science, Stanford University, Stanford, CA, USA
| | - Russell A Poldrack
- Stanford Data Science, Stanford University, Stanford, CA, USA; Department of Psychology, Stanford University, Stanford, CA, USA
| |
Collapse
|
9
|
Sheu RK, Pardeshi MS. A Survey on Medical Explainable AI (XAI): Recent Progress, Explainability Approach, Human Interaction and Scoring System. SENSORS (BASEL, SWITZERLAND) 2022; 22:8068. [PMID: 36298417 PMCID: PMC9609212 DOI: 10.3390/s22208068] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/23/2022] [Revised: 10/18/2022] [Accepted: 10/18/2022] [Indexed: 06/16/2023]
Abstract
The emerging field of eXplainable AI (XAI) in the medical domain is considered to be of utmost importance. Meanwhile, incorporating explanations in the medical domain with respect to legal and ethical AI is necessary to understand detailed decisions, results, and current status of the patient's conditions. Successively, we will be presenting a detailed survey for the medical XAI with the model enhancements, evaluation methods, significant overview of case studies with open box architecture, medical open datasets, and future improvements. Potential differences in AI and XAI methods are provided with the recent XAI methods stated as (i) local and global methods for preprocessing, (ii) knowledge base and distillation algorithms, and (iii) interpretable machine learning. XAI characteristics details with future healthcare explainability is included prominently, whereas the pre-requisite provides insights for the brainstorming sessions before beginning a medical XAI project. Practical case study determines the recent XAI progress leading to the advance developments within the medical field. Ultimately, this survey proposes critical ideas surrounding a user-in-the-loop approach, with an emphasis on human-machine collaboration, to better produce explainable solutions. The surrounding details of the XAI feedback system for human rating-based evaluation provides intelligible insights into a constructive method to produce human enforced explanation feedback. For a long time, XAI limitations of the ratings, scores and grading are present. Therefore, a novel XAI recommendation system and XAI scoring system are designed and approached from this work. Additionally, this paper encourages the importance of implementing explainable solutions into the high impact medical field.
Collapse
Affiliation(s)
- Ruey-Kai Sheu
- Department of Computer Science, Tunghai University, No. 1727, Section 4, Taiwan Blvd, Xitun District, Taichung 407224, Taiwan
| | - Mayuresh Sunil Pardeshi
- AI Center, Tunghai University, No. 1727, Section 4, Taiwan Blvd, Xitun District, Taichung 407224, Taiwan
| |
Collapse
|
10
|
Kim D, Lee J, Moon J, Moon T. Interpretable deep learning-based hippocampal sclerosis classification. Epilepsia Open 2022; 7:747-757. [PMID: 36177546 PMCID: PMC9712484 DOI: 10.1002/epi4.12655] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2022] [Accepted: 09/26/2022] [Indexed: 12/30/2022] Open
Abstract
OBJECTIVE To evaluate the performance of a deep learning model for hippocampal sclerosis classification on the clinical dataset and suggest plausible visual interpretation for the model prediction. METHODS T2-weighted oblique coronal images of the brain MRI epilepsy protocol performed on patients were used. The training set included 320 participants with 160 no, 100 left and 60 right hippocampal sclerosis, and cross-validation was implemented. The test set consisted of 302 participants with 252 no, 25 left and 25 right hippocampal sclerosis. As the test set was imbalanced, we took an average of the accuracy achieved within each group to measure a balanced accuracy for multiclass and binary classifications. The dataset was composed to include not only healthy participants but also participants with abnormalities besides hippocampal sclerosis in the control group. We visualized the reasons for the model prediction using the layer-wise relevance propagation method. RESULTS When evaluated on the validation of the training set, we achieved multiclass and binary classification accuracy of 87.5% and 88.8% from the voting ensemble of six models. Evaluated on the test sets, we achieved multiclass and binary classification accuracy of 91.5% and 89.76%. The distinctly sparse visual interpretations were provided for each individual participant and group to suggest the contribution of each input voxel to the prediction on the MRI. SIGNIFICANCE The current interpretable deep learning-based model is promising for adapting effectively to clinical settings by utilizing commonly used data, such as MRI, with realistic abnormalities faced by neurologists to support the diagnosis of hippocampal sclerosis with plausible visual interpretation.
Collapse
Affiliation(s)
- Dohyun Kim
- Department of Artificial IntelligenceSungkyunkwan UniversitySuwonSouth Korea
| | - Jungtae Lee
- Application Engineering Team, Memory BusinessSamsung Electronics Co., Ltd.SuwonSouth Korea
| | - Jangsup Moon
- Department of NeurologySeoul National University HospitalSeoulSouth Korea,Department of Genomic MedicineSeoul National University HospitalSeoulSouth Korea
| | - Taesup Moon
- Department of Electrical and Computer EngineeringSeoul National UniversitySeoulSouth Korea,ASRI/INMC/IPAI/AIISSeoul National UniversitySeoulSouth Korea
| |
Collapse
|
11
|
Hofmann SM, Beyer F, Lapuschkin S, Goltermann O, Loeffler M, Müller KR, Villringer A, Samek W, Witte AV. Towards the Interpretability of Deep Learning Models for Multi-modal Neuroimaging: Finding Structural Changes of the Ageing Brain. Neuroimage 2022; 261:119504. [PMID: 35882272 DOI: 10.1016/j.neuroimage.2022.119504] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2022] [Revised: 07/15/2022] [Accepted: 07/21/2022] [Indexed: 11/17/2022] Open
Abstract
Brain-age (BA) estimates based on deep learning are increasingly used as neuroimaging biomarker for brain health; however, the underlying neural features have remained unclear. We combined ensembles of convolutional neural networks with Layer-wise Relevance Propagation (LRP) to detect which brain features contribute to BA. Trained on magnetic resonance imaging (MRI) data of a population-based study (n=2637, 18-82 years), our models estimated age accurately based on single and multiple modalities, regionally restricted and whole-brain images (mean absolute errors 3.37-3.86 years). We find that BA estimates capture aging at both small and large-scale changes, revealing gross enlargements of ventricles and subarachnoid spaces, as well as white matter lesions, and atrophies that appear throughout the brain. Divergence from expected aging reflected cardiovascular risk factors and accelerated aging was more pronounced in the frontal lobe. Applying LRP, our study demonstrates how superior deep learning models detect brain-aging in healthy and at-risk individuals throughout adulthood.
Collapse
Affiliation(s)
- Simon M Hofmann
- Department of Neurology, Max Planck Institute for Human Cognitive and Brain Sciences, 04103 Leipzig, Germany; Department of Artificial Intelligence, Fraunhofer Institute Heinrich Hertz, 10587 Berlin, Germany; Clinic for Cognitive Neurology, University of Leipzig Medical Center, 04103 Leipzig, Germany.
| | - Frauke Beyer
- Department of Neurology, Max Planck Institute for Human Cognitive and Brain Sciences, 04103 Leipzig, Germany; Clinic for Cognitive Neurology, University of Leipzig Medical Center, 04103 Leipzig, Germany
| | - Sebastian Lapuschkin
- Department of Artificial Intelligence, Fraunhofer Institute Heinrich Hertz, 10587 Berlin, Germany
| | - Ole Goltermann
- Department of Neurology, Max Planck Institute for Human Cognitive and Brain Sciences, 04103 Leipzig, Germany; Max Planck School of Cognition, 04103 Leipzig, Germany; Institute of Systems Neuroscience, University Medical Center Hamburg-Eppendorf, Germany
| | | | - Klaus-Robert Müller
- Department of Electrical Engineering and Computer Science, Technical University Berlin, 10623 Berlin, Germany; Department of Artificial Intelligence, Korea University, 02841 Seoul, Korea (the Republic of); Brain Team, Google Research, 10117 Berlin, Germany; Max Planck Institute for Informatics, 66123 Saarbrücken, Germany; BIFOLD - Berlin Institute for the Foundations of Learning and Data, 10587 Berlin, Germany
| | - Arno Villringer
- Department of Neurology, Max Planck Institute for Human Cognitive and Brain Sciences, 04103 Leipzig, Germany; Clinic for Cognitive Neurology, University of Leipzig Medical Center, 04103 Leipzig, Germany; MindBrainBody Institute, Berlin School of Mind and Brain, Humboldt-Universität zu Berlin, 10099 Berlin, Germany; Center for Stroke Research, Charité - Universitätsmedizin Berlin, 10117 Berlin, Germany
| | - Wojciech Samek
- Department of Artificial Intelligence, Fraunhofer Institute Heinrich Hertz, 10587 Berlin, Germany; Department of Electrical Engineering and Computer Science, Technical University Berlin, 10623 Berlin, Germany; BIFOLD - Berlin Institute for the Foundations of Learning and Data, 10587 Berlin, Germany
| | - A Veronica Witte
- Department of Neurology, Max Planck Institute for Human Cognitive and Brain Sciences, 04103 Leipzig, Germany; Clinic for Cognitive Neurology, University of Leipzig Medical Center, 04103 Leipzig, Germany
| |
Collapse
|
12
|
Abstract
Brain dynamics are highly complex and yet hold the key to understanding brain function and dysfunction. The dynamics captured by resting-state functional magnetic resonance imaging data are noisy, high-dimensional, and not readily interpretable. The typical approach of reducing this data to low-dimensional features and focusing on the most predictive features comes with strong assumptions and can miss essential aspects of the underlying dynamics. In contrast, introspection of discriminatively trained deep learning models may uncover disorder-relevant elements of the signal at the level of individual time points and spatial locations. Yet, the difficulty of reliable training on high-dimensional low sample size datasets and the unclear relevance of the resulting predictive markers prevent the widespread use of deep learning in functional neuroimaging. In this work, we introduce a deep learning framework to learn from high-dimensional dynamical data while maintaining stable, ecologically valid interpretations. Results successfully demonstrate that the proposed framework enables learning the dynamics of resting-state fMRI directly from small data and capturing compact, stable interpretations of features predictive of function and dysfunction.
Collapse
|
13
|
Jiang R, Woo CW, Qi S, Wu J, Sui J. Interpreting Brain Biomarkers: Challenges and solutions in interpreting machine learning-based predictive neuroimaging. IEEE SIGNAL PROCESSING MAGAZINE 2022; 39:107-118. [PMID: 36712588 PMCID: PMC9880880 DOI: 10.1109/msp.2022.3155951] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
Predictive modeling of neuroimaging data (predictive neuroimaging) for evaluating individual differences in various behavioral phenotypes and clinical outcomes is of growing interest. However, the field is experiencing challenges regarding the interpretability of the results. Approaches to defining the specific contribution of functional connections, regions, or networks in prediction models are urgently needed, which may help explore the underlying mechanisms. In this article, we systematically review the methods and applications for interpreting brain signatures derived from predictive neuroimaging based on a survey of 326 research articles. Strengths, limitations, and the suitable conditions for major interpretation strategies are also deliberated. In-depth discussion of common issues in existing literature and the corresponding recommendations to address these pitfalls are provided. We highly recommend exhaustive validation on the reliability and interpretability of the biomarkers across multiple datasets and contexts, which thereby could translate technical advances in neuroimaging into concrete improvements in precision medicine.
Collapse
Affiliation(s)
- Rongtao Jiang
- Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA, 06520
| | - Choong-Wan Woo
- Center for Neuroscience Imaging Research, Institute for Basic Science, Suwon, South Korea, 16419
- Department of Biomedical Engineering, Sungkyunkwan University, Suwon, South Korea, 16419
- Department of Intelligent Precision Healthcare Convergence, Sungkyunkwan University, Suwon, Korea, 16419
| | - Shile Qi
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China, 211106
| | - Jing Wu
- Department of Medical Oncology, Beijing You-An Hospital, Capital Medical University, Beijing, China, 100069
| | - Jing Sui
- State Key Laboratory of Brain Cognition and Learning, Beijing Normal University, Beijing, China, 100875
| |
Collapse
|
14
|
Kalmady SV, Paul AK, Narayanaswamy JC, Agrawal R, Shivakumar V, Greenshaw AJ, Dursun SM, Greiner R, Venkatasubramanian G, Reddy YCJ. Prediction of Obsessive-Compulsive Disorder: Importance of Neurobiology-Aided Feature Design and Cross-Diagnosis Transfer Learning. BIOLOGICAL PSYCHIATRY. COGNITIVE NEUROSCIENCE AND NEUROIMAGING 2022; 7:735-746. [PMID: 34929344 DOI: 10.1016/j.bpsc.2021.12.003] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/05/2021] [Revised: 11/25/2021] [Accepted: 12/07/2021] [Indexed: 06/14/2023]
Abstract
BACKGROUND Machine learning applications using neuroimaging provide a multidimensional, data-driven approach that captures the level of complexity necessary for objectively aiding diagnosis and prognosis in psychiatry. However, models learned from small training samples often have limited generalizability, which continues to be a problem with automated diagnosis of mental illnesses such as obsessive-compulsive disorder (OCD). Earlier studies have shown that features incorporating prior neurobiological knowledge of brain function and combining brain parcellations from various sources can potentially improve the overall prediction. However, it is unknown whether such knowledge-driven methods can provide a performance that is comparable to state-of-the-art approaches based on neural networks. METHODS In this study, we apply a transparent and explainable multiparcellation ensemble learning framework EMPaSchiz (Ensemble algorithm with Multiple Parcellations for Schizophrenia prediction) to the task of predicting OCD, based on a resting-state functional magnetic resonance imaging dataset of 350 subjects. Furthermore, we apply transfer learning using the features found effective for schizophrenia to OCD to leverage the commonality in brain alterations across these psychiatric diagnoses. RESULTS We show that our knowledge-based approach leads to a prediction performance of 80.3% accuracy for OCD diagnosis that is better than domain-agnostic and automated feature design using neural networks. Furthermore, we show that a selection of reduced feature sets can be transferred from schizophrenia to the OCD prediction model without significant loss in prediction performance. CONCLUSIONS This study presents a machine learning framework for OCD prediction with neurobiology-aided feature design using resting-state functional magnetic resonance imaging that is generalizable and reasonably interpretable.
Collapse
Affiliation(s)
- Sunil Vasu Kalmady
- Alberta Machine Intelligence Institute, Department of Computing Science, University of Alberta, Edmonton, Alberta, Canada; Canadian VIGOUR Centre, University of Alberta, Edmonton, Alberta, Canada.
| | - Animesh Kumar Paul
- Alberta Machine Intelligence Institute, Department of Computing Science, University of Alberta, Edmonton, Alberta, Canada
| | - Janardhanan C Narayanaswamy
- OCD Clinic, Department of Psychiatry, National Institute of Mental Health and Neuro Sciences, Bangalore, India; Translational Psychiatry Laboratory, Neurobiology Research Centre, National Institute of Mental Health and Neuro Sciences, Bangalore, India
| | - Rimjhim Agrawal
- Translational Psychiatry Laboratory, Neurobiology Research Centre, National Institute of Mental Health and Neuro Sciences, Bangalore, India
| | - Venkataram Shivakumar
- OCD Clinic, Department of Psychiatry, National Institute of Mental Health and Neuro Sciences, Bangalore, India; Translational Psychiatry Laboratory, Neurobiology Research Centre, National Institute of Mental Health and Neuro Sciences, Bangalore, India
| | - Andrew J Greenshaw
- Department of Psychiatry, Faculty of Medicine and Dentistry, University of Alberta, Edmonton, Alberta, Canada
| | - Serdar M Dursun
- Department of Psychiatry, Faculty of Medicine and Dentistry, University of Alberta, Edmonton, Alberta, Canada
| | - Russell Greiner
- Alberta Machine Intelligence Institute, Department of Computing Science, University of Alberta, Edmonton, Alberta, Canada; Department of Psychiatry, Faculty of Medicine and Dentistry, University of Alberta, Edmonton, Alberta, Canada
| | - Ganesan Venkatasubramanian
- OCD Clinic, Department of Psychiatry, National Institute of Mental Health and Neuro Sciences, Bangalore, India; Translational Psychiatry Laboratory, Neurobiology Research Centre, National Institute of Mental Health and Neuro Sciences, Bangalore, India.
| | - Y C Janardhan Reddy
- OCD Clinic, Department of Psychiatry, National Institute of Mental Health and Neuro Sciences, Bangalore, India; Translational Psychiatry Laboratory, Neurobiology Research Centre, National Institute of Mental Health and Neuro Sciences, Bangalore, India
| |
Collapse
|
15
|
Lee S, Bradlow ET, Kable JW. Fast construction of interpretable whole-brain decoders. CELL REPORTS METHODS 2022; 2:100227. [PMID: 35784649 PMCID: PMC9243546 DOI: 10.1016/j.crmeth.2022.100227] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 10/29/2021] [Revised: 04/11/2022] [Accepted: 05/16/2022] [Indexed: 01/15/2023]
Abstract
Researchers often seek to decode mental states from brain activity measured with functional MRI. Rigorous decoding requires the use of formal neural prediction models, which are likely to be the most accurate if they use the whole brain. However, the computational burden and lack of interpretability of off-the-shelf statistical methods can make whole-brain decoding challenging. Here, we propose a method to build whole-brain neural decoders that are both interpretable and computationally efficient. We extend the partial least squares algorithm to build a regularized model with variable selection that offers a unique "fit once, tune later" approach: users need to fit the model only once and can choose the best tuning parameters post hoc. We show in real data that our method scales well with increasing data size and yields interpretable predictors. The algorithm is publicly available in multiple languages in the hope that interpretable whole-brain predictors can be implemented more widely in neuroimaging research.
Collapse
Affiliation(s)
- Sangil Lee
- Department of Psychology, School of Arts and Sciences, University of Pennsylvania, Philadelphia, PA 19104, USA
- Marketing Department, Wharton School, University of Pennsylvania, PA 19104, USA
- Social Science Matrix, University of California, Berkeley, Berkeley, CA 94720, USA
| | - Eric T. Bradlow
- Marketing Department, Wharton School, University of Pennsylvania, PA 19104, USA
| | - Joseph W. Kable
- Department of Psychology, School of Arts and Sciences, University of Pennsylvania, Philadelphia, PA 19104, USA
- Marketing Department, Wharton School, University of Pennsylvania, PA 19104, USA
| |
Collapse
|
16
|
Mertes S, Huber T, Weitz K, Heimerl A, André E. GANterfactual—Counterfactual Explanations for Medical Non-experts Using Generative Adversarial Learning. Front Artif Intell 2022; 5:825565. [PMID: 35464995 PMCID: PMC9024220 DOI: 10.3389/frai.2022.825565] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2021] [Accepted: 03/10/2022] [Indexed: 11/13/2022] Open
Abstract
With the ongoing rise of machine learning, the need for methods for explaining decisions made by artificial intelligence systems is becoming a more and more important topic. Especially for image classification tasks, many state-of-the-art tools to explain such classifiers rely on visual highlighting of important areas of the input data. Contrary, counterfactual explanation systems try to enable a counterfactual reasoning by modifying the input image in a way such that the classifier would have made a different prediction. By doing so, the users of counterfactual explanation systems are equipped with a completely different kind of explanatory information. However, methods for generating realistic counterfactual explanations for image classifiers are still rare. Especially in medical contexts, where relevant information often consists of textural and structural information, high-quality counterfactual images have the potential to give meaningful insights into decision processes. In this work, we present GANterfactual, an approach to generate such counterfactual image explanations based on adversarial image-to-image translation techniques. Additionally, we conduct a user study to evaluate our approach in an exemplary medical use case. Our results show that, in the chosen medical use-case, counterfactual explanations lead to significantly better results regarding mental models, explanation satisfaction, trust, emotions, and self-efficacy than two state-of-the art systems that work with saliency maps, namely LIME and LRP.
Collapse
|
17
|
Salahuddin Z, Woodruff HC, Chatterjee A, Lambin P. Transparency of deep neural networks for medical image analysis: A review of interpretability methods. Comput Biol Med 2022; 140:105111. [PMID: 34891095 DOI: 10.1016/j.compbiomed.2021.105111] [Citation(s) in RCA: 71] [Impact Index Per Article: 35.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2021] [Revised: 12/01/2021] [Accepted: 12/02/2021] [Indexed: 02/03/2023]
Abstract
Artificial Intelligence (AI) has emerged as a useful aid in numerous clinical applications for diagnosis and treatment decisions. Deep neural networks have shown the same or better performance than clinicians in many tasks owing to the rapid increase in the available data and computational power. In order to conform to the principles of trustworthy AI, it is essential that the AI system be transparent, robust, fair, and ensure accountability. Current deep neural solutions are referred to as black-boxes due to a lack of understanding of the specifics concerning the decision-making process. Therefore, there is a need to ensure the interpretability of deep neural networks before they can be incorporated into the routine clinical workflow. In this narrative review, we utilized systematic keyword searches and domain expertise to identify nine different types of interpretability methods that have been used for understanding deep learning models for medical image analysis applications based on the type of generated explanations and technical similarities. Furthermore, we report the progress made towards evaluating the explanations produced by various interpretability methods. Finally, we discuss limitations, provide guidelines for using interpretability methods and future directions concerning the interpretability of deep neural networks for medical imaging analysis.
Collapse
Affiliation(s)
- Zohaib Salahuddin
- The D-Lab, Department of Precision Medicine, GROW - School for Oncology and Developmental Biology, Maastricht University, Maastricht, the Netherlands.
| | - Henry C Woodruff
- The D-Lab, Department of Precision Medicine, GROW - School for Oncology and Developmental Biology, Maastricht University, Maastricht, the Netherlands; Department of Radiology and Nuclear Medicine, GROW - School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, the Netherlands
| | - Avishek Chatterjee
- The D-Lab, Department of Precision Medicine, GROW - School for Oncology and Developmental Biology, Maastricht University, Maastricht, the Netherlands
| | - Philippe Lambin
- The D-Lab, Department of Precision Medicine, GROW - School for Oncology and Developmental Biology, Maastricht University, Maastricht, the Netherlands; Department of Radiology and Nuclear Medicine, GROW - School for Oncology and Developmental Biology, Maastricht University Medical Centre+, Maastricht, the Netherlands
| |
Collapse
|
18
|
Hu R, Peng Z, Zhu X, Gan J, Zhu Y, Ma J, Wu G. Multi-Band Brain Network Analysis for Functional Neuroimaging Biomarker Identification. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3843-3855. [PMID: 34310294 PMCID: PMC8931676 DOI: 10.1109/tmi.2021.3099641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
The functional connectomic profile is one of the non-invasive imaging biomarkers in the computer-assisted diagnostic system for many neuro-diseases. However, the diagnostic power of functional connectivity is challenged by mixed frequency-specific neuronal oscillations in the brain, which makes the single Functional Connectivity Network (FCN) often underpowered to capture the disease-related functional patterns. To address this challenge, we propose a novel functional connectivity analysis framework to conduct joint feature learning and personalized disease diagnosis, in a semi-supervised manner, aiming at focusing on putative multi-band functional connectivity biomarkers from functional neuroimaging data. Specifically, we first decompose the Blood Oxygenation Level Dependent (BOLD) signals into multiple frequency bands by the discrete wavelet transform, and then cast the alignment of all fully-connected FCNs derived from multiple frequency bands into a parameter-free multi-band fusion model. The proposed fusion model fuses all fully-connected FCNs to obtain a sparsely-connected FCN (sparse FCN for short) for each individual subject, as well as lets each sparse FCN be close to its neighbored sparse FCNs and be far away from its furthest sparse FCNs. Furthermore, we employ the l1 -SVM to conduct joint brain region selection and disease diagnosis. Finally, we evaluate the effectiveness of our proposed framework on various neuro-diseases, i.e., Fronto-Temporal Dementia (FTD), Obsessive-Compulsive Disorder (OCD), and Alzheimer's Disease (AD), and the experimental results demonstrate that our framework shows more reasonable results, compared to state-of-the-art methods, in terms of classification performance and the selected brain regions. The source code can be visited by the url https://github.com/reynard-hu/mbbna.
Collapse
|
19
|
Tjoa E, Guan C. A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2021; 32:4793-4813. [PMID: 33079674 DOI: 10.1109/tnnls.2020.3027314] [Citation(s) in RCA: 311] [Impact Index Per Article: 103.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/06/2023]
Abstract
Recently, artificial intelligence and machine learning in general have demonstrated remarkable performances in many tasks, from image processing to natural language processing, especially with the advent of deep learning (DL). Along with research progress, they have encroached upon many different fields and disciplines. Some of them require high level of accountability and thus transparency, for example, the medical sector. Explanations for machine decisions and predictions are thus needed to justify their reliability. This requires greater interpretability, which often means we need to understand the mechanism underlying the algorithms. Unfortunately, the blackbox nature of the DL is still unresolved, and many machine decisions are still poorly understood. We provide a review on interpretabilities suggested by different research works and categorize them. The different categories show different dimensions in interpretability research, from approaches that provide "obviously" interpretable information to the studies of complex patterns. By applying the same categorization to interpretability in medical research, it is hoped that: 1) clinicians and practitioners can subsequently approach these methods with caution; 2) insight into interpretability will be born with more considerations for medical practices; and 3) initiatives to push forward data-based, mathematically grounded, and technically grounded medical education are encouraged.
Collapse
|
20
|
Macpherson T, Churchland A, Sejnowski T, DiCarlo J, Kamitani Y, Takahashi H, Hikida T. Natural and Artificial Intelligence: A brief introduction to the interplay between AI and neuroscience research. Neural Netw 2021; 144:603-613. [PMID: 34649035 DOI: 10.1016/j.neunet.2021.09.018] [Citation(s) in RCA: 29] [Impact Index Per Article: 9.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2021] [Revised: 09/15/2021] [Accepted: 09/21/2021] [Indexed: 10/20/2022]
Abstract
Neuroscience and artificial intelligence (AI) share a long history of collaboration. Advances in neuroscience, alongside huge leaps in computer processing power over the last few decades, have given rise to a new generation of in silico neural networks inspired by the architecture of the brain. These AI systems are now capable of many of the advanced perceptual and cognitive abilities of biological systems, including object recognition and decision making. Moreover, AI is now increasingly being employed as a tool for neuroscience research and is transforming our understanding of brain functions. In particular, deep learning has been used to model how convolutional layers and recurrent connections in the brain's cerebral cortex control important functions, including visual processing, memory, and motor control. Excitingly, the use of neuroscience-inspired AI also holds great promise for understanding how changes in brain networks result in psychopathologies, and could even be utilized in treatment regimes. Here we discuss recent advancements in four areas in which the relationship between neuroscience and AI has led to major advancements in the field; (1) AI models of working memory, (2) AI visual processing, (3) AI analysis of big neuroscience datasets, and (4) computational psychiatry.
Collapse
Affiliation(s)
- Tom Macpherson
- Laboratory for Advanced Brain Functions, Institute for Protein Research, Osaka University, Osaka, Japan
| | - Anne Churchland
- Cold Spring Harbor Laboratory, Neuroscience, Cold Spring Harbor, NY, USA
| | - Terry Sejnowski
- Computational Neurobiology Laboratory, Salk Institute for Biological Studies, CA, USA; Division of Biological Sciences, University of California San Diego, CA, USA
| | - James DiCarlo
- Brain and Cognitive Sciences, Massachusetts Institute of Technology, MA, USA
| | - Yukiyasu Kamitani
- Department of Neuroinformatics, ATR Computational Neuroscience Laboratories, Kyoto, Japan; Graduate School of Informatics, Kyoto University, Kyoto, Japan
| | - Hidehiko Takahashi
- Department of Psychiatry and Behavioral Sciences, Tokyo Medical and Dental University Graduate School, Tokyo, Japan
| | - Takatoshi Hikida
- Laboratory for Advanced Brain Functions, Institute for Protein Research, Osaka University, Osaka, Japan.
| |
Collapse
|
21
|
Misra J, Surampudi SG, Venkatesh M, Limbachia C, Jaja J, Pessoa L. Learning brain dynamics for decoding and predicting individual differences. PLoS Comput Biol 2021; 17:e1008943. [PMID: 34478442 PMCID: PMC8445454 DOI: 10.1371/journal.pcbi.1008943] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2021] [Revised: 09/16/2021] [Accepted: 08/19/2021] [Indexed: 12/04/2022] Open
Abstract
Insights from functional Magnetic Resonance Imaging (fMRI), as well as recordings of large numbers of neurons, reveal that many cognitive, emotional, and motor functions depend on the multivariate interactions of brain signals. To decode brain dynamics, we propose an architecture based on recurrent neural networks to uncover distributed spatiotemporal signatures. We demonstrate the potential of the approach using human fMRI data during movie-watching data and a continuous experimental paradigm. The model was able to learn spatiotemporal patterns that supported 15-way movie-clip classification (∼90%) at the level of brain regions, and binary classification of experimental conditions (∼60%) at the level of voxels. The model was also able to learn individual differences in measures of fluid intelligence and verbal IQ at levels comparable to that of existing techniques. We propose a dimensionality reduction approach that uncovers low-dimensional trajectories and captures essential informational (i.e., classification related) properties of brain dynamics. Finally, saliency maps and lesion analysis were employed to characterize brain-region/voxel importance, and uncovered how dynamic but consistent changes in fMRI activation influenced decoding performance. When applied at the level of voxels, our framework implements a dynamic version of multivariate pattern analysis. Our approach provides a framework for visualizing, analyzing, and discovering dynamic spatially distributed brain representations during naturalistic conditions.
Collapse
Affiliation(s)
- Joyneel Misra
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
| | - Srinivas Govinda Surampudi
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
| | - Manasij Venkatesh
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
| | - Chirag Limbachia
- Department of Psychology and Maryland Neuroimaging Center, University of Maryland, College Park, Maryland, United States of America
| | - Joseph Jaja
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
| | - Luiz Pessoa
- Department of Electrical and Computer Engineering, University of Maryland, College Park, Maryland, United States of America
- Department of Psychology and Maryland Neuroimaging Center, University of Maryland, College Park, Maryland, United States of America
| |
Collapse
|
22
|
Strodthoff N, Strodthoff C, Becher T, Weiler N, Frerichs I. Inferring Respiratory and Circulatory Parameters from Electrical Impedance Tomography With Deep Recurrent Models. IEEE J Biomed Health Inform 2021; 25:3105-3111. [PMID: 33577463 DOI: 10.1109/jbhi.2021.3059016] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Electrical impedance tomography (EIT) is a noninvasive imaging modality that allows a continuous assessment of changes in regional bioimpedance of different organs. One of its most common biomedical applications is monitoring regional ventilation distribution in critically ill patients treated in intensive care units. In this work, we put forward a proof-of-principle study that demonstrates how one can reconstruct synchronously measured respiratory or circulatory parameters from the EIT image sequence using a deep learning model trained in an end-to-end fashion. For this purpose, we devise an architecture with a convolutional feature extractor whose output is processed by a recurrent neural network. We demonstrate that one can accurately infer absolute volume, absolute flow, normalized airway pressure and within certain limitations even the normalized arterial blood pressure from the EIT signal alone, in a way that generalizes to unseen patients without prior calibration. As an outlook with direct clinical relevance, we furthermore demonstrate the feasibility of reconstructing the absolute transpulmonary pressure from a combination of EIT and absolute airway pressure, as a way to potentially replace the invasive measurement of esophageal pressure. With these results, we hope to stimulate further studies building on the framework put forward in this work.
Collapse
|
23
|
Thanjavur K, Babul A, Foran B, Bielecki M, Gilchrist A, Hristopulos DT, Brucar LR, Virji-Babul N. Recurrent neural network-based acute concussion classifier using raw resting state EEG data. Sci Rep 2021; 11:12353. [PMID: 34117309 PMCID: PMC8196170 DOI: 10.1038/s41598-021-91614-4] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2020] [Accepted: 05/24/2021] [Indexed: 02/05/2023] Open
Abstract
Concussion is a global health concern. Despite its high prevalence, a sound understanding of the mechanisms underlying this type of diffuse brain injury remains elusive. It is, however, well established that concussions cause significant functional deficits; that children and youths are disproportionately affected and have longer recovery time than adults; and that individuals suffering from a concussion are more prone to experience additional concussions, with each successive injury increasing the risk of long term neurological and mental health complications. Currently, the most significant challenge in concussion management is the lack of objective, clinically- accepted, brain-based approaches for determining whether an athlete has suffered a concussion. Here, we report on our efforts to address this challenge. Specifically, we introduce a deep learning long short-term memory (LSTM)-based recurrent neural network that is able to distinguish between non-concussed and acute post-concussed adolescent athletes using only short (i.e. 90 s long) samples of resting state EEG data as input. The athletes were neither required to perform a specific task nor expected to respond to a stimulus during data collection. The acquired EEG data were neither filtered, cleaned of artefacts, nor subjected to explicit feature extraction. The LSTM network was trained and validated using data from 27 male, adolescent athletes with sports related concussion, benchmarked against 35 non-concussed adolescent athletes. During rigorous testing, the classifier consistently identified concussions with an accuracy of > 90% and achieved an ensemble median Area Under the Receiver Operating Characteristic Curve (ROC/AUC) equal to 0.971. This is the first instance of a high-performing classifier that relies only on easy-to-acquire resting state, raw EEG data. Our concussion classifier represents a promising first step towards the development of an easy-to-use, objective, brain-based, automatic classification of concussion at an individual level.
Collapse
Affiliation(s)
- Karun Thanjavur
- Department of Physics and Astronomy, University of Victoria, Victoria, BC, V8P 5C2, Canada.
| | - Arif Babul
- Department of Physics and Astronomy, University of Victoria, Victoria, BC, V8P 5C2, Canada
| | - Brandon Foran
- Department of Computer Science, Middlesex College, Western University, London, ON, N6A 5B7, Canada
| | - Maya Bielecki
- Department of Computer Science, Middlesex College, Western University, London, ON, N6A 5B7, Canada
| | - Adam Gilchrist
- Department of Computer Science, Middlesex College, Western University, London, ON, N6A 5B7, Canada
| | - Dionissios T Hristopulos
- School of Electrical and Computer Engineering, Technical University of Crete, 73100, Chania, Greece
| | - Leyla R Brucar
- Djavad Mowafaghian Centre for Brain Health, University of British Columbia, Vancouver, BC, V6T 1Z3, Canada
| | - Naznin Virji-Babul
- Djavad Mowafaghian Centre for Brain Health, University of British Columbia, Vancouver, BC, V6T 1Z3, Canada
- Department of Physical Therapy, Faculty of Medicine, University of British Columbia, Vancouver, BC, V6T 1Z3, Canada
| |
Collapse
|
24
|
Abstract
Purpose We present image classifiers based on Dense Convolutional Networks and transfer learning to classify chest X-ray images according to three labels: COVID-19, pneumonia, and normal. Methods We fine-tuned neural networks pretrained on ImageNet and applied a twice transfer learning approach, using NIH ChestX-ray14 dataset as an intermediate step. We also suggested a novelty called output neuron keeping, which changes the twice transfer learning technique. In order to clarify the modus operandi of the models, we used Layer-wise Relevance Propagation (LRP) to generate heatmaps. Results We were able to reach test accuracy of 100% on our test dataset. Twice transfer learning and output neuron keeping showed promising results improving performances, mainly in the beginning of the training process. Although LRP revealed that words on the X-rays can influence the networks’ predictions, we discovered this had only a very small effect on accuracy. Conclusion Although clinical studies and larger datasets are still needed to further ensure good generalization, the state-of-the-art performances we achieved show that, with the help of artificial intelligence, chest X-rays can become a cheap and accurate auxiliary method for COVID-19 diagnosis. Heatmaps generated by LRP improve the interpretability of the deep neural networks and indicate an analytical path for future research on diagnosis. Twice transfer learning with output neuron keeping improved DNN performance.
Collapse
Affiliation(s)
- Pedro R. A. S. Bassi
- Department of Computer Engineering and Industrial Automation, School of Electrical and Computer Engineering, University of Campinas, Campinas, Brazil
| | - Romis Attux
- Department of Computer Engineering and Industrial Automation, School of Electrical and Computer Engineering, University of Campinas, Campinas, Brazil
| |
Collapse
|
25
|
Yotsutsuji S, Lei M, Akama H. Evaluation of Task fMRI Decoding With Deep Learning on a Small Sample Dataset. Front Neuroinform 2021; 15:577451. [PMID: 33679360 PMCID: PMC7928289 DOI: 10.3389/fninf.2021.577451] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2020] [Accepted: 01/25/2021] [Indexed: 01/12/2023] Open
Abstract
Recently, several deep learning methods have been applied to decoding in task-related fMRI, and their advantages have been exploited in a variety of ways. However, this paradigm is sometimes problematic, due to the difficulty of applying deep learning to high-dimensional data and small sample size conditions. The difficulties in gathering a large amount of data to develop predictive machine learning models with multiple layers from fMRI experiments with complicated designs and tasks are well-recognized. Group-level, multi-voxel pattern analysis with small sample sizes results in low statistical power and large accuracy evaluation errors; failure in such instances is ascribed to the individual variability that risks information leakage, a particular issue when dealing with a limited number of subjects. In this study, using a small-size fMRI dataset evaluating bilingual language switch in a property generation task, we evaluated the relative fit of different deep learning models, incorporating moderate split methods to control the amount of information leakage. Our results indicated that using the session shuffle split as the data folding method, along with the multichannel 2D convolutional neural network (M2DCNN) classifier, recorded the best authentic classification accuracy, which outperformed the efficiency of 3D convolutional neural network (3DCNN). In this manuscript, we discuss the tolerability of within-subject or within-session information leakage, of which the impact is generally considered small but complex and essentially unknown; this requires clarification in future studies.
Collapse
Affiliation(s)
- Sunao Yotsutsuji
- School of Life Science and Technology, Tokyo Institute of Technology, Tokyo, Japan
| | - Miaomei Lei
- Ex-Graduate School of Science and Technology, Tokyo Institute of Technology, Tokyo, Japan
| | - Hiroyuki Akama
- School of Life Science and Technology, Tokyo Institute of Technology, Tokyo, Japan.,Institute of Liberal Arts, Tokyo Institute of Technology, Tokyo, Japan
| |
Collapse
|
26
|
Abstract
Deep learning (DL) approaches to medical image analysis tasks have recently become popular; however, they suffer from a lack of human interpretability critical for both increasing understanding of the methods' operation and enabling clinical translation. This review summarizes currently available methods for performing image model interpretation and critically evaluates published uses of these methods for medical imaging applications. We divide model interpretation in two categories: (1) understanding model structure and function and (2) understanding model output. Understanding model structure and function summarizes ways to inspect the learned features of the model and how those features act on an image. We discuss techniques for reducing the dimensionality of high-dimensional data and cover autoencoders, both of which can also be leveraged for model interpretation. Understanding model output covers attribution-based methods, such as saliency maps and class activation maps, which produce heatmaps describing the importance of different parts of an image to the model prediction. We describe the mathematics behind these methods, give examples of their use in medical imaging, and compare them against one another. We summarize several published toolkits for model interpretation specific to medical imaging applications, cover limitations of current model interpretation methods, provide recommendations for DL practitioners looking to incorporate model interpretation into their task, and offer general discussion on the importance of model interpretation in medical imaging contexts.
Collapse
Affiliation(s)
- Daniel T. Huff
- Department of Medical Physics, University of Wisconsin-Madison, Madison WI
| | - Amy J. Weisman
- Department of Medical Physics, University of Wisconsin-Madison, Madison WI
| | - Robert Jeraj
- Department of Medical Physics, University of Wisconsin-Madison, Madison WI
- Faculty of Mathematics and Physics, University of Ljubljana, Ljubljana, Slovenia
| |
Collapse
|
27
|
Rittman T. Neurological update: neuroimaging in dementia. J Neurol 2020; 267:3429-3435. [PMID: 32638104 PMCID: PMC7578138 DOI: 10.1007/s00415-020-10040-0] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2020] [Revised: 06/26/2020] [Accepted: 06/30/2020] [Indexed: 12/18/2022]
Abstract
Neuroimaging for dementia has made remarkable progress in recent years, shedding light on diagnostic subtypes of dementia, predicting prognosis and monitoring pathology. This review covers some updates in the understanding of dementia using structural imaging, positron emission tomography (PET), structural and functional connectivity, and using big data and artificial intelligence. Progress with neuroimaging methods allows neuropathology to be examined in vivo, providing a suite of biomarkers for understanding neurodegeneration and for application in clinical trials. In addition, we highlight quantitative susceptibility imaging as an exciting new technique that may prove to be a sensitive biomarker for a range of neurodegenerative diseases. There are challenges in translating novel imaging techniques to clinical practice, particularly in developing standard methodologies and overcoming regulatory issues. It is likely that clinicians will need to lead the way if these obstacles are to be overcome. Continued efforts applying neuroimaging to understand mechanisms of neurodegeneration and translating them to clinical practice will complete a revolution in neuroimaging.
Collapse
Affiliation(s)
- Timothy Rittman
- Department of Neurosciences, University of Cambridge, Cambridge, UK.
| |
Collapse
|
28
|
Samek W, Müller KR. Towards Explainable Artificial Intelligence. EXPLAINABLE AI: INTERPRETING, EXPLAINING AND VISUALIZING DEEP LEARNING 2019. [DOI: 10.1007/978-3-030-28954-6_1] [Citation(s) in RCA: 137] [Impact Index Per Article: 27.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/24/2023]
|