51
|
Xiao D, Lian C, Deng H, Kuang T, Liu Q, Ma L, Kim D, Lang Y, Chen X, Gateno J, Shen SG, Xia JJ, Yap PT. Estimating Reference Bony Shape Models for Orthognathic Surgical Planning Using 3D Point-Cloud Deep Learning. IEEE J Biomed Health Inform 2021; 25:2958-2966. [PMID: 33497345 DOI: 10.1109/jbhi.2021.3054494] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Orthognathic surgical outcomes rely heavily on the quality of surgical planning. Automatic estimation of a reference facial bone shape significantly reduces experience-dependent variability and improves planning accuracy and efficiency. We propose an end-to-end deep learning framework to estimate patient-specific reference bony shape models for patients with orthognathic deformities. Specifically, we apply a point-cloud network to learn a vertex-wise deformation field from a patient's deformed bony shape, represented as a point cloud. The estimated deformation field is then used to correct the deformed bony shape to output a patient-specific reference bony surface model. To train our network effectively, we introduce a simulation strategy to synthesize deformed bones from any given normal bone, producing a relatively large and diverse dataset of shapes for training. Our method was evaluated using both synthetic and real patient data. Experimental results show that our framework estimates realistic reference bony shape models for patients with varying deformities. The performance of our method is consistently better than an existing method and several deep point-cloud networks. Our end-to-end estimation framework based on geometric deep learning shows great potential for improving clinical workflows.
Collapse
|
52
|
Xiao D, Deng H, Lian C, Kuang T, Liu Q, Ma L, Lang Y, Chen X, Kim D, Gateno J, Shen SG, Shen D, Yap PT, Xia JJ. Unsupervised learning of reference bony shapes for orthognathic surgical planning with a surface deformation network. Med Phys 2021; 48:7735-7746. [PMID: 34309844 DOI: 10.1002/mp.15126] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/17/2021] [Revised: 07/15/2021] [Accepted: 07/16/2021] [Indexed: 11/11/2022] Open
Abstract
PURPOSE The purpose of this study was to reduce the experience dependence during the orthognathic surgical planning that involves virtually simulating the corrective procedure for jaw deformities. METHODS We introduce a geometric deep learning framework for generating reference facial bone shape models for objective guidance in surgical planning. First, we propose a surface deformation network to warp a patient's deformed bone to a set of normal bones for generating a dictionary of patient-specific normal bony shapes. Subsequently, sparse representation learning is employed to estimate a reference shape model based on the dictionary. RESULTS We evaluated our method on a clinical dataset containing 24 patients, and compared it with a state-of-the-art method that relies on landmark-based sparse representation. Our method yields significantly higher accuracy than the competing method for estimating normal jaws and maintains the midfaces of patients' facial bones as well as the conventional way. CONCLUSIONS Experimental results indicate that our method generates accurate shape models that meet clinical standards.
Collapse
|
53
|
Li G, Liu Y, Zheng Y, Wu Y, Li D, Liang X, Chen Y, Cui Y, Yap PT, Qiu S, Zhang H, Shen D. Multiscale neural modeling of resting-state fMRI reveals executive-limbic malfunction as a core mechanism in major depressive disorder. Neuroimage Clin 2021; 31:102758. [PMID: 34284335 PMCID: PMC8313604 DOI: 10.1016/j.nicl.2021.102758] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2020] [Revised: 06/30/2021] [Accepted: 07/03/2021] [Indexed: 11/15/2022]
Abstract
Major depressive disorder (MDD) represents a grand challenge to human health and society, but the underlying pathophysiological mechanisms remain elusive. Previous neuroimaging studies have suggested that MDD is associated with abnormal interactions and dynamics in two major neural systems including the default mode - salience (DMN-SAL) network and the executive - limbic (EXE-LIM) network, but it is not clear which network plays a central role and which network plays a subordinate role in MDD pathophysiology. To address this question, we refined a newly developed Multiscale Neural Model Inversion (MNMI) framework and applied it to test whether MDD is more affected by impaired circuit interactions in the DMN-SAL network or the EXE-LIM network. The model estimates the directed connection strengths between different neural populations both within and between brain regions based on resting-state fMRI data collected from normal healthy subjects and patients with MDD. Results show that MDD is primarily characterized by abnormal circuit interactions in the EXE-LIM network rather than the DMN-SAL network. Specifically, we observe reduced frontoparietal effective connectivity that potentially contributes to hypoactivity in the dorsolateral prefrontal cortex (dlPFC), and decreased intrinsic inhibition combined with increased excitation from the superior parietal cortex (SPC) that potentially lead to amygdala hyperactivity, together resulting in activation imbalance in the PFC-amygdala circuit that pervades in MDD. Moreover, the model reveals reduced PFC-to-hippocampus excitation but decreased SPC-to-thalamus inhibition in MDD population that potentially lead to hypoactivity in the hippocampus and hyperactivity in the thalamus, consistent with previous experimental data. Overall, our findings provide strong support for the long-standing limbic-cortical dysregulation model in major depression but also offer novel insights into the multiscale pathophysiology of this debilitating disease.
Collapse
|
54
|
Guan H, Liu Y, Yang E, Yap PT, Shen D, Liu M. Multi-site MRI harmonization via attention-guided deep domain adaptation for brain disorder identification. Med Image Anal 2021; 71:102076. [PMID: 33930828 PMCID: PMC8184627 DOI: 10.1016/j.media.2021.102076] [Citation(s) in RCA: 40] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2020] [Revised: 12/21/2020] [Accepted: 04/03/2021] [Indexed: 01/18/2023]
Abstract
Structural magnetic resonance imaging (MRI) has shown great clinical and practical values in computer-aided brain disorder identification. Multi-site MRI data increase sample size and statistical power, but are susceptible to inter-site heterogeneity caused by different scanners, scanning protocols, and subject cohorts. Multi-site MRI harmonization (MMH) helps alleviate the inter-site difference for subsequent analysis. Some MMH methods performed at imaging level or feature extraction level are concise but lack robustness and flexibility to some extent. Even though several machine/deep learning-based methods have been proposed for MMH, some of them require a portion of labeled data in the to-be-analyzed target domain or ignore the potential contributions of different brain regions to the identification of brain disorders. In this work, we propose an attention-guided deep domain adaptation (AD2A) framework for MMH and apply it to automated brain disorder identification with multi-site MRIs. The proposed framework does not need any category label information of target data, and can also automatically identify discriminative regions in whole-brain MR images. Specifically, the proposed AD2A is composed of three key modules: (1) an MRI feature encoding module to extract representations of input MRIs, (2) an attention discovery module to automatically locate discriminative dementia-related regions in each whole-brain MRI scan, and (3) a domain transfer module trained with adversarial learning for knowledge transfer between the source and target domains. Experiments have been performed on 2572 subjects from four benchmark datasets with T1-weighted structural MRIs, with results demonstrating the effectiveness of the proposed method in both tasks of brain disorder identification and disease progression prediction.
Collapse
|
55
|
Xu X, Lian C, Wang S, Zhu T, Chen RC, Wang AZ, Royce TJ, Yap PT, Shen D, Lian J. Asymmetric multi-task attention network for prostate bed segmentation in computed tomography images. Med Image Anal 2021; 72:102116. [PMID: 34217953 DOI: 10.1016/j.media.2021.102116] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2020] [Revised: 05/18/2021] [Accepted: 05/21/2021] [Indexed: 10/21/2022]
Abstract
Post-prostatectomy radiotherapy requires accurate annotation of the prostate bed (PB), i.e., the residual tissue after the operative removal of the prostate gland, to minimize side effects on surrounding organs-at-risk (OARs). However, PB segmentation in computed tomography (CT) images is a challenging task, even for experienced physicians. This is because PB is almost a "virtual" target with non-contrast boundaries and highly variable shapes depending on neighboring OARs. In this work, we propose an asymmetric multi-task attention network (AMTA-Net) for the concurrent segmentation of PB and surrounding OARs. Our AMTA-Net mimics experts in delineating the non-contrast PB by explicitly leveraging its critical dependency on the neighboring OARs (i.e., the bladder and rectum), which are relatively easy to distinguish in CT images. Specifically, we first adopt a U-Net as the backbone network for the low-level (or prerequisite) task of the OAR segmentation. Then, we build an attention sub-network upon the backbone U-Net with a series of cascaded attention modules, which can hierarchically transfer the OAR features and adaptively learn discriminative representations for the high-level (or primary) task of the PB segmentation. We comprehensively evaluate the proposed AMTA-Net on a clinical dataset composed of 186 CT images. According to the experimental results, our AMTA-Net significantly outperforms current clinical state-of-the-arts (i.e., atlas-based segmentation methods), indicating the value of our method in reducing time and labor in the clinical workflow. Our AMTA-Net also presents better performance than the technical state-of-the-arts (i.e., the deep learning-based segmentation methods), especially for the most indistinguishable and clinically critical part of the PB boundaries. Source code is released at https://github.com/superxuang/amta-net.
Collapse
|
56
|
Chang WT, Langella SK, Tang Y, Ahmad S, Zhang H, Yap PT, Giovanello KS, Lin W. Brainwide functional networks associated with anatomically- and functionally-defined hippocampal subfields using ultrahigh-resolution fMRI. Sci Rep 2021; 11:10835. [PMID: 34035413 PMCID: PMC8149395 DOI: 10.1038/s41598-021-90364-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2020] [Accepted: 05/05/2021] [Indexed: 02/04/2023] Open
Abstract
The hippocampus is critical for learning and memory and may be separated into anatomically-defined hippocampal subfields (aHPSFs). Hippocampal functional networks, particularly during resting state, are generally analyzed using aHPSFs as seed regions, with the underlying assumption that the function within a subfield is homogeneous, yet heterogeneous between subfields. However, several prior studies have observed similar resting-state functional connectivity (FC) profiles between aHPSFs. Alternatively, data-driven approaches investigate hippocampal functional organization without a priori assumptions. However, insufficient spatial resolution may result in a number of caveats concerning the reliability of the results. Hence, we developed a functional Magnetic Resonance Imaging (fMRI) sequence on a 7 T MR scanner achieving 0.94 mm isotropic resolution with a TR of 2 s and brain-wide coverage to (1) investigate the functional organization within hippocampus at rest, and (2) compare the brain-wide FC associated with fine-grained aHPSFs and functionally-defined hippocampal subfields (fHPSFs). This study showed that fHPSFs were arranged along the longitudinal axis that were not comparable to the lamellar structures of aHPSFs. For brain-wide FC, the fHPSFs rather than aHPSFs revealed that a number of fHPSFs connected specifically with some of the functional networks. Different functional networks also showed preferential connections with different portions of hippocampal subfields.
Collapse
|
57
|
Chen X, Lian C, Wang L, Deng H, Kuang T, Fung SH, Gateno J, Shen D, Xia JJ, Yap PT. Diverse data augmentation for learning image segmentation with cross-modality annotations. Med Image Anal 2021; 71:102060. [PMID: 33957558 DOI: 10.1016/j.media.2021.102060] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2020] [Revised: 03/20/2021] [Accepted: 03/29/2021] [Indexed: 10/21/2022]
Abstract
The dearth of annotated data is a major hurdle in building reliable image segmentation models. Manual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural features related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualitatively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT.
Collapse
|
58
|
Yao D, Sui J, Wang M, Yang E, Jiaerken Y, Luo N, Yap PT, Liu M, Shen D. A Mutual Multi-Scale Triplet Graph Convolutional Network for Classification of Brain Disorders Using Functional or Structural Connectivity. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1279-1289. [PMID: 33444133 PMCID: PMC8238125 DOI: 10.1109/tmi.2021.3051604] [Citation(s) in RCA: 48] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/20/2023]
Abstract
Brain connectivity alterations associated with mental disorders have been widely reported in both functional MRI (fMRI) and diffusion MRI (dMRI). However, extracting useful information from the vast amount of information afforded by brain networks remains a great challenge. Capturing network topology, graph convolutional networks (GCNs) have demonstrated to be superior in learning network representations tailored for identifying specific brain disorders. Existing graph construction techniques generally rely on a specific brain parcellation to define regions-of-interest (ROIs) to construct networks, often limiting the analysis into a single spatial scale. In addition, most methods focus on the pairwise relationships between the ROIs and ignore high-order associations between subjects. In this letter, we propose a mutual multi-scale triplet graph convolutional network (MMTGCN) to analyze functional and structural connectivity for brain disorder diagnosis. We first employ several templates with different scales of ROI parcellation to construct coarse-to-fine brain connectivity networks for each subject. Then, a triplet GCN (TGCN) module is developed to learn functional/structural representations of brain connectivity networks at each scale, with the triplet relationship among subjects explicitly incorporated into the learning process. Finally, we propose a template mutual learning strategy to train different scale TGCNs collaboratively for disease classification. Experimental results on 1,160 subjects from three datasets with fMRI or dMRI data demonstrate that our MMTGCN outperforms several state-of-the-art methods in identifying three types of brain disorders.
Collapse
|
59
|
Xiao D, Lian C, Wang L, Deng H, Lin HY, Thung KH, Zhu J, Yuan P, Perez L, Gateno J, Shen SG, Yap PT, Xia JJ, Shen D. Estimating Reference Shape Model for Personalized Surgical Reconstruction of Craniomaxillofacial Defects. IEEE Trans Biomed Eng 2021; 68:362-373. [PMID: 32340932 PMCID: PMC8163108 DOI: 10.1109/tbme.2020.2990586] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
OBJECTIVE To estimate a patient-specific reference bone shape model for a patient with craniomaxillofacial (CMF) defects due to facial trauma. METHODS We proposed an automatic facial bone shape estimation framework using pre-traumatic conventional portrait photos and post-traumatic head computed tomography (CT) scans via a 3D face reconstruction and a deformable shape model. Specifically, a three-dimensional (3D) face was first reconstructed from the patient's pre-traumatic portrait photos. Second, a correlation model between the skin and bone surfaces was constructed using a sparse representation based on the CT images of training normal subjects. Third, by feeding the reconstructed 3D face into the correlation model, an initial reference shape model was generated. In addition, we refined the initial estimation by applying non-rigid surface matching between the initially estimated shape and the patient's post-traumatic bone based on the adaptive-focus deformable shape model (AFDSM). Furthermore, a statistical shape model, built from the training normal subjects, was utilized to constrain the deformation process to avoid overfitting. RESULTS AND CONCLUSION The proposed method was evaluated using both synthetic and real patient data. Experimental results show that the patient's abnormal facial bony structure can be recovered using our method, and the estimated reference shape model is considered clinically acceptable by an experienced CMF surgeon. SIGNIFICANCE The proposed method is more suitable to the complex CMF defects for CMF reconstructive surgical planning.
Collapse
|
60
|
Yang E, Liu M, Yao D, Cao B, Lian C, Yap PT, Shen D. Deep Bayesian Hashing With Center Prior for Multi-Modal Neuroimage Retrieval. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:503-513. [PMID: 33048672 PMCID: PMC7909752 DOI: 10.1109/tmi.2020.3030752] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
Multi-modal neuroimage retrieval has greatly facilitated the efficiency and accuracy of decision making in clinical practice by providing physicians with previous cases (with visually similar neuroimages) and corresponding treatment records. However, existing methods for image retrieval usually fail when applied directly to multi-modal neuroimage databases, since neuroimages generally have smaller inter-class variation and larger inter-modal discrepancy compared to natural images. To this end, we propose a deep Bayesian hash learning framework, called CenterHash, which can map multi-modal data into a shared Hamming space and learn discriminative hash codes from imbalanced multi-modal neuroimages. The key idea to tackle the small inter-class variation and large inter-modal discrepancy is to learn a common center representation for similar neuroimages from different modalities and encourage hash codes to be explicitly close to their corresponding center representations. Specifically, we measure the similarity between hash codes and their corresponding center representations and treat it as a center prior in the proposed Bayesian learning framework. A weighted contrastive likelihood loss function is also developed to facilitate hash learning from imbalanced neuroimage pairs. Comprehensive empirical evidence shows that our method can generate effective hash codes and yield state-of-the-art performance in cross-modal retrieval on three multi-modal neuroimage datasets.
Collapse
|
61
|
Liu S, Thung KH, Qu L, Lin W, Shen D, Yap PT. Learning MRI artefact removal with unpaired data. NAT MACH INTELL 2021. [DOI: 10.1038/s42256-020-00270-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
62
|
Huang Y, Ahmad S, Fan J, Shen D, Yap PT. Difficulty-aware hierarchical convolutional neural networks for deformable registration of brain MR images. Med Image Anal 2021; 67:101817. [PMID: 33129152 PMCID: PMC7725910 DOI: 10.1016/j.media.2020.101817] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2019] [Revised: 07/16/2020] [Accepted: 08/31/2020] [Indexed: 10/23/2022]
Abstract
The aim of deformable brain image registration is to align anatomical structures, which can potentially vary with large and complex deformations. Anatomical structures vary in size and shape, requiring the registration algorithm to estimate deformation fields at various degrees of complexity. Here, we present a difficulty-aware model based on an attention mechanism to automatically identify hard-to-register regions, allowing better estimation of large complex deformations. The difficulty-aware model is incorporated into a cascaded neural network consisting of three sub-networks to fully leverage both global and local contextual information for effective registration. The first sub-network is trained at the image level to predict a coarse-scale deformation field, which is then used for initializing the subsequent sub-network. The next two sub-networks progressively optimize at the patch level with different resolutions to predict a fine-scale deformation field. Embedding difficulty-aware learning into the hierarchical neural network allows harder patches to be identified in the deeper sub-networks at higher resolutions for refining the deformation field. Experiments conducted on four public datasets validate that our method achieves promising registration accuracy with better preservation of topology, compared with state-of-the-art registration methods.
Collapse
|
63
|
Chen X, Lian C, Wang L, Deng H, Kuang T, Fung S, Gateno J, Yap PT, Xia JJ, Shen D. Anatomy-Regularized Representation Learning for Cross-Modality Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:274-285. [PMID: 32956048 PMCID: PMC8120796 DOI: 10.1109/tmi.2020.3025133] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
An increasing number of studies are leveraging unsupervised cross-modality synthesis to mitigate the limited label problem in training medical image segmentation models. They typically transfer ground truth annotations from a label-rich imaging modality to a label-lacking imaging modality, under an assumption that different modalities share the same anatomical structure information. However, since these methods commonly use voxel/pixel-wise cycle-consistency to regularize the mappings between modalities, high-level semantic information is not necessarily preserved. In this paper, we propose a novel anatomy-regularized representation learning approach for segmentation-oriented cross-modality image synthesis. It learns a common feature encoding across different modalities to form a shared latent space, where 1) the input and its synthesis present consistent anatomical structure information, and 2) the transformation between two images in one domain is preserved by their syntheses in another domain. We applied our method to the tasks of cross-modality skull segmentation and cardiac substructure segmentation. Experimental results demonstrate the superiority of our method in comparison with state-of-the-art cross-modality medical image segmentation methods.
Collapse
|
64
|
Zhou Y, Chen H, Li Y, Liu Q, Xu X, Wang S, Yap PT, Shen D. Multi-task learning for segmentation and classification of tumors in 3D automated breast ultrasound images. Med Image Anal 2020; 70:101918. [PMID: 33676100 DOI: 10.1016/j.media.2020.101918] [Citation(s) in RCA: 80] [Impact Index Per Article: 20.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2020] [Revised: 11/22/2020] [Accepted: 11/23/2020] [Indexed: 12/12/2022]
Abstract
Tumor classification and segmentation are two important tasks for computer-aided diagnosis (CAD) using 3D automated breast ultrasound (ABUS) images. However, they are challenging due to the significant shape variation of breast tumors and the fuzzy nature of ultrasound images (e.g., low contrast and signal to noise ratio). Considering the correlation between tumor classification and segmentation, we argue that learning these two tasks jointly is able to improve the outcomes of both tasks. In this paper, we propose a novel multi-task learning framework for joint segmentation and classification of tumors in ABUS images. The proposed framework consists of two sub-networks: an encoder-decoder network for segmentation and a light-weight multi-scale network for classification. To account for the fuzzy boundaries of tumors in ABUS images, our framework uses an iterative training strategy to refine feature maps with the help of probability maps obtained from previous iterations. Experimental results based on a clinical dataset of 170 3D ABUS volumes collected from 107 patients indicate that the proposed multi-task framework improves tumor segmentation and classification over the single-task learning counterparts.
Collapse
|
65
|
Huynh KM, Xu T, Wu Y, Wang X, Chen G, Wu H, Thung KH, Lin W, Shen D, Yap PT. Probing Tissue Microarchitecture of the Baby Brain via Spherical Mean Spectrum Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3607-3618. [PMID: 32746109 PMCID: PMC7688284 DOI: 10.1109/tmi.2020.3001175] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
During the first years of life, the human brain undergoes dynamic spatially-heterogeneous changes, invo- lving differentiation of neuronal types, dendritic arbori- zation, axonal ingrowth, outgrowth and retraction, synaptogenesis, and myelination. To better quantify these changes, this article presents a method for probing tissue microarchitecture by characterizing water diffusion in a spectrum of length scales, factoring out the effects of intra-voxel orientation heterogeneity. Our method is based on the spherical means of the diffusion signal, computed over gradient directions for a set of diffusion weightings (i.e., b -values). We decompose the spherical mean profile at each voxel into a spherical mean spectrum (SMS), which essentially encodes the fractions of spin packets undergoing fine- to coarse-scale diffusion proce- sses, characterizing restricted and hindered diffusion stemming respectively from intra- and extra-cellular water compartments. From the SMS, multiple orientation distribution invariant indices can be computed, allowing for example the quantification of neurite density, microscopic fractional anisotropy ( μ FA), per-axon axial/radial diffusivity, and free/restricted isotropic diffusivity. We show that these indices can be computed for the developing brain for greater sensitivity and specificity to development related changes in tissue microstructure. Also, we demonstrate that our method, called spherical mean spectrum imaging (SMSI), is fast, accurate, and can overcome the biases associated with other state-of-the-art microstructure models.
Collapse
|
66
|
Liu S, Thung KH, Lin W, Shen D, Yap PT. Hierarchical Nonlocal Residual Networks for Image Quality Assessment of Pediatric Diffusion MRI With Limited and Noisy Annotations. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3691-3702. [PMID: 32746115 PMCID: PMC7606371 DOI: 10.1109/tmi.2020.3002708] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Fast and automated image quality assessment (IQA) of diffusion MR images is crucial for making timely decisions for rescans. However, learning a model for this task is challenging as the number of annotated data is limited and the annotation labels might not always be correct. As a remedy, we will introduce in this paper an automatic image quality assessment (IQA) method based on hierarchical non-local residual networks for pediatric diffusion MR images. Our IQA is performed in three sequential stages, i.e., 1) slice-wise IQA, where a nonlocal residual network is first pre-trained to annotate each slice with an initial quality rating (i.e., pass/questionable/fail), which is subsequently refined via iterative semi-supervised learning and slice self-training; 2) volume-wise IQA, which agglomerates the features extracted from the slices of a volume, and uses a nonlocal network to annotate the quality rating for each volume via iterative volume self-training; and 3) subject-wise IQA, which ensembles the volumetric IQA results to determine the overall image quality pertaining to a subject. Experimental results demonstrate that our method, trained using only samples of modest size, exhibits great generalizability, and is capable of conducting rapid hierarchical IQA with near-perfect accuracy.
Collapse
|
67
|
Liu F, Feng J, Chen G, Shen D, Yap PT. Gaussianization of Diffusion MRI Data Using Spatially Adaptive Filtering. Med Image Anal 2020; 68:101828. [PMID: 33338870 DOI: 10.1016/j.media.2020.101828] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2019] [Revised: 06/28/2020] [Accepted: 08/07/2020] [Indexed: 11/27/2022]
Abstract
Diffusion MRI magnitude data, typically Rician or noncentral χ distributed, is affected by the noise floor, which falsely elevates signal, reduces image contrast, and biases estimation of diffusion parameters. Noise floor can be avoided by extracting real-valued Gaussian-distributed data from complex diffusion-weighted images via phase correction, which is performed by rotating each complex diffusion-weighted image based on its phase so that the actual image content resides in the real part. The imaginary part can then be discarded, leaving only the real part to form a Gaussian-noise image that is not confounded by the noise floor. The effectiveness of phase correction depends on the estimation of the background phase associated with factors such as brain motion, cardiac pulsation, perfusion, and respiration. Most existing smoothing techniques, applied to the real and imaginary images for phase estimation, assume spatially-stationary noise. This assumption does not necessarily hold in real data. In this paper, we introduce an adaptive filtering approach, called multi-kernel filter (MKF), for image smoothing catering to spatially-varying noise. Inspired by the mechanisms of human vision, MKF employs a bilateral filter with spatially-varying kernels. Extensive experiments demonstrate that MKF significantly improves spatial adaptivity and outperforms various state-of-the-art filters in signal Gaussianization.
Collapse
|
68
|
Huynh KM, Wu Y, Thung KH, Ahmad S, Taylor HP, Shen D, Yap PT. Characterizing Intra-soma Diffusion with Spherical Mean Spectrum Imaging. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12267:354-363. [PMID: 34223563 PMCID: PMC8248904 DOI: 10.1007/978-3-030-59728-3_35] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/03/2024]
Abstract
Most brain microstructure models are dedicated to the quantification of white matter microstructure, using for example sticks, cylinders, and zeppelins to model intra- and extra-axonal environments. Gray matter presents unique micro-architecture with cell bodies (somas) exhibiting diffusion characteristics that differ from axons in white matter. In this paper, we introduce a method to quantify soma microstructure, giving measures such as volume fraction, diffusivity, and kurtosis. Our method captures a spectrum of diffusion patterns and scales and does not rely on restrictive model assumptions. We show that our method yields unique and meaningful contrasts that are in agreement with histological data. We demonstrate its application in the mapping of the distinct spatial patterns of soma density in the cortex.
Collapse
|
69
|
Yao D, Sui J, Yang E, Yap PT, Shen D, Liu M. Temporal-Adaptive Graph Convolutional Network for Automated Identification of Major Depressive Disorder Using Resting-State fMRI. MACHINE LEARNING IN MEDICAL IMAGING. MLMI (WORKSHOP) 2020; 12436:1-10. [PMID: 36383497 PMCID: PMC9645786 DOI: 10.1007/978-3-030-59861-7_1] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Extensive studies focus on analyzing human brain functional connectivity from a network perspective, in which each network contains complex graph structures. Based on resting-state functional MRI (rs-fMRI) data, graph convolutional networks (GCNs) enable comprehensive mapping of brain functional connectivity (FC) patterns to depict brain activities. However, existing studies usually characterize static properties of the FC patterns, ignoring the time-varying dynamic information. In addition, previous GCN methods generally use fixed group-level (e.g., patients or controls) representation of FC networks, and thus, cannot capture subject-level FC specificity. To this end, we propose a Temporal-Adaptive GCN (TAGCN) framework that can not only take advantage of both spatial and temporal information using resting-state FC patterns and time-series but also explicitly characterize subject-level specificity of FC patterns. Specifically, we first segment each ROI-based time-series into multiple overlapping windows, then employ an adaptive GCN to mine topological information. We further model the temporal patterns for each ROI along time to learn the periodic brain status changes. Experimental results on 533 major depressive disorder (MDD) and health control (HC) subjects demonstrate that the proposed TAGCN outperforms several state-of-the-art methods in MDD vs. HC classification, and also can be used to capture dynamic FC alterations and learn valid graph representations.
Collapse
|
70
|
Wu Y, Hong Y, Ahmad S, Lin W, Shen D, Yap PT. Tract Dictionary Learning for Fast and Robust Recognition of Fiber Bundles. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12267:251-259. [PMID: 34195699 PMCID: PMC8238464 DOI: 10.1007/978-3-030-59728-3_25] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/02/2022]
Abstract
In this paper, we propose an efficient framework for parcellation of white matter tractograms using discriminative dictionary learning. Key to our framework is the learning of a compact dictionary for each fiber bundle so that the streamlines within the bundle can be sufficiently represented. Dictionaries for multiple bundles are combined for whole-brain tractogram representation. These dictionaries are learned jointly to encourage inter-bundle incoherence for discriminative power. The proposed method allows tractograms to be assigned to more than one bundle, catering to scenarios where tractograms cannot be clearly separated. Experiments on a bundle-labeled HCP dataset and an infant dataset highlight the ability of our framework in grouping streamlines into anatomically plausible bundles.
Collapse
|
71
|
Lang Y, Lian C, Xiao D, Deng H, Yuan P, Gateno J, Shen SGF, Alfi DM, Yap PT, Xia JJ, Shen D. Automatic Localization of Landmarks in Craniomaxillofacial CBCT Images Using a Local Attention-Based Graph Convolution Network. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12264:817-826. [PMID: 34927175 PMCID: PMC8675277 DOI: 10.1007/978-3-030-59719-1_79] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
Landmark localization is an important step in quantifying craniomaxillofacial (CMF) deformities and designing treatment plans of reconstructive surgery. However, due to the severity of deformities and defects (partially missing anatomy), it is difficult to automatically and accurately localize a large set of landmarks simultaneously. In this work, we propose two cascaded networks for digitizing 60 anatomical CMF landmarks in cone-beam computed tomography (CBCT) images. The first network is a U-Net that outputs heatmaps for landmark locations and landmark features extracted with a local attention mechanism. The second network is a graph convolution network that takes the features extracted by the first network as input and determines whether each landmark exists via binary classification. We evaluated our approach on 50 sets of CBCT scans of patients with CMF deformities and compared them with state-of-the-art methods. The results indicate that our approach can achieve an average detection error of 1.47mm with a false positive rate of 19%, outperforming related methods.
Collapse
|
72
|
Wei D, Ahmad S, Huo J, Huang P, Yap PT, Xue Z, Sun J, Li W, Shen D, Wang Q. SLIR: Synthesis, localization, inpainting, and registration for image-guided thermal ablation of liver tumors. Med Image Anal 2020; 65:101763. [DOI: 10.1016/j.media.2020.101763] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2019] [Revised: 04/10/2020] [Accepted: 06/19/2020] [Indexed: 12/31/2022]
|
73
|
Lian C, Wang F, Deng HH, Wang L, Xiao D, Kuang T, Lin HY, Gateno J, Shen SGF, Yap PT, Xia JJ, Shen D. Multi-task Dynamic Transformer Network for Concurrent Bone Segmentation and Large-Scale Landmark Localization with Dental CBCT. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12264:807-816. [PMID: 34935006 PMCID: PMC8687703 DOI: 10.1007/978-3-030-59719-1_78] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
Abstract
Accurate bone segmentation and anatomical landmark localization are essential tasks in computer-aided surgical simulation for patients with craniomaxillofacial (CMF) deformities. To leverage the complementarity between the two tasks, we propose an efficient end-to-end deep network, i.e., multi-task dynamic transformer network (DTNet), to concurrently segment CMF bones and localize large-scale landmarks in one-pass from large volumes of cone-beam computed tomography (CBCT) data. Our DTNet was evaluated quantitatively using CBCTs of patients with CMF deformities. The results demonstrated that our method outperforms the other state-of-the-art methods in both tasks of the bony segmentation and the landmark digitization. Our DTNet features three main technical contributions. First, a collaborative two-branch architecture is designed to efficiently capture both fine-grained image details and complete global context for high-resolution volume-to-volume prediction. Second, leveraging anatomical dependencies between landmarks, regionalized dynamic learners (RDLs) are designed in the concept of "learns to learn" to jointly regress large-scale 3D heatmaps of all landmarks under limited computational costs. Third, adaptive transformer modules (ATMs) are designed for the flexible learning of task-specific feature embedding from common feature bases.
Collapse
|
74
|
Cheng F, Chen Y, Zong X, Lin W, Shen D, Yap PT. Acceleration of High-Resolution 3D MR Fingerprinting via a Graph Convolutional Network. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12262:158-166. [PMID: 38504822 PMCID: PMC10950303 DOI: 10.1007/978-3-030-59713-9_16] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/21/2024]
Abstract
Magnetic resonance fingerprinting (MRF) is a novel imaging framework for fast and simultaneous quantification of multiple tissue properties. Recently, 3D MRF methods have been developed, but the acquisition speed needs to be improved before they can be adopted for clinical use. The purpose of this study is to develop a novel deep learning approach to accelerate 3D MRF acquisition along the slice-encoding direction in k-space. We introduce a graph-based convolutional neural network that caters to non-Cartesian spiral trajectories commonly used for MRF acquisition. We improve tissue quantification accuracy compared with the state of the art. Our method enables fast 3D MRF with high spatial resolution, allowing whole-brain coverage within 5min, making MRF more feasible in clinical settings.
Collapse
|
75
|
Chen G, Hong Y, Zhang Y, Kim J, Huynh KM, Ma J, Lin W, Shen D, Yap PT. Estimating Tissue Microstructure with Undersampled Diffusion Data via Graph Convolutional Neural Networks. MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION : MICCAI ... INTERNATIONAL CONFERENCE ON MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION 2020; 12267:280-290. [PMID: 34308440 PMCID: PMC8294782 DOI: 10.1007/978-3-030-59728-3_28] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/03/2022]
Abstract
Advanced diffusion models for tissue microstructure are widely employed to study brain disorders. However, these models usually require diffusion MRI (DMRI) data with densely sampled q-space, which is prohibitive in clinical settings. This problem can be resolved by using deep learning techniques, which learn the mapping between sparsely sampled q-space data and the high-quality diffusion microstructural indices estimated from densely sampled data. However, most existing methods simply view the input DMRI data as a vector without considering data structure in the q-space. In this paper, we propose to overcome this limitation by representing DMRI data using graphs and utilizing graph convolutional neural networks to estimate tissue microstructure. Our method makes full use of the q-space angular neighboring information to improve estimation accuracy. Experimental results based on data from the Baby Connectome Project demonstrate that our method outperforms state-of-the-art methods both qualitatively and quantitatively.
Collapse
|