1
|
Electrophysiological brain imaging based on simulation-driven deep learning in the context of epilepsy. Neuroimage 2024; 285:120490. [PMID: 38103624 DOI: 10.1016/j.neuroimage.2023.120490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Revised: 11/30/2023] [Accepted: 12/06/2023] [Indexed: 12/19/2023] Open
Abstract
Identifying the location, the spatial extent and the electrical activity of distributed brain sources in the context of epilepsy through ElectroEncephaloGraphy (EEG) recordings is a challenging task because of the highly ill-posed nature of the underlying Electrophysiological Source Imaging (ESI) problem. To guarantee a unique solution, most existing ESI methods pay more attention to solve this inverse problem by imposing physiological constraints. This paper proposes an efficient ESI approach based on simulation-driven deep learning. Epileptic High-resolution 256-channels scalp EEG (Hr-EEG) signals are simulated in a realistic manner to train the proposed patient-specific model. More particularly, a computational neural mass model developed in our team is used to generate the temporal dynamics of the activity of each dipole while the forward problem is solved using a patient-specific three-shell realistic head model and the boundary element method. A Temporal Convolutional Network (TCN) is considered in the proposed model to capture local spatial patterns. To enable the model to observe the EEG signals from different scale levels, the multi-scale strategy is leveraged to capture the overall features and fine-grain features by adjusting the convolutional kernel size. Then, the Long Short-Term Memory (LSTM) is used to extract temporal dependencies among the computed spatial features. The performance of the proposed method is evaluated through three different scenarios of realistic synthetic interictal Hr-EEG data as well as on real interictal Hr-EEG data acquired in three patients with drug-resistant partial epilepsy, during their presurgical evaluation. A performance comparison study is also conducted with two other deep learning-based methods and four classical ESI techniques. The proposed model achieved a Dipole Localization Error (DLE) of 1.39 and Normalized Hamming Distance (NHD) of 0.28 in the case of one patch with SNR of 10 dB. In the case of two uncorrelated patches with an SNR of 10 dB, obtained DLE and NHD were respectively 1.50 and 0.28. Even in the more challenging scenario of two correlated patches with an SNR of 10 dB, the proposed approach still achieved a DLE of 3.74 and an NHD of 0.43. The results obtained on simulated data demonstrate that the proposed method outperforms the existing methods for different signal-to-noise and source configurations. The good behavior of the proposed method is also confirmed on real interictal EEG data. The robustness with respect to noise makes it a promising and alternative tool to localize epileptic brain areas and to reconstruct their electrical activities from EEG signals.
Collapse
|
2
|
Deep learning-based combination of [18F]-FDG PET and CT images for producing pulmonary perfusion image. Med Phys 2023; 50:7779-7790. [PMID: 37387645 DOI: 10.1002/mp.16566] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2023] [Accepted: 06/07/2023] [Indexed: 07/01/2023] Open
Abstract
BACKGROUND The main application of [18F] FDG-PET (18 FDG-PET) and CT images in oncology is tumor identification and quantification. Combining PET and CT images to mine pulmonary perfusion information for functional lung avoidance radiation therapy (FLART) is desirable but remains challenging. PURPOSE To develop a deep-learning-based (DL) method to combine 18 FDG-PET and CT images for producing pulmonary perfusion images (PPI). METHODS Pulmonary technetium-99 m-labeled macroaggregated albumin SPECT (PPISPECT ), 18 FDG-PET, and CT images obtained from 53 patients were enrolled. CT and PPISPECT images were rigidly registered, and registration displacement was subsequently used to align 18 FDG-PET and PPISPECT images. The left/right lung was separated and rigidly registered again to improve the registration accuracy. A DL model based on 3D Unet architecture was constructed to directly combine multi-modality 18 FDG-PET and CT images for producing PPI (PPIDLM ). 3D Unet architecture was used as the basic architecture, and the input was expanded from a single-channel to a dual-channel to combine multi-modality images. For comparative evaluation, 18 FDG-PET images were also used alone to generate PPIDLPET . Sixty-seven samples were randomly selected for training and cross-validation, and 36 were used for testing. The Spearman correlation coefficient (rs ) and multi-scale structural similarity index measure (MS-SSIM) between PPIDLM /PPIDLPET and PPISPECT were computed to assess the statistical and perceptual image similarities. The Dice similarity coefficient (DSC) was calculated to determine the similarity between high-/low- functional lung (HFL/LFL) volumes. RESULTS The voxel-wise rs and MS-SSIM of PPIDLM /PPIDLPET were 0.78 ± 0.04/0.57 ± 0.03, 0.93 ± 0.01/0.89 ± 0.01 for cross-validation and 0.78 ± 0.11/0.55 ± 0.18, 0.93 ± 0.03/0.90 ± 0.04 for testing. PPIDLM /PPIDLPET achieved averaged DSC values of 0.78 ± 0.03/0.64 ± 0.02 for HFL and 0.83 ± 0.01/0.72 ± 0.03 for LFL in the training dataset and 0.77 ± 0.11/0.64 ± 0.12, 0.82 ± 0.05/0.72 ± 0.06 in the testing dataset. PPIDLM yielded a stronger correlation and higher MS-SSIM with PPISPECT than PPIDLPET (p < 0.001). CONCLUSIONS The DL-based method integrates lung metabolic and anatomy information for producing PPI and significantly improved the accuracy over methods based on metabolic information alone. The generated PPIDLM can be applied for pulmonary perfusion volume segmentation, which is potentially beneficial for FLART treatment plan optimization.
Collapse
|
3
|
Estimation of intravoxel incoherent motion (IVIM) parameters in vertebral bone marrow: a comparative study of five algorithms. MAGMA (NEW YORK, N.Y.) 2023; 36:837-847. [PMID: 36715885 DOI: 10.1007/s10334-023-01064-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/04/2022] [Revised: 01/15/2023] [Accepted: 01/16/2023] [Indexed: 01/31/2023]
Abstract
OBJECTIVES To access the performances of different algorithms for quantification of Intravoxel incoherent motion (IVIM) parameters D, f, [Formula: see text] in Vertebral Bone Marrow (VBM). MATERIALS AND METHODS Five algorithms were studied: four deterministic algorithms (the One-Step and three segmented methods: Two-Step, Three-Step, and Fixed-[Formula: see text] algorithm) based on the least-squares (LSQ) method and a Bayesian probabilistic algorithm. Numerical simulations and quantification of IVIM parameters D, f, [Formula: see text] in vivo in vertebral bone marrow, were done on six healthy volunteers. The One-way repeated-measures analysis of variance (ANOVA) followed by Bonferroni's multiple comparison test (p value = 0.05) was applied. RESULTS In numerical simulations, the Bayesian algorithm provided the best estimation of D, f, [Formula: see text] compared to the deterministic algorithms. In vivo VBM-IVIM, the values of D and f estimated by the Bayesian algorithm were close to those of the One-Step method, in contrast to the three segmented methods. DISCUSSION The comparison of the five algorithms indicates that the Bayesian algorithm provides the best estimation of VBM-IVIM parameters, in both numerical simulations and in vivo data.
Collapse
|
4
|
Online prediction for respiratory movement compensation: a patient-specific gating control for MRI-guided radiotherapy. Radiat Oncol 2023; 18:149. [PMID: 37697360 PMCID: PMC10496354 DOI: 10.1186/s13014-023-02341-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2023] [Accepted: 08/31/2023] [Indexed: 09/13/2023] Open
Abstract
BACKGROUND This study aims to validate the effectiveness of linear regression for motion prediction of internal organs or tumors on 2D cine-MR and to present an online gating signal prediction scheme that can improve the accuracy of MR-guided radiotherapy for liver and lung cancer. MATERIALS AND METHODS We collected 2D cine-MR sequences of 21 liver cancer patients and 10 lung cancer patients to develop a binary gating signal prediction algorithm that forecasts the crossing-time of tumor motion traces relative to the target threshold. Both 0.4 s and 0.6 s prediction windows were tested using three linear predictors and three recurrent neural networks (RNNs), given the system delay of 0.5 s. Furthermore, an adaptive linear regression model was evaluated using only the first 30 s as the burn-in period, during which the model parameters were adapted during the online prediction process. The accuracy of the predicted traces was measured using amplitude metrics (MAE, RMSE, and R2), and in addition, we proposed three temporal metrics, namely crossing error, gating error, and gating accuracy, which are more relevant to the nature of the gating signals. RESULTS In both 0.6 s and 0.4 s prediction cases, linear regression outperformed other methods, demonstrating significantly smaller amplitude errors compared to the RNNs (P < 0.05). The proposed algorithm with adaptive linear regression had the best performance with an average gating accuracy of 98.3% and 98.0%, a gating error of 44 ms and 45 ms, for liver cancer and lung cancer patients, respectively. CONCLUSION A functional online gating control scheme was developed with an adaptive linear regression that is both more cost-efficient and accurate than sophisticated RNN based methods in all studied metrics.
Collapse
|
5
|
Multi-Scale Feature Fusion Network for Low-Dose CT Denoising. J Digit Imaging 2023; 36:1808-1825. [PMID: 36914854 PMCID: PMC10406773 DOI: 10.1007/s10278-023-00805-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Revised: 03/01/2023] [Accepted: 03/03/2023] [Indexed: 03/16/2023] Open
Abstract
Computed tomography (CT) is an imaging technique extensively used in medical treatment, but too much radiation dose in a CT scan will cause harm to the human body. Decreasing the dose of radiation will result in increased noise and artifacts in the reconstructed image, blurring the internal tissue and edge details. To get high-quality CT images, we present a multi-scale feature fusion network (MSFLNet) for low-dose CT (LDCT) denoising. In our MSFLNet, we combined multiple feature extraction modules, effective noise reduction modules, and fusion modules constructed using the attention mechanism to construct a horizontally connected multi-scale structure as the overall architecture of the network, which is used to construct different levels of feature maps at all scales. We innovatively define a composite loss function composed of pixel-level loss based on MS-SSIM-L1 and edge-based edge loss for LDCT denoising. In short, our approach learns a rich set of features that combine contextual information from multiple scales while maintaining the spatial details of denoised CT images. Our laboratory results indicate that compared with the existing methods, the peak signal-to-noise ratio (PSNR) value of CT images of the AAPM dataset processed by the new model is 33.6490, and the structural similarity (SSIM) value is 0.9174, which also achieves good results on the Piglet dataset with different doses. The results also show that the method removes noise and artifacts while effectively preserving CT images' architecture and grain information.
Collapse
|
6
|
TSUnet-CC: Temporal Spectrogram Unet embedding Cross Channel-wise attention mechanism for MDD identification. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2023; 2023:1-4. [PMID: 38083642 DOI: 10.1109/embc40787.2023.10340299] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2023]
Abstract
Automatic detection of major depressive disorder (MDD) with multiple-channel electroencephalography (EEG) signals is of great significance for treatment of the mental diseases. In a U-net network, clear EEG signals are fed to obtain temporal feature tensor through encoder and decoder networks with several convolution operations. Moreover, the clear EEG signals can be converted into multi-scale spectrogram to obtain the rich saliency information and then the spectrogram feature tensor can be extracted by another symmetrical U-net. The temporal and spectrogram feature tensors can provide more comprehensive information, but may also contain redundant information, which may affect the detection of MDD. To deal with such issue, this paper proposed a novel Temporal Spectrogram Unet (TSUnet-CC), which embeds the cross channel-wise attention mechanism for multiple-channel EEGbased MDD identification. We make three novel contributions: 1) multi-scale saliency-encoded spectrogram using Fourierbased approach to capture rich saliency information under different scales, 2) TSUnet network using a symmetrical twostream U-net architecture that learns multiple temporal and spectrogram feature tensors in time and frequency domains, and 3) cross channel-wise block enabling the larger weights of key feature channels that contain MDD information. The leaveone-subject-out experiments show that our proposed TSUnetCC gains high performance with a classification accuracy up to 98.55% and 99.22% in eyes closed and eyes open datasets, which outperformed some state-of-the-art methods and revealed its clinical potential.
Collapse
|
7
|
BPAT-UNet: Boundary preserving assembled transformer UNet for ultrasound thyroid nodule segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 238:107614. [PMID: 37244233 DOI: 10.1016/j.cmpb.2023.107614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Revised: 04/23/2023] [Accepted: 05/15/2023] [Indexed: 05/29/2023]
Abstract
BACKGROUND AND OBJECTIVE Accurate and efficient segmentation of thyroid nodules on ultrasound images is critical for computer-aided nodule diagnosis and treatment. For ultrasound images, Convolutional neural networks (CNNs) and Transformers, which are widely used in natural images, cannot obtain satisfactory segmentation results, because they either cannot obtain precise boundaries or segment small objects. METHODS To address these issues, we propose a novel Boundary-preserving assembly Transformer UNet (BPAT-UNet) for ultrasound thyroid nodule segmentation. In the proposed network, a Boundary point supervision module (BPSM), which adopts two novel self-attention pooling approaches, is designed to enhance boundary features and generate ideal boundary points through a novel method. Meanwhile, an Adaptive multi-scale feature fusion module (AMFFM) is constructed to fuse features and channel information at different scales. Finally, to fully integrate the characteristics of high-frequency local and low-frequency global, the Assembled transformer module (ATM) is placed at the bottleneck of the network. The correlation between deformable features and features-among computation is characterized by introducing them into the above two modules of AMFFM and ATM. As the design goal and eventually demonstrated, BPSM and ATM promote the proposed BPAT-UNet to further constrain boundaries, whereas AMFFM assists to detect small objects. RESULTS Compared to other classical segmentation networks, the proposed BPAT-UNet displays superior segmentation performance in visualization results and evaluation metrics. Significant improvement of segmentation accuracy was shown on the public thyroid dataset of TN3k with Dice similarity coefficient (DSC) of 81.64% and 95th percentage of the asymmetric Hausdorff distance (HD95) of 14.06, whereas those on our private dataset were with DSC of 85.63% and HD95 of 14.53, respectively. CONCLUSIONS This paper presents a method for thyroid ultrasound image segmentation, which achieves high accuracy and meets the clinical requirements. Code is available at https://github.com/ccjcv/BPAT-UNet.
Collapse
|
8
|
Dual-domain fusion deep convolutional neural network for low-dose CT denoising. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2023:XST230020. [PMID: 37212059 DOI: 10.3233/xst-230020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
Abstract
BACKGROUND In view of the underlying health risks posed by X-ray radiation, the main goal of the present research is to achieve high-quality CT images at the same time as reducing x-ray radiation. In recent years, convolutional neural network (CNN) has shown excellent performance in removing low-dose CT noise. However, previous work mainly focused on deepening and feature extraction work on CNN without considering fusion of features from frequency domain and image domain. OBJECTIVE To address this issue, we propose to develop and test a new LDCT image denoising method based on a dual-domain fusion deep convolutional neural network (DFCNN). METHODS This method deals with two domains, namely, the DCT domain and the image domain. In the DCT domain, we design a new residual CBAM network to enhance the internal and external relations of different channels while reducing noise to promote richer image structure information. For the image domain, we propose a top-down multi-scale codec network as a denoising network to obtain more acceptable edges and textures while obtaining multi-scale information. Then, the feature images of the two domains are fused by a combination network. RESULTS The proposed method was validated on the Mayo dataset and the Piglet dataset. The denoising algorithm is optimal in both subjective and objective evaluation indexes as compared to other state-of-the-art methods reported in previous studies. CONCLUSIONS The study results demonstrate that by applying the new fusion model denoising, denoising results in both image domain and DCT domain are better than other models developed using features extracted in the single image domain.
Collapse
|
9
|
Evaluating the accuracy of automated cephalometric analysis based on artificial intelligence. BMC Oral Health 2023; 23:191. [PMID: 37005593 DOI: 10.1186/s12903-023-02881-8.pmid:37005593;pmcid:pmc10067288] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Accepted: 03/14/2023] [Indexed: 05/26/2023] Open
Abstract
BACKGROUND The purpose of this study was to evaluate the accuracy of automatic cephalometric landmark localization and measurements using cephalometric analysis via artificial intelligence (AI) compared with computer-assisted manual analysis. METHODS Reconstructed lateral cephalograms (RLCs) from cone-beam computed tomography (CBCT) in 85 patients were selected. Computer-assisted manual analysis (Dolphin Imaging 11.9) and AI automatic analysis (Planmeca Romexis 6.2) were used to locate 19 landmarks and obtain 23 measurements. Mean radial error (MRE) and successful detection rate (SDR) values were calculated to assess the accuracy of automatic landmark digitization. Paired t tests and Bland‒Altman plots were used to compare the differences and consistencies in cephalometric measurements between manual and automatic analysis programs. RESULTS The MRE for 19 cephalometric landmarks was 2.07 ± 1.35 mm with the automatic program. The average SDR within 1 mm, 2 mm, 2.5 mm, 3 and 4 mm were 18.82%, 58.58%, 71.70%, 82.04% and 91.39%, respectively. Soft tissue landmarks (1.54 ± 0.85 mm) had the most consistency, while dental landmarks (2.37 ± 1.55 mm) had the most variation. In total, 15 out of 23 measurements were within the clinically acceptable level of accuracy, 2 mm or 2°. The rates of consistency within the 95% limits of agreement were all above 90% for all measurement parameters. CONCLUSION Automatic analysis software collects cephalometric measurements almost effectively enough to be acceptable in clinical work. Nevertheless, automatic cephalometry is not capable of completely replacing manual tracing. Additional manual supervision and adjustment for automatic programs can increase accuracy and efficiency.
Collapse
|
10
|
Evaluating the accuracy of automated cephalometric analysis based on artificial intelligence. BMC Oral Health 2023; 23:191. [PMID: 37005593 PMCID: PMC10067288 DOI: 10.1186/s12903-023-02881-8] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Accepted: 03/14/2023] [Indexed: 04/04/2023] Open
Abstract
BACKGROUND The purpose of this study was to evaluate the accuracy of automatic cephalometric landmark localization and measurements using cephalometric analysis via artificial intelligence (AI) compared with computer-assisted manual analysis. METHODS Reconstructed lateral cephalograms (RLCs) from cone-beam computed tomography (CBCT) in 85 patients were selected. Computer-assisted manual analysis (Dolphin Imaging 11.9) and AI automatic analysis (Planmeca Romexis 6.2) were used to locate 19 landmarks and obtain 23 measurements. Mean radial error (MRE) and successful detection rate (SDR) values were calculated to assess the accuracy of automatic landmark digitization. Paired t tests and Bland‒Altman plots were used to compare the differences and consistencies in cephalometric measurements between manual and automatic analysis programs. RESULTS The MRE for 19 cephalometric landmarks was 2.07 ± 1.35 mm with the automatic program. The average SDR within 1 mm, 2 mm, 2.5 mm, 3 and 4 mm were 18.82%, 58.58%, 71.70%, 82.04% and 91.39%, respectively. Soft tissue landmarks (1.54 ± 0.85 mm) had the most consistency, while dental landmarks (2.37 ± 1.55 mm) had the most variation. In total, 15 out of 23 measurements were within the clinically acceptable level of accuracy, 2 mm or 2°. The rates of consistency within the 95% limits of agreement were all above 90% for all measurement parameters. CONCLUSION Automatic analysis software collects cephalometric measurements almost effectively enough to be acceptable in clinical work. Nevertheless, automatic cephalometry is not capable of completely replacing manual tracing. Additional manual supervision and adjustment for automatic programs can increase accuracy and efficiency.
Collapse
|
11
|
Deep-learning approach to automate the segmentation of aorta in non-contrast CTs. J Med Imaging (Bellingham) 2023; 10:024001. [PMID: 36875637 PMCID: PMC9981198 DOI: 10.1117/1.jmi.10.2.024001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2022] [Accepted: 02/13/2023] [Indexed: 03/05/2023] Open
Abstract
Purpose Segmentation of vascular structures in preoperative computed tomography (CT) is a preliminary step for computer-assisted endovascular navigation. It is a challenging issue when contrast medium enhancement is reduced or impossible, as in the case of endovascular abdominal aneurysm repair for patients with severe renal impairment. In non-contrast-enhanced CTs, the segmentation tasks are currently hampered by the problems of low contrast, similar topological form, and size imbalance. To tackle these problems, we propose a novel fully automatic approach based on convolutional neural network. Approach The proposed method is implemented by fusing the features from different dimensions by three kinds of mechanisms, i.e., channel concatenation, dense connection, and spatial interpolation. The fusion mechanisms are regarded as the enhancement of features in non-contrast CTs where the boundary of aorta is ambiguous. Results All of the networks are validated by three-fold cross-validation on our dataset of non-contrast CTs, which contains 5749 slices in total from 30 individual patients. Our methods achieve a Dice score of 88.7% as the overall performance, which is better than the results reported in the related works. Conclusions The analysis indicates that our methods yield a competitive performance by overcoming the above-mentioned problems in most general cases. Further, experiments on our non-contrast CTs demonstrate the superiority of the proposed methods, especially in low-contrast, similar-shaped, and extreme-sized cases.
Collapse
|
12
|
A total variation prior unrolling approach for computed tomography reconstruction. Med Phys 2023; 50:2816-2834. [PMID: 36791315 DOI: 10.1002/mp.16307] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2022] [Revised: 01/09/2023] [Accepted: 02/06/2023] [Indexed: 02/17/2023] Open
Abstract
BACKGROUND With the rapid development of deep learning technology, deep neural networks can effectively enhance the performance of computed tomography (CT) reconstructions. One kind of commonly used method to construct CT reconstruction networks is to unroll the conventional iterative reconstruction (IR) methods to convolutional neural networks (CNNs). However, most unrolling methods primarily unroll the fidelity term of IR methods to CNNs, without unrolling the prior terms. The prior terms are always directly replaced by neural networks. PURPOSE In conventional IR methods, the prior terms play a vital role in improving the visual quality of reconstructed images. Unrolling the hand-crafted prior terms to CNNs may provide a more specialized unrolling approach to further improve the performance of CT reconstruction. In this work, a primal-dual network (PD-Net) was proposed by unrolling both the data fidelity term and the total variation (TV) prior term, which effectively preserves the image edges and textures in the reconstructed images. METHODS By further deriving the Chambolle-Pock (CP) algorithm instance for CT reconstruction, we discovered that the TV prior updates the reconstructed images with its divergences in each iteration of the solution process. Based on this discovery, CNNs were applied to yield the divergences of the feature maps for the reconstructed image generated in each iteration. Additionally, a loss function was applied to the predicted divergences of the reconstructed image to guarantee that the CNNs' results were the divergences of the corresponding feature maps in the iteration. In this manner, the proposed CNNs seem to play the same roles in the PD-Net as the TV prior in the IR methods. Thus, the TV prior in the CP algorithm instance can be directly unrolled to CNNs. RESULTS The datasets from the Low-Dose CT Image and Projection Data and the Piglet dataset were employed to assess the effectiveness of our proposed PD-Net. Compared with conventional CT reconstruction methods, our proposed method effectively preserves the structural and textural information in reference to ground truth. CONCLUSIONS The experimental results show that our proposed PD-Net framework is feasible for the implementation of CT reconstruction tasks. Owing to the promising results yielded by our proposed neural network, this study is intended to inspire further development of unrolling approaches by enabling the direct unrolling of hand-crafted prior terms to CNNs.
Collapse
|
13
|
CMF-Net: craniomaxillofacial landmark localization on CBCT images using geometric constraint and transformer. Phys Med Biol 2023; 68. [PMID: 36652722 DOI: 10.1088/1361-6560/acb483] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2022] [Accepted: 01/18/2023] [Indexed: 01/20/2023]
Abstract
Accurate and robust anatomical landmark localization is a mandatory and crucial step in deformation diagnosis and treatment planning for patients with craniomaxillofacial (CMF) malformations. In this paper, we propose a trainable end-to-end cephalometric landmark localization framework on CBCT scans, referred to as CMF-Net, which combines the appearance with transformers, geometric constraint, and adaptive wing (AWing) loss. More precisely: 1) We decompose the localization task into two branches: the appearance branch integrates transformers for identifying the exact positions of candidates, while the geometric constraint branch at low resolution allows the implicit spatial relationships to be effectively learned on the reduced training data. 2) We use the AWing loss to leverage the difference between the pixel values of the target heatmaps and the automatic prediction heatmaps. We verify our CMF-Net by identifying the 24 most relevant clinical landmarks on 150 dental CBCT scans with complicated scenarios collected from real-world clinics. Comprehensive experiments show that it performs better than the state-of-the-art deep learning methods, with an average localization error of 1.108 mm (the clinically acceptable precision range being 1.5 mm) and a correct landmark detection rate equal to 79.28%. Our CMF-Net is time-efficient and able to locate skull landmarks with high accuracy and significant robustness. This approach could be applied in 3D cephalometric measurement, analysis, and surgical planning.
Collapse
|
14
|
CTANet: Confidence-based Threshold Adaption Network for Semi-supervised Segmentation of Uterine Regions from MR Images for HIFU Treatment. Ing Rech Biomed 2023. [DOI: 10.1016/j.irbm.2022.100747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023]
|
15
|
Automatic segmentation for plan-of-the-day selection in CBCT-guided adaptive radiation therapy of cervical cancer. Phys Med Biol 2022; 67. [PMID: 36541494 DOI: 10.1088/1361-6560/aca5e5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2022] [Accepted: 11/24/2022] [Indexed: 11/25/2022]
Abstract
Objective.Plan-of-the-day (PoD) adaptive radiation therapy (ART) is based on a library of treatment plans, among which, at each treatment fraction, the PoD is selected using daily images. However, this strategy is limited by PoD selection uncertainties. This work aimed to propose and evaluate a workflow to automatically and quantitatively identify the PoD for cervix cancer ART based on daily CBCT images.Approach.The quantification was based on the segmentation of the main structures of interest in the CBCT images (clinical target volume [CTV], rectum, bladder, and bowel bag) using a deep learning model. Then, the PoD was selected from the treatment plan library according to the geometrical coverage of the CTV. For the evaluation, the resulting PoD was compared to the one obtained considering reference CBCT delineations.Main results.In experiments on a database of 23 patients with 272 CBCT images, the proposed method obtained an agreement between the reference PoD and the automatically identified PoD for 91.5% of treatment fractions (99.6% when considering a 5% margin on CTV coverage).Significance.The proposed automatic workflow automatically selected PoD for ART using deep-learning methods. The results showed the ability of the proposed process to identify the optimal PoD in a treatment plan library.
Collapse
|
16
|
Development and verification of radiomics framework for computed tomography image segmentation. Med Phys 2022; 49:6527-6537. [PMID: 35917213 PMCID: PMC9805121 DOI: 10.1002/mp.15904] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Revised: 07/19/2022] [Accepted: 07/21/2022] [Indexed: 01/09/2023] Open
Abstract
BACKGROUND Radiomics has been considered an imaging marker for capturing quantitative image information (QII). The introduction of radiomics to image segmentation is desirable but challenging. PURPOSE This study aims to develop and validate a radiomics-based framework for image segmentation (RFIS). METHODS RFIS is designed using features extracted from volume (svfeatures) created by sliding window (swvolume). The 53 svfeatures are extracted from 11 phantom series. Outliers in the svfeature datasets are detected by isolation forest (iForest) and specified as the mean value. The percentage coefficient of variation (%COV) is calculated to evaluate the reproducibility of svfeatures. RFIS is constructed and applied to the gross target volume (GTV) segmentation from the peritumoral region (GTV with a 10 mm margin) to assess its feasibility. The 127 lung cancer images are enrolled. The test-retest method, correlation matrix, and Mann-Whitney U test (p < 0.05) are used to select non-redundant svfeatures of statistical significance from the reproducible svfeatures. The synthetic minority over-sampling technique is utilized to balance the minority group in the training sets. The support vector machine is employed for RFIS construction, which is tuned in the training set using 10-fold stratified cross-validation and then evaluated in the test sets. The swvolumes with the consistent classification results are grouped and merged. Mode filtering is performed to remove very small subvolumes and create relatively large regions of completely uniform character. In addition, RFIS performance is evaluated by the area under the receiver operating characteristic (ROC) curve (AUC), accuracy, sensitivity, specificity, and Dice similarity coefficient (DSC). RESULTS 30249 phantom and 145008 patient image swvolumes were analyzed. Forty-nine (92.45% of 53) svfeatures represented excellent reproducibility(%COV<15). Forty-five features (91.84% of 49) included five categories that passed test-retest analysis. Thirteen svfeatures (28.89% of 45) svfeatures were selected for RFIS construction. RFIS showed an average (95% confidence interval) sensitivity of 0.848 (95% CI:0.844-0.883), a specificity of 0.821 (95% CI: 0.818-0.825), an accuracy of 83.48% (95% CI: 83.27%-83.70%), and an AUC of 0.906 (95% CI: 0.904-0.908) with cross-validation. The sensitivity, specificity, accuracy, and AUC were equal to 0.762 (95% CI: 0.754-0.770), 0.840 (95% CI: 0.837-0.844), 82.29% (95% CI: 81.90%-82.60%), and 0.877 (95% CI: 0.873-0.881) in the test set, respectively. GTV was segmented by grouping and merging swvolume with identical classification results. The mean DSC after mode filtering was 0.707 ± 0.093 in the training sets and 0.688 ± 0.072 in the test sets. CONCLUSION Reproducible svfeatures can capture the differences in QII among swvolumes. RFIS can be applied to swvolume classification, which achieves image segmentation by grouping and merging the swvolume with similar QII.
Collapse
|
17
|
Edge feature extraction-based dual CNN for LDCT denoising. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2022; 39:1929-1938. [PMID: 36215566 DOI: 10.1364/josaa.462923] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/02/2022] [Accepted: 09/14/2022] [Indexed: 06/16/2023]
Abstract
In low-dose computed tomography (LDCT) denoising tasks, it is often difficult to balance edge/detail preservation and noise/artifact reduction. To solve this problem, we propose a dual convolutional neural network (CNN) based on edge feature extraction (Ed-DuCNN) for LDCT. Ed-DuCNN consists of two branches. One branch is the edge feature extraction subnet (Edge_Net) that can fully extract the edge details in the image. The other branch is the feature fusion subnet (Fusion_Net) that introduces an attention mechanism to fuse edge features and noisy image features. Specifically, first, shallow edge-specific detail features are extracted by trainable Sobel convolutional blocks and then are integrated into Edge_Net together with the LDCT images to obtain deep edge detail features. Finally, the input image, shallow edge detail, and deep edge detail features are fused in Fusion_Net to generate the final denoised image. The experimental results show that the proposed Ed-DuCNN can achieve competitive performance in terms of quantitative metrics and visual perceptual quality compared with that of state-of-the-art methods.
Collapse
|
18
|
Convolutional Modulation Theory: A bridge between Convolutional Neural Networks and Signal Modulation Theory. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.09.088] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
19
|
Structure boundary-preserving U-Net for prostate ultrasound image segmentation. Front Oncol 2022; 12:900340. [PMID: 35965563 PMCID: PMC9366193 DOI: 10.3389/fonc.2022.900340] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2022] [Accepted: 06/30/2022] [Indexed: 11/19/2022] Open
Abstract
Prostate cancer diagnosis is performed under ultrasound-guided puncture for pathological cell extraction. However, determining accurate prostate location remains a challenge from two aspects: (1) prostate boundary in ultrasound images is always ambiguous; (2) the delineation of radiologists always occupies multiple pixels, leading to many disturbing points around the actual contour. We proposed a boundary structure-preserving U-Net (BSP U-Net) in this paper to achieve precise prostate contour. BSP U-Net incorporates prostate shape prior to traditional U-Net. The prior shape is built by the key point selection module, which is an active shape model-based method. Then, the module plugs into the traditional U-Net structure network to achieve prostate segmentation. The experiments were conducted on two datasets: PH2 + ISBI 2016 challenge and our private prostate ultrasound dataset. The results on PH2 + ISBI 2016 challenge achieved a Dice similarity coefficient (DSC) of 95.94% and a Jaccard coefficient (JC) of 88.58%. The results of prostate contour based on our method achieved a higher pixel accuracy of 97.05%, a mean intersection over union of 93.65%, a DSC of 92.54%, and a JC of 93.16%. The experimental results show that the proposed BSP U-Net has good performance on PH2 + ISBI 2016 challenge and prostate ultrasound image segmentation and outperforms other state-of-the-art methods.
Collapse
|
20
|
Artificial intelligence assisted cytological detection for early esophageal squamous epithelial lesions by using low-grade squamous intraepithelial lesion as diagnostic threshold. Cancer Med 2022; 12:1228-1236. [PMID: 35766144 PMCID: PMC9883535 DOI: 10.1002/cam4.4984] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2022] [Revised: 05/19/2022] [Accepted: 06/13/2022] [Indexed: 02/02/2023] Open
Abstract
BACKGROUND Manual cytological diagnosis for early esophageal squamous cell carcinoma (early ESCC) and high-grade intraepithelial neoplasia (HGIN) is unsatisfactory. Herein, we have introduced an artificial intelligence (AI)-assisted cytological diagnosis for such lesions. METHODS Low-grade squamous intraepithelial lesion or worse was set as the diagnostic threshold for AI-assisted diagnosis. The performance of AI-assisted diagnosis was evaluated and compared to that of manual diagnosis. Feasibility in large-scale screening was also assessed. RESULTS AI-assisted diagnosis for abnormal cells was superior to manual reading by presenting a higher efficiency for each slide (50.9 ± 0.8 s vs 236.8 ± 3.9 s, p = 1.52 × 10-76 ) and a better interobserver agreement (93.27% [95% CI, 92.76%-93.74%] vs 65.29% [95% CI, 64.35%-66.22%], p = 1.03 × 10-84 ). AI-assisted detection showed a higher diagnostic accuracy (96.89% [92.38%-98.57%] vs 72.54% [65.85%-78.35%], p = 1.42 × 10-14 ), sensitivity (99.35% [95.92%-99.97%] vs 68.39% [60.36%-75.48%], p = 7.11 × 10-15 ), and negative predictive value (NPV) (97.06% [82.95%-99.85%] vs 40.96% [30.46%-52.31%], p = 1.42 × 10-14 ). Specificity and positive predictive value (PPV) were not significantly differed. AI-assisted diagnosis demonstrated a smaller proportion of participants of interest (3.73%, [79/2117] vs.12.84% [272/2117], p = 1.59 × 10-58 ), a higher consistence between cytology and endoscopy (40.51% [32/79] vs. 12.13% [33/272], p = 1.54 × 10- 8), specificity (97.74% [96.98%-98.32%] vs 88.52% [87.05%-89.84%], p = 3.19 × 10-58 ), and PPV (40.51% [29.79%-52.15%] vs 12.13% [8.61%-16.75%], p = 1.54 × 10-8 ) in community-based screening. Sensitivity and NPV were not significantly differed. AI-assisted diagnosis as primary screening significantly reduced average cost for detecting positive cases. CONCLUSION Our study provides a novel cytological method for detecting and screening early ESCC and HGIN.
Collapse
|
21
|
Investigation of two neural mass models for DCM-based effective connectivity inference in temporal epilepsy. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 221:106840. [PMID: 35550455 DOI: 10.1016/j.cmpb.2022.106840] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/17/2021] [Revised: 04/13/2022] [Accepted: 04/25/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Recently, spectral Dynamic Causal Modelling (DCM) has been used increasingly to infer effective connectivity from epileptic intracranial electroencephalographic (iEEG) signals. In this context, the Physiology-Based Model (PBM), a neural mass model, is used as a generative model. However, previous studies have highlighted out the inability of PBM to properly describe iEEG signals with specific power spectral densities (PSDs). More precisely, PSDs that have multiple peaks around β and γ rhythms (i.e. spectral characteristics at seizure onset) are concerned. METHODS To cope with this limitation, an alternative neural mass model, called the complete PBM (cPBM), is investigated. The spectral DCM and two recent variants are used to evaluate the relevance of cPBM over PBM. RESULTS The study is conducted on both simulated signals and real epileptic iEEG recordings. Our results confirm that, compared to PBM, cPBM shows (i) more ability to model the desired PSDs and (ii) lower numerical complexity whatever the method. CONCLUSIONS Thanks to its intrinsic and extrinsic connectivity parameters as well as the input coming into the fast inhibitory subpopulation, the cPBM provides a more expressive model of PSDs, leading to a better understanding of epileptic patterns and DCM-based effective connectivity inference.
Collapse
|
22
|
Epileptic Seizure Prediction Using Deep Neural Networks via Transfer Learning and Multi-Feature Fusion. Int J Neural Syst 2022; 32:2250032. [DOI: 10.1142/s0129065722500320] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
23
|
Landmark Localization for Cephalometric Analysis using Multiscale Image Patch-based Graph Convolutional Networks. IEEE J Biomed Health Inform 2022; 26:3015-3024. [PMID: 35259123 DOI: 10.1109/jbhi.2022.3157722] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Accurate and robust cephalometric image analysis plays an essential role in orthodontic diagnosis, treatment assessment and surgical planning. This paper proposes a novel landmark localization method for cephalometric analysis using multiscale image patch-based graph convolutional networks. In detail, image patches with the same size are hierarchically sampled from the Gaussian pyramid to well preserve multiscale context information. We combine local appearance and shape information into spatialized features with an attention module to enrich node representations in graph. The spatial relationships of landmarks are built with the incorporation of three-layer graph convolutional networks, and multiple landmarks are simultaneously updated and moved toward the targets in a cascaded coarse-to-fine process. Quantitative results obtained on publicly available cephalometric X-ray images have exhibited superior performance compared with other state-of-the-art methods in terms of mean radial error and successful detection rate within various precision ranges. Our approach performs significantly better especially in the clinically accepted range of 2 mm and this makes it suitable in cephalometric analysis and orthognathic surgery.
Collapse
|
24
|
[Analysis of screening results and risk factors of high-risk populations of lung cancer in Nanchang city from 2018 to 2019]. ZHONGHUA ZHONG LIU ZA ZHI [CHINESE JOURNAL OF ONCOLOGY] 2021; 43:1316-1321. [PMID: 34915643 DOI: 10.3760/cma.j.cn112152-20200615-00559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
Objective: To collate and analyze the screening results of high-risk lung cancer populations in communities in Nanchang from 2018 to 2019, and to explore the lung-positive nodules and risk factors for lung cancer. Methods: Data of the screening subjects in 8 administrative districts and 15 street health service centers in Nanchang city, Jiangxi province from November 2018 to October 2019 were collected, people at high risk of lung cancer was assessed, clinical screening of high-risk groups of lung cancer was conducted by low-dose helical computed tomography (LDCT), and risk factors for suspected lung cancer and lung-positive nodules were analyzed. Results: Of the 25 871 people participated in screening, 5 220 were at high risk for lung cancer and 15 374 without other malignant tumors were at high risk. There were 2 417 cases participated in clinical LDCT screening, including 193 cases of lung-positive nodules, 67 cases of suspected lung cancer, 912 cases of other lung diseases, the positive rate of lung cancer or lung-positive nodules was 10.76% (260/2 417). Univariate analysis showed that age, coarse grain intake, oil intake, housing heating, passive smoking, alcohol consumption and mental trauma were associated with positive pulmonary nodules or lung cancer (all P<0.05). Multivariate analysis showed that gender, age, housing heating, smoking and drinking were related to the occurrence of lung nodules or lung cancer (all P<0.05). Conclusions: Men are more likely to develop lung cancer or lung-positive nodules than women. The age is an independent risk factor for lung-positive nodules or lung cancer. In a certain range, age will increase the incidence of lung cancer, housing heating may be the protective factor for lung cancer, while smoking and drinking are risk factors.
Collapse
|
25
|
Trajectory Grouping With Curvature Regularization for Tubular Structure Tracking. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 31:405-418. [PMID: 34874858 DOI: 10.1109/tip.2021.3131940] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Tubular structure tracking is a crucial task in the fields of computer vision and medical image analysis. The minimal paths-based approaches have exhibited their strong ability in tracing tubular structures, by which a tubular structure can be naturally modeled as a minimal geodesic path computed with a suitable geodesic metric. However, existing minimal paths-based tracing approaches still suffer from difficulties such as the shortcuts and short branches combination problems, especially when dealing with the images involving complicated tubular tree structures or background. In this paper, we introduce a new minimal paths-based model for minimally interactive tubular structure centerline extraction in conjunction with a perceptual grouping scheme. Basically, we take into account the prescribed tubular trajectories and curvature-penalized geodesic paths to seek suitable shortest paths. The proposed approach can benefit from the local smoothness prior on tubular structures and the global optimality of the used graph-based path searching scheme. Experimental results on both synthetic and real images prove that the proposed model indeed obtains outperformance comparing with the state-of-the-art minimal paths-based tubular structure tracing algorithms.
Collapse
|
26
|
Examinee-Examiner Network: Weakly Supervised Accurate Coronary Lumen Segmentation Using Centerline Constraint. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:9429-9441. [PMID: 34757906 DOI: 10.1109/tip.2021.3125490] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Accurate coronary lumen segmentation on coronary-computed tomography angiography (CCTA) images is crucial for quantification of coronary stenosis and the subsequent computation of fractional flow reserve. Many factors including difficulty in labeling coronary lumens, various morphologies in stenotic lesions, thin structures and small volume ratio with respect to the imaging field complicate the task. In this work, we fused the continuity topological information of centerlines which are easily accessible, and proposed a novel weakly supervised model, Examinee-Examiner Network (EE-Net), to overcome the challenges in automatic coronary lumen segmentation. First, the EE-Net was proposed to address the fracture in segmentation caused by stenoses by combining the semantic features of lumens and the geometric constraints of continuous topology obtained from the centerlines. Then, a Centerline Gaussian Mask Module was proposed to deal with the insensitiveness of the network to the centerlines. Subsequently, a weakly supervised learning strategy, Examinee-Examiner Learning, was proposed to handle the weakly supervised situation with few lumen labels by using our EE-Net to guide and constrain the segmentation with customized prior conditions. Finally, a general network layer, Drop Output Layer, was proposed to adapt to the class imbalance by dropping well-segmented regions and weights the classes dynamically. Extensive experiments on two different data sets demonstrated that our EE-Net has good continuity and generalization ability on coronary lumen segmentation task compared with several widely used CNNs such as 3D-UNet. The results revealed our EE-Net with great potential for achieving accurate coronary lumen segmentation in patients with coronary artery disease. Code at http://github.com/qiyaolei/Examinee-Examiner-Network.
Collapse
|
27
|
Dissected aorta segmentation using convolutional neural networks. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2021; 211:106417. [PMID: 34587564 DOI: 10.1016/j.cmpb.2021.106417] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/18/2021] [Accepted: 09/12/2021] [Indexed: 06/13/2023]
Abstract
BACKGROUND AND OBJECTIVE Aortic dissection is a severe cardiovascular pathology in which an injury of the intimal layer of the aorta allows blood flowing into the aortic wall, forcing the wall layers apart. Such situation presents a high mortality rate and requires an in-depth understanding of the 3-D morphology of the dissected aorta to plan the right treatment. An accurate automatic segmentation algorithm is therefore needed. METHOD In this paper, we propose a deep-learning-based algorithm to segment dissected aorta on computed tomography angiography (CTA) images. The algorithm consists of two steps. Firstly, a 3-D convolutional neural network (CNN) is applied to divide the 3-D volume into two anatomical portions. Secondly, two 2-D CNNs based on pyramid scene parsing network (PSPnet) segment each specific portion separately. An edge extraction branch was added to the 2-D model to get higher segmentation accuracy on intimal flap area. RESULTS The experiments conducted and the comparisons made show that the proposed solution performs well with an average dice index over 92%. The combination of 3-D and 2-D models improves the aorta segmentation accuracy compared to 3-D only models and the segmentation robustness compared to 2-D only models. The edge extraction branch improves the DICE index near aorta boundaries from 73.41% to 81.39%. CONCLUSIONS The proposed algorithm has satisfying performance for capturing the aorta structure while avoiding false positives on the intimal flaps.
Collapse
|
28
|
Fractional Wavelet-Based Generative Scattering Networks. Front Neurorobot 2021; 15:752752. [PMID: 34764862 PMCID: PMC8577828 DOI: 10.3389/fnbot.2021.752752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2021] [Accepted: 09/20/2021] [Indexed: 11/13/2022] Open
Abstract
Generative adversarial networks and variational autoencoders (VAEs) provide impressive image generation from Gaussian white noise, but both are difficult to train, since they need a generator (or encoder) and a discriminator (or decoder) to be trained simultaneously, which can easily lead to unstable training. To solve or alleviate these synchronous training problems of generative adversarial networks (GANs) and VAEs, researchers recently proposed generative scattering networks (GSNs), which use wavelet scattering networks (ScatNets) as the encoder to obtain features (or ScatNet embeddings) and convolutional neural networks (CNNs) as the decoder to generate an image. The advantage of GSNs is that the parameters of ScatNets do not need to be learned, while the disadvantage of GSNs is that their ability to obtain representations of ScatNets is slightly weaker than that of CNNs. In addition, the dimensionality reduction method of principal component analysis (PCA) can easily lead to overfitting in the training of GSNs and, therefore, affect the quality of generated images in the testing process. To further improve the quality of generated images while keeping the advantages of GSNs, this study proposes generative fractional scattering networks (GFRSNs), which use more expressive fractional wavelet scattering networks (FrScatNets), instead of ScatNets as the encoder to obtain features (or FrScatNet embeddings) and use similar CNNs of GSNs as the decoder to generate an image. Additionally, this study develops a new dimensionality reduction method named feature-map fusion (FMF) instead of performing PCA to better retain the information of FrScatNets,; it also discusses the effect of image fusion on the quality of the generated image. The experimental results obtained on the CIFAR-10 and CelebA datasets show that the proposed GFRSNs can lead to better generated images than the original GSNs on testing datasets. The experimental results of the proposed GFRSNs with deep convolutional GAN (DCGAN), progressive GAN (PGAN), and CycleGAN are also given.
Collapse
|
29
|
MVSGAN: Spatial-aware Multi-view CMR Fusion for accurate 3D Left Ventricular Myocardium Segmentation. IEEE J Biomed Health Inform 2021; 26:2264-2275. [PMID: 34699378 DOI: 10.1109/jbhi.2021.3122581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
The accurate 3D left ventricular (LV) myocardium segmentation in short-axis (SAX) view of cardiac magnetic resonance (CMR) is challenged by the sparse spatial structure of CMR. The strategy of multi-view CMR fusion can provide fine-grained spatial structure for accurate segmentation. However, the large information misalignment & lack of dense 3D CMR as fusion target in multi-view CMR fusion, and the different spatial resolution between the fused cardiac model and the ground truth of segmentation in segmentation limit the strategy. In this study, we propose a multi-view spatial-aware adversarial network (MVSGAN). It studies the perception of fine-grained cardiac structure for accurate segmentation by the spatially consistent fusion of multi-view CMR. It consists of three modules: (1) A residual adversarial fusion (RAF) module takes inter-slices deep correlation and anatomical prior of multi-view CMR to refine the spatial structures by residual supplement and adversarial optimization. (2) A structural perception-aggregation (SPA) module establishes the spatial correlation between the spatially dense cardiac model and sparse segmentation label for accurate 3D CMR LV myocardium segmentation. (3) A joint training strategy utilizes the spatial dense SAX volume as explicit and implicit goals to jointly optimize the framework. The experiments are applied on a public dataset and a clinical dataset to evaluate the performance of MVSGAN. The average Dice and Jaccard score of LV myocardium segmentation obtained by MVSGAN are highest among seven existing state-of-the-art methods, which are up to 0.92 and 0.75. It is concluded that the spatial-aware multiview CMR fusion can provide meaningful spatial correlation for accurate 3D SAX LV myocardium segmentation.
Collapse
|
30
|
An adaptive optimal viewing angle determination algorithm for TEVAR operation. BMC Med Imaging 2021; 21:141. [PMID: 34600478 PMCID: PMC8487120 DOI: 10.1186/s12880-021-00676-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2021] [Accepted: 09/28/2021] [Indexed: 11/13/2022] Open
Abstract
BACKGROUND The determination of the right x-ray angiography viewing angle is an important issue during the treatment of thoracic endovascular aortic repair (TEVAR). An inaccurate projection angle (manually determined today by the physicians according to their personal experience) may affect the placement of the stent and cause vascular occlusion or endoleak. METHODS Based on the acquisition of a computed tomography angiography (CTA) image before TEVAR, an adaptive optimization algorithm is proposed to determine the optimal viewing angle of the angiogram automatically. This optimal view aims at avoiding any overlapping between the left common carotid artery and the left subclavian artery. Moreover, the proposed optimal procedure exploits the patient-specific morphology to adaptively reduce the potential foreshortening effect. RESULTS Experimental results conducted on thirty-five patients demonstrate that the optimal angiographic viewing angle based on the proposed method has no significant difference when compared with the expert practice (p = 0.0678). CONCLUSION We propose a method that utilizes the CTA image acquired before TEVAR to automatically calculate the optimal C-arm angle. This method has the potential to assist surgeons during their interventional procedure by providing a shorter procedure time, less radiation exposure, and less contrast injection.
Collapse
|
31
|
Few-shot Learning for Deformable Medical Image Registration with Perception-Correspondence Decoupling and Reverse Teaching. IEEE J Biomed Health Inform 2021; 26:1177-1187. [PMID: 34232899 DOI: 10.1109/jbhi.2021.3095409] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Deformable medical image registration estimates corresponding deformation to align the regions of interest (ROIs) of two images to a same spatial coordinate system. However, recent unsupervised registration models only have correspondence ability without perception, making misalignment on blurred anatomies and distortion on task-unconcerned backgrounds. Label-constrained (LC) registration models embed the perception ability via labels, but the lack of texture constraints in labels and the expensive labeling costs causes distortion internal ROIs and overfitted perception. We propose the first few-shot deformable medical image registration framework, Perception-Correspondence Registration (PC-Reg), which embeds perception ability to registration models only with few labels, thus greatly improving registration accuracy and reducing distortion. 1) We propose the Perception-Correspondence Decoupling which decouples the perception and correspondence actions of registration to two CNNs. Therefore, independent optimizations and feature representations are available avoiding interference of the correspondence due to the lack of texture constraints. 2) For few-shot learning, we propose Reverse Teaching which aligns labeled and unlabeled images to each other to provide supervision information to the structure and style knowledge in unlabeled images, thus generating additional training data. Therefore, these data will reversely teach our perception CNN more style and structure knowledge, improving its generalization ability. Our experiments on three datasets with only five labels demonstrate that our PC-Reg has competitive registration accuracy and effective distortion-reducing ability. Compared with LC-VoxelMorph(lambda=1), we achieve the 12.5%, 6.3% and 1.0% Reg-DSC improvements on three datasets, revealing our framework with great potential in clinical application.
Collapse
|
32
|
Spatio-temporal graph convolutional network for diagnosis and treatment response prediction of major depressive disorder from functional connectivity. Hum Brain Mapp 2021; 42:3922-3933. [PMID: 33969930 PMCID: PMC8288094 DOI: 10.1002/hbm.25529] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2021] [Revised: 04/17/2021] [Accepted: 05/02/2021] [Indexed: 12/14/2022] Open
Abstract
The pathophysiology of major depressive disorder (MDD) has been explored to be highly associated with the dysfunctional integration of brain networks. It is therefore imperative to explore neuroimaging biomarkers to aid diagnosis and treatment. In this study, we developed a spatiotemporal graph convolutional network (STGCN) framework to learn discriminative features from functional connectivity for automatic diagnosis and treatment response prediction of MDD. Briefly, dynamic functional networks were first obtained from the resting-state fMRI with the sliding temporal window method. Secondly, a novel STGCN approach was proposed by introducing the modules of spatial graph attention convolution (SGAC) and temporal fusion. A novel SGAC was proposed to improve the feature learning ability and special anatomy prior guided pooling was developed to enable the feature dimension reduction. A temporal fusion module was proposed to capture the dynamic features of functional connectivity between adjacent sliding windows. Finally, the STGCN proposed approach was utilized to the tasks of diagnosis and antidepressant treatment response prediction for MDD. Performances of the framework were comprehensively examined with large cohorts of clinical data, which demonstrated its effectiveness in classifying MDD patients and predicting the treatment response. The sound performance suggests the potential of the STGCN for the clinical use in diagnosis and treatment prediction.
Collapse
|
33
|
Characterisation of Wheat–Oat Flour Extrudate: Physicochemical and Β-Glucan Attributes. ACTA ALIMENTARIA 2020. [DOI: 10.1556/066.2020.49.4.8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Abstract
The objective of this study was to evaluate the effect of wheat–oat flour ratio on the physical properties and β-glucan characteristics of extrudates. Results showed that increasing the wheat–oat flour ratio resulted in a decrease in the water solubility index (r2=0.8567) and hardness (r2=0.9316), whereas the expansion ratio (r2=0.9307) and water absorption index (r2=0.9061) increased. Wheat flour generally caused an increase in L values from 57.81 to 62.94 providing bright samples. Few cells were observed at high wheat–oat flour ratios under a scanning electron microscope, and a smooth surface was noted. Meanwhile, the total (r2=0.9867) and soluble (r2=0.9848) β-glucan contents were inversely proportional to the wheat–oat flour ratio. Extrudates with added wheat flour had a high molecular weight, but wheat flour had no significant (P<0.05) effect on the viscosity of β-glucan extracts. Conclusively, incorporation of wheat flour at a wheat–oat flour ratio of 2.33 provides ready-to-eat food based on whole oat flour, on account of improving the texture and providing sufficient β-glucan contents (0.806 g/100 g) without significantly affecting β-glucan viscosity.
Collapse
|
34
|
HIFUNet: Multi-Class Segmentation of Uterine Regions From MR Images Using Global Convolutional Networks for HIFU Surgery Planning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:3309-3320. [PMID: 32356741 DOI: 10.1109/tmi.2020.2991266] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Accurate segmentation of uterus, uterine fibroids, and spine from MR images is crucial for high intensity focused ultrasound (HIFU) therapy but remains still difficult to achieve because of 1) the large shape and size variations among individuals, 2) the low contrast between adjacent organs and tissues, and 3) the unknown number of uterine fibroids. To tackle this problem, in this paper, we propose a large kernel Encoder-Decoder Network based on a 2D segmentation model. The use of this large kernel can capture multi-scale contexts by enlarging the valid receptive field. In addition, a deep multiple atrous convolution block is also employed to enlarge the receptive field and extract denser feature maps. Our approach is compared to both conventional and other deep learning methods and the experimental results conducted on a large dataset show its effectiveness.
Collapse
|
35
|
Dense biased networks with deep priori anatomy and hard region adaptation: Semi-supervised learning for fine renal artery segmentation. Med Image Anal 2020; 63:101722. [DOI: 10.1016/j.media.2020.101722] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/09/2019] [Revised: 05/02/2020] [Accepted: 05/06/2020] [Indexed: 12/24/2022]
|
36
|
|
37
|
Vessel Structure Extraction using Constrained Minimal Path Propagation. Artif Intell Med 2020; 105:101846. [PMID: 32505425 DOI: 10.1016/j.artmed.2020.101846] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2018] [Revised: 10/23/2019] [Accepted: 03/20/2020] [Indexed: 11/18/2022]
Abstract
Minimal path method has been widely recognized as an efficient tool for extracting vascular structures in medical imaging. In a previous paper, a method termed minimal path propagation with backtracking (MPP-BT) was derived to deal with curve-like structures such as vessel centerlines. A robust approach termed CMPP (constrained minimal path propagation) is here proposed to extend this work. The proposed method utilizes another minimal path propagation procedure to extract the complete vessel lumen after the centerlines have been found. Moreover, a process named local MPP-BT is applied to handle structure missing caused by the so-called close loop problems. This approach is fast and unsupervised with only one roughly set start point required in the whole process to get the entire vascular structure. A variety of datasets, including 2D cardiac angiography, 2D retinal images and 3D kidney CT angiography, are used for validation. A quantitative evaluation, together with a comparison to recently reported methods, is performed on retinal images for which a ground truth is available. The proposed method leads to specificity (Sp) and sensitivity (Se) values equal to 0.9750 and 0.6591. This evaluation is also extended to 3D synthetic vascular datasets and shows that the specificity (Sp) and sensitivity (Se) values are higher than 0.99. Parameter setting and computation cost are analyzed in this paper.
Collapse
|
38
|
Weakly-supervised convolutional neural networks of renal tumor segmentation in abdominal CTA images. BMC Med Imaging 2020; 20:37. [PMID: 32293303 PMCID: PMC7161012 DOI: 10.1186/s12880-020-00435-w] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2019] [Accepted: 03/20/2020] [Indexed: 11/23/2022] Open
Abstract
Background Renal cancer is one of the 10 most common cancers in human beings. The laparoscopic partial nephrectomy (LPN) is an effective way to treat renal cancer. Localization and delineation of the renal tumor from pre-operative CT Angiography (CTA) is an important step for LPN surgery planning. Recently, with the development of the technique of deep learning, deep neural networks can be trained to provide accurate pixel-wise renal tumor segmentation in CTA images. However, constructing the training dataset with a large amount of pixel-wise annotations is a time-consuming task for the radiologists. Therefore, weakly-supervised approaches attract more interest in research. Methods In this paper, we proposed a novel weakly-supervised convolutional neural network (CNN) for renal tumor segmentation. A three-stage framework was introduced to train the CNN with the weak annotations of renal tumors, i.e. the bounding boxes of renal tumors. The framework includes pseudo masks generation, group and weighted training phases. Clinical abdominal CT angiographic images of 200 patients were applied to perform the evaluation. Results Extensive experimental results show that the proposed method achieves a higher dice coefficient (DSC) of 0.826 than the other two existing weakly-supervised deep neural networks. Furthermore, the segmentation performance is close to the fully supervised deep CNN. Conclusions The proposed strategy improves not only the efficiency of network training but also the precision of the segmentation.
Collapse
|
39
|
Fast and accurate segmentation method of active shape model with Rayleigh mixture model clustering for prostate ultrasound images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2020; 184:105097. [PMID: 31634807 DOI: 10.1016/j.cmpb.2019.105097] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/18/2018] [Revised: 09/24/2019] [Accepted: 09/25/2019] [Indexed: 06/10/2023]
Abstract
BACKGROUND AND OBJECTIVE The prostate cancer interventions, which need an accurate prostate segmentation, are performed under ultrasound imaging guidance. However, prostate ultrasound segmentation is facing two challenges. The first is the low signal-to-noise ratio and inhomogeneity of the ultrasound image. The second is the non-standardized shape and size of the prostate. METHODS For prostate ultrasound image segmentation, this paper proposed an accurate and efficient method of Active shape model (ASM) with Rayleigh mixture model Clustering (ASM-RMMC). Firstly, Rayleigh mixture model (RMM) is adopted for clustering the image regions which present similar speckle distributions. These content-based clustered images are then used to initialize and guide the deformation of an ASM model. RESULTS The performance of the proposed method is assessed on 30 prostate ultrasound images using four metrics as Mean Average Distance (MAD), Dice Similarity Coefficient (DSC), False Positive Error (FPE) and False Negative Error (FNE). The proposed ASM-RMMC reaches high segmentation accuracy with 95% ± 0.81% for DSC, 1.86 ± 0.02 pixels for MAD, 2.10% ± 0.36% for FPE and 2.78% ± 0.71% for FNE, respectively. Moreover, the average segmentation time is less than 8 s when treating a single prostate ultrasound image through ASM-RMMC. CONCLUSIONS This paper presents a method for prostate ultrasound image segmentation, which achieves high accuracy with less computational complexity and meets the clinical requirements.
Collapse
|
40
|
Abstract
Total variation (TV) based models are very popular in image denoising but suffer from some drawbacks. For example, local TV methods often cannot preserve edges and textures well when they face excessive smoothing. Non-local TV methods constitute an alternative, but their computational cost is huge. To overcome these issues, we propose an image denoising method named non-local patch graph total variation (NPGTV). Its main originality stands for the graph total variation method, which combines the total variation with graph signal processing. Schematically, we first construct a K-nearest graph from the original image using a non-local patch-based method. Then the model is solved with the Douglas-Rachford Splitting algorithm. By doing so, the image details can be well preserved while being denoised. Experiments conducted on several standard natural images illustrate the effectiveness of our method when compared to some other state-of-the-art denoising methods like classical total variation, non-local means filter (NLM), non-local graph based transform (NLGBT), adaptive graph-based total variation (AGTV).
Collapse
|
41
|
Domain Progressive 3D Residual Convolution Network to Improve Low-Dose CT Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2019; 38:2903-2913. [PMID: 31107644 DOI: 10.1109/tmi.2019.2917258] [Citation(s) in RCA: 90] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/19/2023]
Abstract
The wide applications of X-ray computed tomography (CT) bring low-dose CT (LDCT) into a clinical prerequisite, but reducing the radiation exposure in CT often leads to significantly increased noise and artifacts, which might lower the judgment accuracy of radiologists. In this paper, we put forward a domain progressive 3D residual convolution network (DP-ResNet) for the LDCT imaging procedure that contains three stages: sinogram domain network (SD-net), filtered back projection (FBP), and image domain network (ID-net). Though both are based on the residual network structure, the SD-net and ID-net provide complementary effect on improving the final LDCT quality. The experimental results with both simulated and real projection data show that this domain progressive deep-learning network achieves significantly improved performance by combing the network processing in the two domains.
Collapse
|
42
|
Prescription Value-Based Automatic Optimization of Importance Factors in Inverse Planning. Technol Cancer Res Treat 2019; 18:1533033819892259. [PMID: 31782353 PMCID: PMC6886287 DOI: 10.1177/1533033819892259] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/25/2022] Open
Abstract
Objective: An automatic method for the optimization of importance factors was proposed to improve the efficiency of inverse planning. Methods: The automatic method consists of 3 steps: (1) First, the importance factors are automatically and iteratively adjusted based on our proposed penalty strategies. (2) Then, plan evaluation is performed to determine whether the obtained plan is acceptable. (3) If not, a higher penalty is assigned to the unsatisfied objective by multiplying it by a compensation coefficient. The optimization processes are performed alternately until an acceptable plan is obtained or the maximum iteration Nmax of step (3) is reached. Results: Tested on 2 kinds of clinical cases and compared with manual method, the results showed that the quality of the proposed automatic plan was comparable to, or even better than, the manual plan in terms of the dose–volume histogram and dose distributions. Conclusions: The proposed algorithm has potential to significantly improve the efficiency of the existing manual adjustment methods for importance factors and contributes to the development of fully automated planning. Especially, the more the subobjective functions, the more obvious the advantage of our algorithm.
Collapse
|
43
|
Deep iterative reconstruction estimation (DIRE): approximate iterative reconstruction estimation for low dose CT imaging. ACTA ACUST UNITED AC 2019; 64:135007. [DOI: 10.1088/1361-6560/ab18db] [Citation(s) in RCA: 21] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
|
44
|
|
45
|
Abstract
OBJECTIVE This study introduces a fractional wavelet scattering network (FrScatNet), which is a generalized translation invariant version of the classical wavelet scattering network. METHODS In our approach, the FrScatNet is constructed based on the fractional wavelet transform (FRWT). The fractional scattering coefficients are iteratively computed using FRWTs and modulus operators. The feature vectors constructed by fractional scattering coefficients are usually used for signal classification. In this paper, an application example of the FrScatNet is provided in order to assess its performance on pathological images. First, the FrScatNet extracts feature vectors from patches of the original histological images under different orders. Then we classify those patches into target (benign or malignant) and background groups. And the FrScatNet property is analyzed by comparing error rates computed from different fractional orders, respectively. Based on the above pathological image classification, a gland segmentation algorithm is proposed by combining the boundary information and the gland location. RESULTS The error rates for different fractional orders of FrScatNet are examined and show that the classification accuracy is improved in fractional scattering domain. We also compare the FrScatNet-based gland segmentation method with those proposed in the 2015 MICCAI Gland Segmentation Challenge and our method achieves comparable results. CONCLUSION The FrScatNet is shown to achieve accurate and robust results. More stable and discriminative fractional scattering coefficients are obtained by the FrScatNet in this paper. SIGNIFICANCE The added fractional order parameter is able to analyze the image in the fractional scattering domain.
Collapse
|
46
|
Abstract
Background Accurate segmentation of brain tissues from magnetic resonance imaging (MRI) is of significant importance in clinical applications and neuroscience research. Accurate segmentation is challenging due to the tissue heterogeneity, which is caused by noise, bias filed and partial volume effects. Methods To overcome this limitation, this paper presents a novel algorithm for brain tissue segmentation based on supervoxel and graph filter. Firstly, an effective supervoxel method is employed to generate effective supervoxels for the 3D MRI image. Secondly, the supervoxels are classified into different types of tissues based on filtering of graph signals. Results The performance is evaluated on the BrainWeb 18 dataset and the Internet Brain Segmentation Repository (IBSR) 18 dataset. The proposed method achieves mean dice similarity coefficient (DSC) of 0.94, 0.92 and 0.90 for the segmentation of white matter (WM), grey matter (GM) and cerebrospinal fluid (CSF) for BrainWeb 18 dataset, and mean DSC of 0.85, 0.87 and 0.57 for the segmentation of WM, GM and CSF for IBSR18 dataset. Conclusions The proposed approach can well discriminate different types of brain tissues from the brain MRI image, which has high potential to be applied for clinical applications.
Collapse
|
47
|
Discriminative Feature Representation to Improve Projection Data Inconsistency for Low Dose CT Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2017; 36:2499-2509. [PMID: 28816658 DOI: 10.1109/tmi.2017.2739841] [Citation(s) in RCA: 46] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
In low dose computed tomography (LDCT) imaging, the data inconsistency of measured noisy projections can significantly deteriorate reconstruction images. To deal with this problem, we propose here a new sinogram restoration approach, the sinogram- discriminative feature representation (S-DFR) method. Different from other sinogram restoration methods, the proposed method works through a 3-D representation-based feature decomposition of the projected attenuation component and the noise component using a well-designed composite dictionary containing atoms with discriminative features. This method can be easily implemented with good robustness in parameter setting. Its comparison to other competing methods through experiments on simulated and real data demonstrated that the S-DFR method offers a sound alternative in LDCT.
Collapse
|
48
|
Discriminative Prior - Prior Image Constrained Compressed Sensing Reconstruction for Low-Dose CT Imaging. Sci Rep 2017; 7:13868. [PMID: 29066731 PMCID: PMC5655040 DOI: 10.1038/s41598-017-13520-y] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/09/2017] [Accepted: 09/25/2017] [Indexed: 12/15/2022] Open
Abstract
X-ray computed tomography (CT) has been widely used to provide patient-specific anatomical information in the forms of tissue attenuation. However, the cumulative radiation induced in CT scan has raised extensive concerns in recently years. How to maintain reconstruction image quality is a major challenge for low-dose CT (LDCT) imaging. Generally, LDCT imaging can be greatly improved by incorporating prior knowledge in some specific forms. A joint estimation framework termed discriminative prior-prior image constrained compressed sensing (DP-PICCS) reconstruction is proposed in this paper. This DP-PICCS algorithm utilizes discriminative prior knowledge via two feature dictionary constraints which built on atoms from the samples of tissue attenuation feature patches and noise-artifacts residual feature patches, respectively. Also, the prior image construction relies on a discriminative feature representation (DFR) processing by two feature dictionary. Its comparison to other competing methods through experiments on low-dose projections acquired from torso phantom simulation study and clinical abdomen study demonstrated that the DP-PICCS method achieved promising improvement in terms of the effectively-suppressed noise and the well-retained structures.
Collapse
|
49
|
Two-dimensional active raypath separation using examination of the roots of the spectrum polynomial. THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA 2017; 142:EL408. [PMID: 29092564 DOI: 10.1121/1.5008434] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
The application of the root multiple signal classification algorithm for raypath separation was motivated by the dramatic reduction in processing time of the multiple-signal classification algorithm. However, the algorithm provides classification only in the direction of the arrival domain and fails to separate raypaths arriving at the array with similar directions of arrival. Moreover, for many applications in shallow water (such as ocean acoustic tomography and active sonar), the emitted signal is known and can be used as a priori information to improve the resolution. Thus, in this study, a two-dimensional active wideband classification algorithm is developed using the examination of the roots of the spectrum polynomial in the angle versus time domain. A two-step strategy is developed to enable extension to the two-dimensional case. The results of simulations confirm that the proposed algorithm achieves almost identical resolution as the existing two-dimensional algorithms while offering a significant reduction in computation time.
Collapse
|
50
|
A local adjustment strategy for the initialization of dynamic causal modelling to infer effective connectivity in brain epileptic structures. Comput Biol Med 2017; 84:30-44. [PMID: 28340406 DOI: 10.1016/j.compbiomed.2017.03.006] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2016] [Revised: 03/03/2017] [Accepted: 03/04/2017] [Indexed: 01/22/2023]
Abstract
This paper addresses the question of effective connectivity in the human cerebral cortex in the context of epilepsy. Among model based approaches to infer brain connectivity, spectral Dynamic Causal Modelling is a conventional technique for which we propose an alternative to estimate cross spectral density. The proposed strategy we investigated tackles the sub-estimation of the free energy using the well-known variational Expectation-Maximization algorithm highly sensitive to the initialization of the parameters vector by a permanent local adjustment of the initialization process. The performance of the proposed strategy in terms of effective connectivity identification is assessed using simulated data generated by a neuronal mass model (simulating unidirectional and bidirectional flows) and real epileptic intracerebral Electroencephalographic signals. Results show the efficiency of proposed approach compared to the conventional Dynamic Causal Modelling and the one wherein a deterministic annealing scheme is employed.
Collapse
|