1
|
Zhu H, Cheng W, Hu K, He G. DSnet: a new dual-branch network for hippocampus subfield segmentation. Sci Rep 2024; 14:15317. [PMID: 38961218 PMCID: PMC11222372 DOI: 10.1038/s41598-024-66415-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Accepted: 07/01/2024] [Indexed: 07/05/2024] Open
Abstract
The hippocampus is a critical component of the brain and is associated with many neurological disorders. It can be further subdivided into several subfields, and accurate segmentation of these subfields is of great significance for diagnosis and research. However, the structures of hippocampal subfields are irregular and have complex boundaries, and their voxel values are close to surrounding brain tissues, making the segmentation task highly challenging. Currently, many automatic segmentation tools exist for hippocampal subfield segmentation, but they suffer from high time costs and low segmentation accuracy. In this paper, we propose a new dual-branch segmentation network structure (DSnet) based on deep learning for hippocampal subfield segmentation. While traditional convolutional neural network-based methods are effective in capturing hierarchical structures, they struggle to establish long-term dependencies. The DSnet integrates the Transformer architecture and a hybrid attention mechanism, enhancing the network's global perceptual capabilities. Moreover, the dual-branch structure of DSnet leverages the segmentation results of the hippocampal region to facilitate the segmentation of its subfields. We validate the efficacy of our algorithm on the public Kulaga-Yoskovitz dataset. Experimental results indicate that our method is more effective in segmenting hippocampal subfields than conventional single-branch network structures. Compared to the classic 3D U-Net, our proposed DSnet improves the average Dice accuracy of hippocampal subfield segmentation by 0.57%.
Collapse
Affiliation(s)
- Hancan Zhu
- School of Mathematics, Physics and Information, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China
- Institute of Artificial Intelligence, Shaoxing University, Shaoxing, 312000, Zhejiang, China
| | - Wangang Cheng
- School of Mathematics, Physics and Information, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China
| | - Keli Hu
- Institute of Artificial Intelligence, Shaoxing University, Shaoxing, 312000, Zhejiang, China
| | - Guanghua He
- School of Mathematics, Physics and Information, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China.
| |
Collapse
|
2
|
He G, Zhang G, Zhou L, Zhu H. Deep convolutional neural network for hippocampus segmentation with boundary region refinement. Med Biol Eng Comput 2023; 61:2329-2339. [PMID: 37067776 DOI: 10.1007/s11517-023-02836-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2022] [Accepted: 04/05/2023] [Indexed: 04/18/2023]
Abstract
Accurately segmenting the hippocampus from magnetic resonance (MR) brain images is a crucial step in studying brain disorders. However, this task is challenging due to the low signal contrast of hippocampal images, the irregular shape, and small structural size of the hippocampi. In recent years, several deep convolutional networks have been proposed for hippocampus segmentation, which have achieved state-of-the-art performance. These methods typically use large image patches for training the network, as larger patches are beneficial for capturing long-range contextual information. However, this approach increases the computational burden and overlooks the significance of the boundary region. In this study, we propose a deep learning-based method for hippocampus segmentation with boundary region refinement. Our method involves two main steps. First, we propose a convolutional network that takes large image patches as input for initial segmentation. Then, we extract small image patches around the hippocampal boundary for training the second convolutional neural network, which refines the segmentation in the boundary regions. We validate our proposed method on a publicly available dataset and demonstrate that it significantly improves the performance of convolutional neural networks that use single-size image patches as input. In conclusion, our study proposes a novel method for hippocampus segmentation, which improves upon the current state-of-the-art methods. By incorporating a boundary refinement step, our approach achieves higher accuracy in hippocampus segmentation and may facilitate research on brain disorders.
Collapse
Affiliation(s)
- Guanghua He
- School of Mathematics, Physics, and Information Science, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China
| | - Guying Zhang
- School of Mathematics, Physics, and Information Science, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China
| | - Lianlian Zhou
- School of Mathematics, Physics, and Information Science, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China
| | - Hancan Zhu
- School of Mathematics, Physics, and Information Science, Shaoxing University, 900 ChengNan Rd, Shaoxing, 312000, Zhejiang, China.
| |
Collapse
|
3
|
Zhang G, Zhou J, He G, Zhu H. Deep fusion of multi-modal features for brain tumor image segmentation. Heliyon 2023; 9:e19266. [PMID: 37664757 PMCID: PMC10468380 DOI: 10.1016/j.heliyon.2023.e19266] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Revised: 08/09/2023] [Accepted: 08/17/2023] [Indexed: 09/05/2023] Open
Abstract
Accurate segmentation of pathological regions in brain magnetic resonance images (MRI) is essential for the diagnosis and treatment of brain tumors. Multi-modality MRIs, which offer diverse feature information, are commonly utilized in brain tumor image segmentation. Deep neural networks have become prevalent in this field; however, many approaches simply concatenate different modalities and input them directly into the neural network for segmentation, disregarding the unique characteristics and complementarity of each modality. In this study, we propose a brain tumor image segmentation method that leverages deep residual learning with multi-modality image feature fusion. Our approach involves extracting and fusing distinct and complementary features from various modalities, fully exploiting the multi-modality information within a deep convolutional neural network to enhance the performance of brain tumor image segmentation. We evaluate the effectiveness of our proposed method using the BraTS2021 dataset and demonstrate that deep residual learning with multi-modality image feature fusion significantly improves segmentation accuracy. Our method achieves competitive segmentation results, with Dice values of 83.3, 89.07, and 91.44 for enhanced tumor, tumor core, and whole tumor, respectively. These findings highlight the potential of our method in improving brain tumor diagnosis and treatment through accurate segmentation of pathological regions in brain MRIs.
Collapse
Affiliation(s)
- Guying Zhang
- School of Mathematics, Physics and Information, Shaoxing University, Shaoxing, Zhejiang, 312000, China
| | - Jia Zhou
- Cancer Center, Gamma Knife Treatment Center, Zhejiang Provincial People's Hospital, Affiliated People's Hospital, Hangzhou Medical College, Hangzhou, Zhejiang, 310014, China
| | - Guanghua He
- School of Mathematics, Physics and Information, Shaoxing University, Shaoxing, Zhejiang, 312000, China
- Institute of Artificial Intelligence, Shaoxing University, Shaoxing, Zhejiang, 312000, China
| | - Hancan Zhu
- School of Mathematics, Physics and Information, Shaoxing University, Shaoxing, Zhejiang, 312000, China
- Institute of Artificial Intelligence, Shaoxing University, Shaoxing, Zhejiang, 312000, China
| |
Collapse
|
4
|
Chen X, Xie H, Li Z, Cheng G, Leng M, Wang FL. Information fusion and artificial intelligence for smart healthcare: a bibliometric study. Inf Process Manag 2023. [DOI: 10.1016/j.ipm.2022.103113] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
5
|
Cheng H, Wu K, Tian J, Ma K, Gu C, Guan X. Colon tissue image segmentation with MWSI-NET. Med Biol Eng Comput 2022; 60:727-737. [PMID: 35044614 DOI: 10.1007/s11517-022-02501-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2021] [Accepted: 12/28/2021] [Indexed: 11/26/2022]
Abstract
Developments in deep learning have resulted in computer-aided diagnosis for many types of cancer. Previously, pathologists manually performed the labeling work in the analysis of colon tissues, which is both time-consuming and labor-intensive. Results are easily affected by subjective conditions. Therefore, it is beneficial to identify the cancerous regions of colon cancer with the assistance of computer-aided technology. Pathological images are often difficult to process due to their irregularity, similarity between cancerous and non-cancerous tissues and large size. We propose a multi-scale perceptual field fusion structure based on a dilated convolutional network. Using this model, a structure of dilated convolution kernels with different aspect ratios is inserted, which can process cancerous regions of different sizes and generate larger receptive fields. Thus, the model can fuse detailed information at different scales for better semantic segmentation. Two different attention mechanisms are adopted to highlight the cancerous areas. A large, open-source dataset was used to verify improved efficacy when compared to previously disclosed methods.
Collapse
Affiliation(s)
- Hao Cheng
- Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Kaijie Wu
- Department of Automation, Shanghai Jiao Tong University, Shanghai, China.
| | - Jie Tian
- Department of Automation, Shanghai Jiao Tong University, Shanghai, China
| | - Kai Ma
- Department of Automation, Shanghai Jiao Tong University, Shanghai, China
| | - Chaocheng Gu
- Department of Automation, Shanghai Jiao Tong University, Shanghai, China
| | - Xinping Guan
- Department of Automation, Shanghai Jiao Tong University, Shanghai, China
| |
Collapse
|
6
|
Zhang Q, Du Y, Wei Z, Liu H, Yang X, Zhao D. Spine Medical Image Segmentation Based on Deep Learning. JOURNAL OF HEALTHCARE ENGINEERING 2021; 2021:1917946. [PMID: 34956558 PMCID: PMC8694979 DOI: 10.1155/2021/1917946] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/11/2021] [Revised: 09/13/2021] [Accepted: 10/25/2021] [Indexed: 11/18/2022]
Abstract
The aim was to further explore the clinical value of deep learning algorithm in the field of spinal medical image segmentation, and this study designed an improved U-shaped network (BN-U-Net) algorithm and applied it to the spinal MRI medical image segmentation of 22 research objects. The application value of this algorithm in MRI image processing was comprehensively evaluated by accuracy (Acc), sensitivity (Sen), specificity (Spe), and area under curve (AUC). The results show that the image processing time of fully convolutional network (FCN) algorithm and U-Net algorithm is greater than 6 min, while the processing time of BN-U-Net algorithm is only 5-10 s, and the processing time is significantly shortened (P < 0.05). The Acc, Sen, and Spe results of BN-U-Net segmentation algorithm were 94.54 ± 3.56%, 88.76 ± 2.67%, and 86.27 ± 6.23%, respectively, which were significantly improved compared with FCN algorithm and U-Net algorithm (P < 0.05). In summary, the improved U-Net network algorithm used in this study significantly improves the quality of spinal MRI images by automatic segmentation of MRI images, which is worthy of further promotion in the field of spinal medical image segmentation.
Collapse
Affiliation(s)
- Qingfeng Zhang
- Beijing University of Chinese Medicine Third Affiliated Hospital/Spin,Department, Beijing 100029, China
| | - Yun Du
- The Second School of Clinical Medicine, Beijing University of Chinese Medicine, Beijing 100078, China
| | - Zhiqiang Wei
- Dongfang Hospital Beijing University of Chinese Medicine/Orthopaedics, Beijing 100078, China
| | - Hengping Liu
- Beijing University of Chinese Medicine Third Affiliated Hospital/Spin,Department, Beijing 100029, China
| | - Xiaoxia Yang
- Beijing University of Chinese Medicine Third Affiliated Hospital/Spin,Department, Beijing 100029, China
| | - Dongfang Zhao
- Dongfang Hospital Beijing University of Chinese Medicine/Orthopaedics, Beijing 100078, China
| |
Collapse
|
7
|
Li Y, Zhou D, Liu TT, Shen XZ. Application of deep learning in image recognition and diagnosis of gastric cancer. Artif Intell Gastrointest Endosc 2021; 2:12-24. [DOI: 10.37126/aige.v2.i2.12] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/15/2021] [Revised: 03/30/2021] [Accepted: 04/20/2021] [Indexed: 02/06/2023] Open
Abstract
In recent years, artificial intelligence has been extensively applied in the diagnosis of gastric cancer based on medical imaging. In particular, using deep learning as one of the mainstream approaches in image processing has made remarkable progress. In this paper, we also provide a comprehensive literature survey using four electronic databases, PubMed, EMBASE, Web of Science, and Cochrane. The literature search is performed until November 2020. This article provides a summary of the existing algorithm of image recognition, reviews the available datasets used in gastric cancer diagnosis and the current trends in applications of deep learning theory in image recognition of gastric cancer. covers the theory of deep learning on endoscopic image recognition. We further evaluate the advantages and disadvantages of the current algorithms and summarize the characteristics of the existing image datasets, then combined with the latest progress in deep learning theory, and propose suggestions on the applications of optimization algorithms. Based on the existing research and application, the label, quantity, size, resolutions, and other aspects of the image dataset are also discussed. The future developments of this field are analyzed from two perspectives including algorithm optimization and data support, aiming to improve the diagnosis accuracy and reduce the risk of misdiagnosis.
Collapse
Affiliation(s)
- Yu Li
- Department of Gastroenterology and Hepatology, Zhongshan Hospital Affiliated to Fudan University, Shanghai 200032, China
| | - Da Zhou
- Department of Gastroenterology and Hepatology, Zhongshan Hospital Affiliated to Fudan University, Shanghai 200032, China
| | - Tao-Tao Liu
- Department of Gastroenterology and Hepatology, Zhongshan Hospital Affiliated to Fudan University, Shanghai 200032, China
| | - Xi-Zhong Shen
- Department of Gastroenterology and Hepatology, Zhongshan Hospital Affiliated to Fudan University, Shanghai 200032, China
| |
Collapse
|
8
|
Road Extraction from Very-High-Resolution Remote Sensing Images via a Nested SE-Deeplab Model. REMOTE SENSING 2020. [DOI: 10.3390/rs12182985] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
Abstract
Automatic road extraction from very-high-resolution remote sensing images has become a popular topic in a wide range of fields. Convolutional neural networks are often used for this purpose. However, many network models do not achieve satisfactory extraction results because of the elongated nature and varying sizes of roads in images. To improve the accuracy of road extraction, this paper proposes a deep learning model based on the structure of Deeplab v3. It incorporates squeeze-and-excitation (SE) module to apply weights to different feature channels, and performs multi-scale upsampling to preserve and fuse shallow and deep information. To solve the problems associated with unbalanced road samples in images, different loss functions and backbone network modules are tested in the model’s training process. Compared with cross entropy, dice loss can improve the performance of the model during training and prediction. The SE module is superior to ResNext and ResNet in improving the integrity of the extracted roads. Experimental results obtained using the Massachusetts Roads Dataset show that the proposed model (Nested SE-Deeplab) improves F1-Score by 2.4% and Intersection over Union by 2.0% compared with FC-DenseNet. The proposed model also achieves better segmentation accuracy in road extraction compared with other mainstream deep-learning models including Deeplab v3, SegNet, and UNet.
Collapse
|
9
|
Yu S, Chen M, Zhang E, Wu J, Yu H, Yang Z, Ma L, Gu X, Lu W. Robustness study of noisy annotation in deep learning based medical image segmentation. Phys Med Biol 2020; 65:175007. [PMID: 32503027 PMCID: PMC7567130 DOI: 10.1088/1361-6560/ab99e5] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
Abstract
Partly due to the use of exhaustive-annotated data, deep networks have achieved impressive performance on medical image segmentation. Medical imaging data paired with noisy annotation are, however, ubiquitous, but little is known about the effect of noisy annotation on deep learning based medical image segmentation. We studied the effect of noisy annotation in the context of mandible segmentation from CT images. First, 202 images of head and neck cancer patients were collected from our clinical database, where the organs-at-risk were annotated by one of twelve planning dosimetrists. The mandibles were roughly annotated as the planning avoiding structure. Then, mandible labels were checked and corrected by a head and neck specialist to get the reference standard. At last, by varying the ratios of noisy labels in the training set, deep networks were trained and tested for mandible segmentation. The trained models were further tested on other two public datasets. Experimental results indicated that the network trained with noisy labels had worse segmentation than that trained with reference standard, and in general, fewer noisy labels led to better performance. When using 20% or less noisy cases for training, no significant difference was found on the segmentation results between the models trained by noisy or reference annotation. Cross-dataset validation results verified that the models trained with noisy data achieved competitive performance to that trained with reference standard. This study suggests that the involved network is robust to noisy annotation to some extent in mandible segmentation from CT images. It also highlights the importance of labeling quality in deep learning. In the future work, extra attention should be paid to how to utilize a small number of reference standard samples to improve the performance of deep learning with noisy annotation.
Collapse
Affiliation(s)
- Shaode Yu
- Medical Artificial Intelligence and Automation Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, United States of America
| | | | | | | | | | | | | | | | | |
Collapse
|