1
|
Li Z, Xie S. EBC-Net: 3D semi-supervised segmentation of pancreas based on edge-biased consistency regularization in dual perturbation space. Med Phys 2024. [PMID: 39042373 DOI: 10.1002/mp.17323] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2023] [Revised: 05/17/2024] [Accepted: 07/08/2024] [Indexed: 07/24/2024] Open
Abstract
BACKGROUND Deep learning technology has made remarkable progress in pancreatic image segmentation tasks. However, annotating 3D medical images is time-consuming and requires expertise, and existing semi-supervised segmentation methods perform poorly in the segmentation task of organs with blurred edges in enhanced CT such as the pancreas. PURPOSE To address the challenges of limited labeled data and indistinct boundaries of regions of interest (ROI). METHODS We propose Edge-Biased Consistency Regularization (EBC-Net). 3D edge detection is employed to construct edge perturbations and integrate edge prior information into limited data, aiding the network in learning from unlabeled data. Additionally, due to the one-sidedness of a single perturbation space, we expand the dual-level perturbation space of both images and features to more efficiently focus the model's attention on the edges of the ROI. Finally, inspired by the clinical habits of doctors, we propose a 3D Anatomical Invariance Extraction Module and Anatomical Attention to capture anatomy-invariant features. RESULTS Extensive experiments have demonstrated that our method outperforms state-of-the-art methods in semi-supervised pancreas image segmentation. Moreover, it can better preserve the morphology of pancreatic organs and excel at edges region accuracy. CONCLUSIONS Incorporated with edge prior knowledge, our method mixes disturbances in dual-perturbation space, which shifts the network's attention to the fuzzy edge region using a few labeled samples. These ideas have been verified on the pancreas segmentation dataset.
Collapse
Affiliation(s)
- Zheng Li
- School of Communications and Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing, China
| | - Shipeng Xie
- School of Communications and Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing, China
| |
Collapse
|
2
|
Wang K, Jin K, Cheng Z, Liu X, Wang C, Guan X, Xu X, Ye J, Wang W, Wang S. Multi-scale consistent self-training network for semi-supervised orbital tumor segmentation. Med Phys 2024; 51:4859-4871. [PMID: 38277474 DOI: 10.1002/mp.16945] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2022] [Revised: 11/20/2023] [Accepted: 12/10/2023] [Indexed: 01/28/2024] Open
Abstract
PURPOSE Segmentation of orbital tumors in CT images is of great significance for orbital tumor diagnosis, which is one of the most prevalent diseases of the eye. However, the large variety of tumor sizes and shapes makes the segmentation task very challenging, especially when the available annotation data is limited. METHODS To this end, in this paper, we propose a multi-scale consistent self-training network (MSCINet) for semi-supervised orbital tumor segmentation. Specifically, we exploit the semantic-invariance features by enforcing the consistency between the predictions of different scales of the same image to make the model more robust to size variation. Moreover, we incorporate a new self-training strategy, which adopts iterative training with an uncertainty filtering mechanism to filter the pseudo-labels generated by the model, to eliminate the accumulation of pseudo-label error predictions and increase the generalization of the model. RESULTS For evaluation, we have built two datasets, the orbital tumor binary segmentation dataset (Orbtum-B) and the orbital multi-organ segmentation dataset (Orbtum-M). Experimental results on these two datasets show that our proposed method can both achieve state-of-the-art performance. In our datasets, there are a total of 55 patients containing 602 2D images. CONCLUSION In this paper, we develop a new semi-supervised segmentation method for orbital tumors, which is designed for the characteristics of orbital tumors and exhibits excellent performance compared to previous semi-supervised algorithms.
Collapse
Affiliation(s)
- Keyi Wang
- School of Mechanical, Electrical and Information Engineering at Shandong University, Weihai, China
| | - Kai Jin
- Department of Ophthalmology, the Second Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, China
| | - Zhiming Cheng
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
| | - Xindi Liu
- Department of Ophthalmology, the Second Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, China
| | - Changjun Wang
- Department of Ophthalmology, the Second Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, China
| | - Xiaojun Guan
- Department of Radiology, the Second Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, China
| | - Xiaojun Xu
- Department of Radiology, the Second Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, China
| | - Juan Ye
- Department of Ophthalmology, the Second Affiliated Hospital, Zhejiang University School of Medicine, Hangzhou, China
| | - Wenyu Wang
- School of Mechanical, Electrical and Information Engineering at Shandong University, Weihai, China
| | - Shuai Wang
- School of Mechanical, Electrical and Information Engineering at Shandong University, Weihai, China
- Suzhou Research Institute of Shandong University, Suzhou, China
| |
Collapse
|
3
|
Chen Y, Ma Y, Mei X, Zhang L, Fu Z, Ma J. Triple-task mutual consistency for semi-supervised 3D medical image segmentation. Comput Biol Med 2024; 175:108506. [PMID: 38688127 DOI: 10.1016/j.compbiomed.2024.108506] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Revised: 03/18/2024] [Accepted: 04/21/2024] [Indexed: 05/02/2024]
Abstract
Semi-supervised deep learning algorithm is an effective means of medical image segmentation. Among these methods, multi-task learning with consistency regularization has achieved outstanding results. However, most of the existing methods usually simply embed the Signed Distance Map (SDM) task into the network, which underestimates the potential ability of SDM in edge awareness and leads to excessive dependence between tasks. In this work, we propose a novel triple-task mutual consistency (TTMC) framework to enhance shape and edge awareness capabilities, and overcome the task dependence problem underestimated in previous work. Specifically, we innovatively construct the Signed Attention Map (SAM), a novel fusion image with attention mechanism, and use it as an auxiliary task for segmentation to enhance the edge awareness ability. Then we implement a triple-task deep network, which jointly predicts the voxel-wise classification map, the Signed Distance Map and the Signed Attention Map. In our proposed framework, an optimized differentiable transformation layer associates SDM with voxel-wise classification map and SAM prediction, while task-level consistency regularization utilizes unlabeled data in an unsupervised manner. Evaluated on the public Left Atrium dataset and NIH Pancreas dataset, our proposed framework achieves significant performance gains by effectively utilizing unlabeled data, outperforming recent state-of-the-art semi-supervised segmentation methods. Code is available at https://github.com/Saocent/TTMC.
Collapse
Affiliation(s)
- Yantao Chen
- School of Electronic Information, Wuhan University, Wuhan 430072, China.
| | - Yong Ma
- School of Electronic Information, Wuhan University, Wuhan 430072, China.
| | - Xiaoguang Mei
- School of Electronic Information, Wuhan University, Wuhan 430072, China.
| | - Lin Zhang
- Department of Urology, Zhongnan Hospital of Wuhan University, Wuhan, China.
| | - Zhigang Fu
- Department of Interventional Radiology, Yichang Central People's Hospital, First College of Clinical Medical Science, China Three Gorges University, Yichang, 443003, China.
| | - Jiayi Ma
- School of Electronic Information, Wuhan University, Wuhan 430072, China.
| |
Collapse
|
4
|
He Y, Kong J, Li J, Zheng C. Entropy and distance-guided super self-ensembling for optic disc and cup segmentation. BIOMEDICAL OPTICS EXPRESS 2024; 15:3975-3992. [PMID: 38867792 PMCID: PMC11166439 DOI: 10.1364/boe.521778] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Revised: 04/14/2024] [Accepted: 05/06/2024] [Indexed: 06/14/2024]
Abstract
Segmenting the optic disc (OD) and optic cup (OC) is crucial to accurately detect changes in glaucoma progression in the elderly. Recently, various convolutional neural networks have emerged to deal with OD and OC segmentation. Due to the domain shift problem, achieving high-accuracy segmentation of OD and OC from different domain datasets remains highly challenging. Unsupervised domain adaptation has taken extensive focus as a way to address this problem. In this work, we propose a novel unsupervised domain adaptation method, called entropy and distance-guided super self-ensembling (EDSS), to enhance the segmentation performance of OD and OC. EDSS is comprised of two self-ensembling models, and the Gaussian noise is added to the weights of the whole network. Firstly, we design a super self-ensembling (SSE) framework, which can combine two self-ensembling to learn more discriminative information about images. Secondly, we propose a novel exponential moving average with Gaussian noise (G-EMA) to enhance the robustness of the self-ensembling framework. Thirdly, we propose an effective multi-information fusion strategy (MFS) to guide and improve the domain adaptation process. We evaluate the proposed EDSS on two public fundus image datasets RIGA+ and REFUGE. Large amounts of experimental results demonstrate that the proposed EDSS outperforms state-of-the-art segmentation methods with unsupervised domain adaptation, e.g., the Dicemean score on three test sub-datasets of RIGA+ are 0.8442, 0.8772 and 0.9006, respectively, and the Dicemean score on the REFUGE dataset is 0.9154.
Collapse
Affiliation(s)
- Yanlin He
- College of Information Sciences and Technology, Northeast Normal University, Changchun 130117, China
| | - Jun Kong
- College of Information Sciences and Technology, Northeast Normal University, Changchun 130117, China
| | - Juan Li
- Jilin Engineering Normal University, Changchun 130052, China
- Business School, Northeast Normal University, Changchun 130117, China
| | - Caixia Zheng
- College of Information Sciences and Technology, Northeast Normal University, Changchun 130117, China
- Key Laboratory of Applied Statistics of MOE, Northeast Normal University, Changchun 130024, China
| |
Collapse
|
5
|
Zhou Q, Yu B, Xiao F, Ding M, Wang Z, Zhang X. Robust Semi-Supervised 3D Medical Image Segmentation With Diverse Joint-Task Learning and Decoupled Inter-Student Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2317-2331. [PMID: 38319753 DOI: 10.1109/tmi.2024.3362837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/08/2024]
Abstract
Semi-supervised segmentation is highly significant in 3D medical image segmentation. The typical solutions adopt a teacher-student dual-model architecture, and they constrain the two models' decision consistency on the same segmentation task. However, the scarcity of medical samples can lower the diversity of tasks, reducing the effectiveness of consistency constraint. The issue can further worsen as the weights of the models gradually become synchronized. In this work, we have proposed to construct diverse joint-tasks using masked image modelling for enhancing the reliability of the consistency constraint, and develop a novel architecture consisting of a single teacher but multiple students to enjoy the additional knowledge decoupled from the synchronized weights. Specifically, the teacher and student models 'see' varied randomly-masked versions of an input, and are trained to segment the same targets but reconstruct different missing regions concurrently. Such joint-task of segmentation and reconstruction can have the two learners capture related but complementary features to derive instructive knowledge when constraining their consistency. Moreover, two extra students join the original one to perform an inter-student learning. The three students share the same encoding but different decoding designs, and learn decoupled knowledge by constraining their mutual consistencies, preventing themselves from suboptimally converging to the biased predictions of the dictatorial teacher. Experimental on four medical datasets show that our approach performs better than six mainstream semi-supervised methods. Particularly, our approach achieves at least 0.61% and 0.36% higher Dice and Jaccard values, respectively, than the most competitive approach on our in-house dataset. The code will be released at https://github.com/zxmboshi/DDL.
Collapse
|
6
|
Su J, Luo Z, Lian S, Lin D, Li S. Mutual learning with reliable pseudo label for semi-supervised medical image segmentation. Med Image Anal 2024; 94:103111. [PMID: 38401271 DOI: 10.1016/j.media.2024.103111] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2023] [Revised: 01/15/2024] [Accepted: 02/20/2024] [Indexed: 02/26/2024]
Abstract
Semi-supervised learning has garnered significant interest as a method to alleviate the burden of data annotation. Recently, semi-supervised medical image segmentation has garnered significant interest that can alleviate the burden of densely annotated data. Substantial advancements have been achieved by integrating consistency-regularization and pseudo-labeling techniques. The quality of the pseudo-labels is crucial in this regard. Unreliable pseudo-labeling can result in the introduction of noise, leading the model to converge to suboptimal solutions. To address this issue, we propose learning from reliable pseudo-labels. In this paper, we tackle two critical questions in learning from reliable pseudo-labels: which pseudo-labels are reliable and how reliable are they? Specifically, we conduct a comparative analysis of two subnetworks to address both challenges. Initially, we compare the prediction confidence of the two subnetworks. A higher confidence score indicates a more reliable pseudo-label. Subsequently, we utilize intra-class similarity to assess the reliability of the pseudo-labels to address the second challenge. The greater the intra-class similarity of the predicted classes, the more reliable the pseudo-label. The subnetwork selectively incorporates knowledge imparted by the other subnetwork model, contingent on the reliability of the pseudo labels. By reducing the introduction of noise from unreliable pseudo-labels, we are able to improve the performance of segmentation. To demonstrate the superiority of our approach, we conducted an extensive set of experiments on three datasets: Left Atrium, Pancreas-CT and Brats-2019. The experimental results demonstrate that our approach achieves state-of-the-art performance. Code is available at: https://github.com/Jiawei0o0/mutual-learning-with-reliable-pseudo-labels.
Collapse
Affiliation(s)
- Jiawei Su
- The Department of Artificial Intelligence, Xiamen University, Fujian, China
| | - Zhiming Luo
- The Department of Artificial Intelligence, Xiamen University, Fujian, China.
| | - Sheng Lian
- The College of Computer and Data Science, Fuzhou University, Fujian, China
| | - Dazhen Lin
- The Department of Artificial Intelligence, Xiamen University, Fujian, China
| | - Shaozi Li
- The Department of Artificial Intelligence, Xiamen University, Fujian, China
| |
Collapse
|
7
|
Gai D, Huang Z, Min W, Geng Y, Wu H, Zhu M, Wang Q. SDMI-Net: Spatially Dependent Mutual Information Network for semi-supervised medical image segmentation. Comput Biol Med 2024; 174:108374. [PMID: 38582003 DOI: 10.1016/j.compbiomed.2024.108374] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2023] [Revised: 02/21/2024] [Accepted: 03/24/2024] [Indexed: 04/08/2024]
Abstract
Semi-supervised medical image segmentation strives to polish deep models with a small amount of labeled data and a large amount of unlabeled data. The efficiency of most semi-supervised medical image segmentation methods based on voxel-level consistency learning is affected by low-confidence voxels. In addition, voxel-level consistency learning fails to consider the spatial correlation between neighboring voxels. To encourage reliable voxel-level consistent learning, we propose a dual-teacher affine consistent uncertainty estimation method to filter out some voxels with high uncertainty. Moreover, we design the spatially dependent mutual information module, which enhances the spatial dependence between neighboring voxels by maximizing the mutual information between the local voxel blocks predicted from the dual-teacher models and the student model, enabling consistent learning at the block level. On two benchmark medical image segmentation datasets, including the Left Atrial Segmentation Challenge dataset and the BraTS-2019 dataset, our method achieves state-of-the-art performance in both quantitative and qualitative aspects.
Collapse
Affiliation(s)
- Di Gai
- School of Mathematics and Computer Science, Nanchang University, Nanchang, 330031, China; Institute of Metaverse, Nanchang University, Nanchang, 330031, China.
| | - Zheng Huang
- School of Mathematics and Computer Science, Nanchang University, Nanchang, 330031, China.
| | - Weidong Min
- School of Mathematics and Computer Science, Nanchang University, Nanchang, 330031, China; Institute of Metaverse, Nanchang University, Nanchang, 330031, China.
| | - Yuhan Geng
- School of Public Health, University of Michigan, Ann Arbor, MI, 48109, USA.
| | - Haifan Wu
- School of Mathematics and Computer Science, Nanchang University, Nanchang, 330031, China.
| | - Meng Zhu
- School of Mathematics and Computer Science, Nanchang University, Nanchang, 330031, China.
| | - Qi Wang
- School of Mathematics and Computer Science, Nanchang University, Nanchang, 330031, China; Institute of Metaverse, Nanchang University, Nanchang, 330031, China.
| |
Collapse
|
8
|
He A, Li T, Yan J, Wang K, Fu H. Bilateral Supervision Network for Semi-Supervised Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1715-1726. [PMID: 38153819 DOI: 10.1109/tmi.2023.3347689] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/30/2023]
Abstract
Massive high-quality annotated data is required by fully-supervised learning, which is difficult to obtain for image segmentation since the pixel-level annotation is expensive, especially for medical image segmentation tasks that need domain knowledge. As an alternative solution, semi-supervised learning (SSL) can effectively alleviate the dependence on the annotated samples by leveraging abundant unlabeled samples. Among the SSL methods, mean-teacher (MT) is the most popular one. However, in MT, teacher model's weights are completely determined by student model's weights, which will lead to the training bottleneck at the late training stages. Besides, only pixel-wise consistency is applied for unlabeled data, which ignores the category information and is susceptible to noise. In this paper, we propose a bilateral supervision network with bilateral exponential moving average (bilateral-EMA), named BSNet to overcome these issues. On the one hand, both the student and teacher models are trained on labeled data, and then their weights are updated with the bilateral-EMA, and thus the two models can learn from each other. On the other hand, pseudo labels are used to perform bilateral supervision for unlabeled data. Moreover, for enhancing the supervision, we adopt adversarial learning to enforce the network generate more reliable pseudo labels for unlabeled data. We conduct extensive experiments on three datasets to evaluate the proposed BSNet, and results show that BSNet can improve the semi-supervised segmentation performance by a large margin and surpass other state-of-the-art SSL methods.
Collapse
|
9
|
Du W, Huo Y, Zhou R, Sun Y, Tang S, Zhao X, Li Y, Li G. Consistency label-activated region generating network for weakly supervised medical image segmentation. Comput Biol Med 2024; 173:108380. [PMID: 38555701 DOI: 10.1016/j.compbiomed.2024.108380] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Revised: 03/04/2024] [Accepted: 03/24/2024] [Indexed: 04/02/2024]
Abstract
The current methods of auto-segmenting medical images are limited due to insufficient and ambiguous pathonmorphological labeling. In clinical practice, rough classification labels (such as disease or normal) are more commonly used than precise segmentation masks. However, there is still much to be explored regarding utilizing these weak clinical labels to accurately determine the lesion mask and guide medical image segmentation. In this paper, we proposed a weakly supervised medical image segmentation model to directly generate the lesion mask through a class activation map (CAM) guided cycle-consistency label-activated region transferring network. Cycle-consistency enforces that the mappings between the two domains should be reversible, which ensures that the original image can be reconstructed from the translated image. We developed a complementary branches fusion module to address the issue of blurry boundaries in CAM-guided segmentation. The complementary branch preserves the original semantic information of the non-lesion region and perfectly fuses the transferred feature of the lesion region with a complementary mask-constrained fake image generation process to clear the boundary of the lesion and non-lesion regions. This module allows the class transformation to focus solely on the label-activated region, resulting in more explicit segmentation. This model can accurately identify different region of medical images at the pixel-level while preserving the overall semantic structure semantion. It organizes disease labels and corresponding regions during image synthesis. Our method utilizes a joint discrimination strategy that significantly enhances the precision of the produced lesion mask. Extensive experiments of the proposed method on BraTs, ISIC and COVID-19 datasets demonstrate superior performance over existing state-of-the-art methods. The code and datasets are available at: https://github.com/mlcb-jlu/MedImgSeg.
Collapse
Affiliation(s)
- Wei Du
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Yongkang Huo
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Rixin Zhou
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Yu Sun
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Shiyi Tang
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Xuan Zhao
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Ying Li
- Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun, 130012, China.
| | - Gaoyang Li
- Translational Medical Center for Stem Cell Therapy and Institute for Regenerative Medicine, Shanghai East Hospital, Shanghai, 200092, China; Bioinformatics Department, School of Life Sciences and Technology, Tongji University, Shanghai, 200092, China.
| |
Collapse
|
10
|
Zhang K, Yang X, Cui Y, Zhao J, Li D. Imaging segmentation mechanism for rectal tumors using improved U-Net. BMC Med Imaging 2024; 24:95. [PMID: 38654162 DOI: 10.1186/s12880-024-01269-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 04/05/2024] [Indexed: 04/25/2024] Open
Abstract
OBJECTIVE In radiation therapy, cancerous region segmentation in magnetic resonance images (MRI) is a critical step. For rectal cancer, the automatic segmentation of rectal tumors from an MRI is a great challenge. There are two main shortcomings in existing deep learning-based methods that lead to incorrect segmentation: 1) there are many organs surrounding the rectum, and the shape of some organs is similar to that of rectal tumors; 2) high-level features extracted by conventional neural networks often do not contain enough high-resolution information. Therefore, an improved U-Net segmentation network based on attention mechanisms is proposed to replace the traditional U-Net network. METHODS The overall framework of the proposed method is based on traditional U-Net. A ResNeSt module was added to extract the overall features, and a shape module was added after the encoder layer. We then combined the outputs of the shape module and the decoder to obtain the results. Moreover, the model used different types of attention mechanisms, so that the network learned information to improve segmentation accuracy. RESULTS We validated the effectiveness of the proposed method using 3773 2D MRI datasets from 304 patients. The results showed that the proposed method achieved 0.987, 0.946, 0.897, and 0.899 for Dice, MPA, MioU, and FWIoU, respectively; these values are significantly better than those of other existing methods. CONCLUSION Due to time savings, the proposed method can help radiologists segment rectal tumors effectively and enable them to focus on patients whose cancerous regions are difficult for the network to segment. SIGNIFICANCE The proposed method can help doctors segment rectal tumors, thereby ensuring good diagnostic quality and accuracy.
Collapse
Affiliation(s)
- Kenan Zhang
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, 030024, China
- Key Laboratory of Big Data Fusion Analysis and Application of Shanxi Province, Taiyuan, 030024, China
| | - Xiaotang Yang
- Department of Radiology, Shanxi Cancer Hospital, Shanxi Medical University, Taiyuan, 030013, China.
| | - Yanfen Cui
- Department of Radiology, Shanxi Cancer Hospital, Shanxi Medical University, Taiyuan, 030013, China
| | - Jumin Zhao
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, 030024, China
- Key Laboratory of Big Data Fusion Analysis and Application of Shanxi Province, Taiyuan, 030024, China
- Intelligent Perception Engineering Technology Center of Shanxi, Taiyuan, 030024, China
| | - Dengao Li
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China.
- Key Laboratory of Big Data Fusion Analysis and Application of Shanxi Province, Taiyuan, 030024, China.
- Intelligent Perception Engineering Technology Center of Shanxi, Taiyuan, 030024, China.
| |
Collapse
|
11
|
Miao J, Zhou SP, Zhou GQ, Wang KN, Yang M, Zhou S, Chen Y. SC-SSL: Self-Correcting Collaborative and Contrastive Co-Training Model for Semi-Supervised Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1347-1364. [PMID: 37995173 DOI: 10.1109/tmi.2023.3336534] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/25/2023]
Abstract
Image segmentation achieves significant improvements with deep neural networks at the premise of a large scale of labeled training data, which is laborious to assure in medical image tasks. Recently, semi-supervised learning (SSL) has shown great potential in medical image segmentation. However, the influence of the learning target quality for unlabeled data is usually neglected in these SSL methods. Therefore, this study proposes a novel self-correcting co-training scheme to learn a better target that is more similar to ground-truth labels from collaborative network outputs. Our work has three-fold highlights. First, we advance the learning target generation as a learning task, improving the learning confidence for unannotated data with a self-correcting module. Second, we impose a structure constraint to encourage the shape similarity further between the improved learning target and the collaborative network outputs. Finally, we propose an innovative pixel-wise contrastive learning loss to boost the representation capacity under the guidance of an improved learning target, thus exploring unlabeled data more efficiently with the awareness of semantic context. We have extensively evaluated our method with the state-of-the-art semi-supervised approaches on four public-available datasets, including the ACDC dataset, M&Ms dataset, Pancreas-CT dataset, and Task_07 CT dataset. The experimental results with different labeled-data ratios show our proposed method's superiority over other existing methods, demonstrating its effectiveness in semi-supervised medical image segmentation.
Collapse
|
12
|
Zhao Y, Zhou X, Pan T, Gao S, Zhang W. Correspondence-based Generative Bayesian Deep Learning for semi-supervised volumetric medical image segmentation. Comput Med Imaging Graph 2024; 113:102352. [PMID: 38341947 DOI: 10.1016/j.compmedimag.2024.102352] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2023] [Revised: 02/03/2024] [Accepted: 02/03/2024] [Indexed: 02/13/2024]
Abstract
Automated medical image segmentation plays a crucial role in diverse clinical applications. The high annotation costs of fully-supervised medical segmentation methods have spurred a growing interest in semi-supervised methods. Existing semi-supervised medical segmentation methods train the teacher segmentation network using labeled data to establish pseudo labels for unlabeled data. The quality of these pseudo labels is constrained as these methods fail to effectively address the significant bias in the data distribution learned from the limited labeled data. To address these challenges, this paper introduces an innovative Correspondence-based Generative Bayesian Deep Learning (C-GBDL) model. Built upon the teacher-student architecture, we design a multi-scale semantic correspondence method to aid the teacher model in generating high-quality pseudo labels. Specifically, our teacher model, embedded with the multi-scale semantic correspondence, learns a better-generalized data distribution from input volumes by feature matching with the reference volumes. Additionally, a double uncertainty estimation schema is proposed to further rectify the noisy pseudo labels. The double uncertainty estimation takes the predictive entropy as the first uncertainty estimation and takes the structural similarity between the input volume and its corresponding reference volumes as the second uncertainty estimation. Four groups of comparative experiments conducted on two public medical datasets demonstrate the effectiveness and the superior performance of our proposed model. Our code is available on https://github.com/yumjoo/C-GBDL.
Collapse
Affiliation(s)
- Yuzhou Zhao
- Shanghai Key Lab of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China
| | - Xinyu Zhou
- Shanghai Key Lab of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China
| | - Tongxin Pan
- Shanghai Key Lab of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China
| | - Shuyong Gao
- Shanghai Key Lab of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China.
| | - Wenqiang Zhang
- Shanghai Key Lab of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China; Shanghai Engineering Research Center of AI & Robotics, Academy for Engineering and Technology, Fudan University, Shanghai, China.
| |
Collapse
|
13
|
Chen Y, Yang Z, Shen C, Wang Z, Zhang Z, Qin Y, Wei X, Lu J, Liu Y, Zhang Y. Evidence-based uncertainty-aware semi-supervised medical image segmentation. Comput Biol Med 2024; 170:108004. [PMID: 38277924 DOI: 10.1016/j.compbiomed.2024.108004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Revised: 12/21/2023] [Accepted: 01/13/2024] [Indexed: 01/28/2024]
Abstract
Semi-Supervised Learning (SSL) has demonstrated great potential to reduce the dependence on a large set of annotated data, which is challenging to collect in clinical practice. One of the most important SSL methods is to generate pseudo labels from the unlabeled data using a network model trained with labeled data, which will inevitably introduce false pseudo labels into the training process and potentially jeopardize performance. To address this issue, uncertainty-aware methods have emerged as a promising solution and have gained considerable attention recently. However, current uncertainty-aware methods usually face the dilemma of balancing the additional computational cost, uncertainty estimation accuracy, and theoretical basis in a unified training paradigm. To address this issue, we propose to integrate the Dempster-Shafer Theory of Evidence (DST) into SSL-based medical image segmentation, dubbed EVidential Inference Learning (EVIL). EVIL performs as a novel consistency regularization-based training paradigm, which enforces consistency on predictions perturbed by two networks with different parameters to enhance generalization Additionally, EVIL provides a theoretically assured solution for precise uncertainty quantification within a single forward pass. By discarding highly unreliable pseudo labels after uncertainty estimation, trustworthy pseudo labels can be generated and incorporated into subsequent model training. The experimental results demonstrate that the proposed approach performs competitively when benchmarked against several state-of-the-art methods on public datasets, i.e., ACDC, MM-WHS, and MonuSeg. The code can be found at https://github.com/CYYukio/EVidential-Inference-Learning.
Collapse
Affiliation(s)
- Yingyu Chen
- College of Computer Science, Sichuan University, China; The Key Laboratory of Data Protection and Intelligent Management, Ministry of Education, Sichuan University, China
| | - Ziyuan Yang
- College of Computer Science, Sichuan University, China
| | - Chenyu Shen
- College of Computer Science, Sichuan University, China
| | - Zhiwen Wang
- College of Computer Science, Sichuan University, China
| | | | - Yang Qin
- College of Computer Science, Sichuan University, China
| | - Xin Wei
- Department of Ophthalmology, West China Hospital, Sichuan University, China
| | - Jingfeng Lu
- School of Cyber Science and Engineering, Sichuan University, China
| | - Yan Liu
- College of Electrical Engineering, Sichuan University, China.
| | - Yi Zhang
- School of Cyber Science and Engineering, Sichuan University, China; The Key Laboratory of Data Protection and Intelligent Management, Ministry of Education, Sichuan University, China
| |
Collapse
|
14
|
Ren J, Che J, Gong P, Wang X, Li X, Li A, Xiao C. Cross comparison representation learning for semi-supervised segmentation of cellular nuclei in immunofluorescence staining. Comput Biol Med 2024; 171:108102. [PMID: 38350398 DOI: 10.1016/j.compbiomed.2024.108102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/07/2023] [Revised: 01/29/2024] [Accepted: 02/04/2024] [Indexed: 02/15/2024]
Abstract
The morphological analysis of cells from optical images is vital for interpreting brain function in disease states. Extracting comprehensive cell morphology from intricate backgrounds, common in neural and some medical images, poses a significant challenge. Due to the huge workload of manual recognition, automated neuron cell segmentation using deep learning algorithms with labeled data is integral to neural image analysis tools. To combat the high cost of acquiring labeled data, we propose a novel semi-supervised cell segmentation algorithm for immunofluorescence-stained cell image datasets (ISC), utilizing a mean-teacher semi-supervised learning framework. We include a "cross comparison representation learning block" to enhance the teacher-student model comparison on high-dimensional channels, thereby improving feature compactness and separability, which results in the extraction of higher-dimensional features from unlabeled data. We also suggest a new network, the Multi Pooling Layer Attention Dense Network (MPAD-Net), serving as the backbone of the student model to augment segmentation accuracy. Evaluations on the immunofluorescence staining datasets and the public CRAG dataset illustrate our method surpasses other top semi-supervised learning methods, achieving average Jaccard, Dice and Normalized Surface Dice (NSD) indicators of 83.22%, 90.95% and 81.90% with only 20% labeled data. The datasets and code are available on the website at https://github.com/Brainsmatics/CCRL.
Collapse
Affiliation(s)
- Jianran Ren
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China
| | - Jingyi Che
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China
| | - Peicong Gong
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China
| | - Xiaojun Wang
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China
| | - Xiangning Li
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China
| | - Anan Li
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China; Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Chi Xiao
- State Key Laboratory of Digital Medical Engineering, School of Biomedical Engineering, Hainan University, Sanya 572025, China; Key Laboratory of Biomedical Engineering of Hainan Province, One Health Institute, Hainan University, Sanya 572025, China.
| |
Collapse
|
15
|
Huang S, Luo J, Ou Y, Shen W, Pang Y, Nie X, Zhang G. Sd-net: a semi-supervised double-cooperative network for liver segmentation from computed tomography (CT) images. J Cancer Res Clin Oncol 2024; 150:79. [PMID: 38316678 PMCID: PMC10844439 DOI: 10.1007/s00432-023-05564-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Accepted: 12/08/2023] [Indexed: 02/07/2024]
Abstract
INTRODUCTION The automatic segmentation of the liver is a crucial step in obtaining quantitative biomarkers for accurate clinical diagnosis and computer-aided decision support systems. This task is challenging due to the frequent presence of noise and sampling artifacts in computerized tomography (CT) images, as well as the complex background, variable shapes, and blurry boundaries of the liver. Standard segmentation of medical images based on full-supervised convolutional networks demands accurate dense annotations. Such a learning framework is built on laborious manual annotation with strict requirements for expertise, leading to insufficient high-quality labels. METHODS To overcome such limitation and exploit massive weakly labeled data, we relaxed the rigid labeling requirement and developed a semi-supervised double-cooperative network (SD- Net). SD-Net is trained to segment the complete liver volume from preoperative abdominal CT images by using limited labeled datasets and large-scale unlabeled datasets. Specifically, to enrich the diversity of unsupervised information, we construct SD-Net consisting of two collaborative network models. Within the supervised training module, we introduce an adaptive mask refinement approach. First, each of the two network models predicts the labeled dataset, after which adaptive mask refinement of the difference predictions is implemented to obtain more accurate liver segmentation results. In the unsupervised training module, a dynamic pseudo-label generation strategy is proposed. First each of the two models predicts unlabeled data and the better prediction is considered as pseudo-labeling before training. RESULTS AND DISCUSSION Based on the experimental findings, the proposed method achieves a dice score exceeding 94%, indicating its high level of accuracy and its suitability for everyday clinical use.
Collapse
Affiliation(s)
- Shixin Huang
- School of Communication and Information Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
- Department of Scientific Research, The People's Hospital of Yubei District of Chongqing city, Chongqing, 401120, China
| | - Jiawei Luo
- West China Biomedical Big Data Center, West China Hospital, Chengdu, 610044, China
| | - Yangning Ou
- School of Electrical and Electronic Engineering, Shanghai Institute of Technology, Shanghai, 201418, China
| | - Wangjun Shen
- Chongqing Human Resources Development Service Center, Chongqing, 400065, China
| | - Yu Pang
- School of Optoelectronic Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
| | - Xixi Nie
- College of Computer Science and Technology, The Chongqing Key Laboratory of Image Cognition, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China.
| | - Guo Zhang
- School of Medical Information and Engineering, Southwest Medical University, Luzhou, 646000, China.
| |
Collapse
|
16
|
Jiao R, Zhang Y, Ding L, Xue B, Zhang J, Cai R, Jin C. Learning with limited annotations: A survey on deep semi-supervised learning for medical image segmentation. Comput Biol Med 2024; 169:107840. [PMID: 38157773 DOI: 10.1016/j.compbiomed.2023.107840] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Revised: 10/30/2023] [Accepted: 12/07/2023] [Indexed: 01/03/2024]
Abstract
Medical image segmentation is a fundamental and critical step in many image-guided clinical approaches. Recent success of deep learning-based segmentation methods usually relies on a large amount of labeled data, which is particularly difficult and costly to obtain, especially in the medical imaging domain where only experts can provide reliable and accurate annotations. Semi-supervised learning has emerged as an appealing strategy and been widely applied to medical image segmentation tasks to train deep models with limited annotations. In this paper, we present a comprehensive review of recently proposed semi-supervised learning methods for medical image segmentation and summarize both the technical novelties and empirical results. Furthermore, we analyze and discuss the limitations and several unsolved problems of existing approaches. We hope this review can inspire the research community to explore solutions to this challenge and further advance the field of medical image segmentation.
Collapse
Affiliation(s)
- Rushi Jiao
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China; School of Engineering Medicine, Beihang University, Beijing, 100191, China; Shanghai Artificial Intelligence Laboratory, Shanghai, 200232, China.
| | - Yichi Zhang
- School of Data Science, Fudan University, Shanghai, 200433, China; Artificial Intelligence Innovation and Incubation Institute, Fudan University, Shanghai, 200433, China.
| | - Le Ding
- School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China.
| | - Bingsen Xue
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China; Shanghai Artificial Intelligence Laboratory, Shanghai, 200232, China.
| | - Jicong Zhang
- School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Hefei Innovation Research Institute, Beihang University, Hefei, 230012, China.
| | - Rong Cai
- School of Engineering Medicine, Beihang University, Beijing, 100191, China; Key Laboratory for Biomechanics and Mechanobiology of Ministry of Education, Beihang University, Beijing, 100191, China.
| | - Cheng Jin
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China; Shanghai Artificial Intelligence Laboratory, Shanghai, 200232, China; Beijing Anding Hospital, Capital Medical University, Beijing, 100088, China.
| |
Collapse
|
17
|
Tian C, Zhang Z, Gao X, Zhou H, Ran R, Jiao Z. An Implicit-Explicit Prototypical Alignment Framework for Semi-Supervised Medical Image Segmentation. IEEE J Biomed Health Inform 2024; 28:929-940. [PMID: 37930923 DOI: 10.1109/jbhi.2023.3330667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2023]
Abstract
Semi-supervised learning methods have been explored to mitigate the scarcity of pixel-level annotation in medical image segmentation tasks. Consistency learning, serving as a mainstream method in semi-supervised training, suffers from low efficiency and poor stability due to inaccurate supervision and insufficient feature representation. Prototypical learning is one potential and plausible way to handle this problem due to the nature of feature aggregation in prototype calculation. However, the previous works have not fully studied how to enhance the supervision quality and feature representation using prototypical learning under the semi-supervised condition. To address this issue, we propose an implicit-explicit alignment (IEPAlign) framework to foster semi-supervised consistency training. In specific, we develop an implicit prototype alignment method based on dynamic multiple prototypes on-the-fly. And then, we design a multiple prediction voting strategy for reliable unlabeled mask generation and prototype calculation to improve the supervision quality. Afterward, to boost the intra-class consistency and inter-class separability of pixel-wise features in semi-supervised segmentation, we construct a region-aware hierarchical prototype alignment, which transmits information from labeled to unlabeled and from certain regions to uncertain regions. We evaluate IEPAlign on three medical image segmentation tasks. The extensive experimental results demonstrate that the proposed method outperforms other popular semi-supervised segmentation methods and achieves comparable performance with fully-supervised training methods.
Collapse
|
18
|
Adiga V S, Dolz J, Lombaert H. Anatomically-aware uncertainty for semi-supervised image segmentation. Med Image Anal 2024; 91:103011. [PMID: 37924752 DOI: 10.1016/j.media.2023.103011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2022] [Revised: 08/11/2023] [Accepted: 10/18/2023] [Indexed: 11/06/2023]
Abstract
Semi-supervised learning relaxes the need of large pixel-wise labeled datasets for image segmentation by leveraging unlabeled data. A prominent way to exploit unlabeled data is to regularize model predictions. Since the predictions of unlabeled data can be unreliable, uncertainty-aware schemes are typically employed to gradually learn from meaningful and reliable predictions. Uncertainty estimation methods, however, rely on multiple inferences from the model predictions that must be computed for each training step, which is computationally expensive. Moreover, these uncertainty maps capture pixel-wise disparities and do not consider global information. This work proposes a novel method to estimate segmentation uncertainty by leveraging global information from the segmentation masks. More precisely, an anatomically-aware representation is first learnt to model the available segmentation masks. The learnt representation thereupon maps the prediction of a new segmentation into an anatomically-plausible segmentation. The deviation from the plausible segmentation aids in estimating the underlying pixel-level uncertainty in order to further guide the segmentation network. The proposed method consequently estimates the uncertainty using a single inference from our representation, thereby reducing the total computation. We evaluate our method on two publicly available segmentation datasets of left atria in cardiac MRIs and of multiple organs in abdominal CTs. Our anatomically-aware method improves the segmentation accuracy over the state-of-the-art semi-supervised methods in terms of two commonly used evaluation metrics.
Collapse
Affiliation(s)
- Sukesh Adiga V
- Computer and Software Engineering Department, ETS Montreal, 1100 Notre Dame St. W., Montreal QC, H3C 1K3, Canada.
| | - Jose Dolz
- Computer and Software Engineering Department, ETS Montreal, 1100 Notre Dame St. W., Montreal QC, H3C 1K3, Canada
| | - Herve Lombaert
- Computer and Software Engineering Department, ETS Montreal, 1100 Notre Dame St. W., Montreal QC, H3C 1K3, Canada
| |
Collapse
|
19
|
Hu J, Qiu L, Wang H, Zhang J. Semi-supervised point consistency network for retinal artery/vein classification. Comput Biol Med 2024; 168:107633. [PMID: 37992471 DOI: 10.1016/j.compbiomed.2023.107633] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2023] [Revised: 10/02/2023] [Accepted: 10/23/2023] [Indexed: 11/24/2023]
Abstract
Recent deep learning methods with convolutional neural networks (CNNs) have boosted advance prosperity of medical image analysis and expedited the automatic retinal artery/vein (A/V) classification. However, it is challenging for these CNN-based approaches in two aspects: (1) specific tubular structures and subtle variations in appearance, contrast, and geometry, which tend to be ignored in CNNs with network layer increasing; (2) limited well-labeled data for supervised segmentation of retinal vessels, which may hinder the effectiveness of deep learning methods. To address these issues, we propose a novel semi-supervised point consistency network (SPC-Net) for retinal A/V classification. SPC-Net consists of an A/V classification (AVC) module and a multi-class point consistency (MPC) module. The AVC module adopts an encoder-decoder segmentation network to generate the prediction probability map of A/V for supervised learning. The MPC module introduces point set representations to adaptively generate point set classification maps of the arteriovenous skeleton, which enjoys its prediction flexibility and consistency (i.e. point consistency) to effectively alleviate arteriovenous confusion. In addition, we propose a consistency regularization between the predicted A/V classification probability maps and point set representations maps for unlabeled data to explore the inherent segmentation perturbation of the point consistency, reducing the need for annotated data. We validate our method on two typical public datasets (DRIVE, HRF) and a private dataset (TR280) with different resolutions. Extensive qualitative and quantitative experimental results demonstrate the effectiveness of our proposed method for supervised and semi-supervised learning.
Collapse
Affiliation(s)
- Jingfei Hu
- School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Hefei Innovation Research Institute, Beihang University, Hefei, 230012, Anhui, China
| | - Linwei Qiu
- School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Hefei Innovation Research Institute, Beihang University, Hefei, 230012, Anhui, China
| | - Hua Wang
- School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Hefei Innovation Research Institute, Beihang University, Hefei, 230012, Anhui, China
| | - Jicong Zhang
- School of Biological Science and Medical Engineering, Beihang University, Beijing, 100191, China; Hefei Innovation Research Institute, Beihang University, Hefei, 230012, Anhui, China; Beijing Advanced Innovation Centre for Big Data-Based Precision Medicine, Beihang University, Beijing, 100083, China.
| |
Collapse
|
20
|
Yang S, Liang Y, Wu S, Sun P, Chen Z. SADSNet: A robust 3D synchronous segmentation network for liver and liver tumors based on spatial attention mechanism and deep supervision. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:707-723. [PMID: 38552134 DOI: 10.3233/xst-230312] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/01/2024]
Abstract
Highlights • Introduce a data augmentation strategy to expand the required different morphological data during the training and learning phase, and improve the algorithm's feature learning ability for complex and diverse tumor morphology CT images.• Design attention mechanisms for encoding and decoding paths to extract fine pixel level features, improve feature extraction capabilities, and achieve efficient spatial channel feature fusion.• The deep supervision layer is used to correct and decode the final image data to provide high accuracy of results.• The effectiveness of this method has been affirmed through validation on the LITS, 3DIRCADb, and SLIVER datasets. BACKGROUND Accurately extracting liver and liver tumors from medical images is an important step in lesion localization and diagnosis, surgical planning, and postoperative monitoring. However, the limited number of radiation therapists and a great number of images make this work time-consuming. OBJECTIVE This study designs a spatial attention deep supervised network (SADSNet) for simultaneous automatic segmentation of liver and tumors. METHOD Firstly, self-designed spatial attention modules are introduced at each layer of the encoder and decoder to extract image features at different scales and resolutions, helping the model better capture liver tumors and fine structures. The designed spatial attention module is implemented through two gate signals related to liver and tumors, as well as changing the size of convolutional kernels; Secondly, deep supervision is added behind the three layers of the decoder to assist the backbone network in feature learning and improve gradient propagation, enhancing robustness. RESULTS The method was testing on LITS, 3DIRCADb, and SLIVER datasets. For the liver, it obtained dice similarity coefficients of 97.03%, 96.11%, and 97.40%, surface dice of 81.98%, 82.53%, and 86.29%, 95% hausdorff distances of 8.96 mm, 8.26 mm, and 3.79 mm, and average surface distances of 1.54 mm, 1.19 mm, and 0.81 mm. Additionally, it also achieved precise tumor segmentation, which with dice scores of 87.81% and 87.50%, surface dice of 89.63% and 84.26%, 95% hausdorff distance of 12.96 mm and 16.55 mm, and average surface distances of 1.11 mm and 3.04 mm on LITS and 3DIRCADb, respectively. CONCLUSION The experimental results show that the proposed method is effective and superior to some other methods. Therefore, this method can provide technical support for liver and liver tumor segmentation in clinical practice.
Collapse
Affiliation(s)
- Sijing Yang
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
| | - Yongbo Liang
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
| | - Shang Wu
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
| | - Peng Sun
- School of Electronic Engineering and Automation, Guilin University of Electronic Technology, Guilin, China
| | - Zhencheng Chen
- School of Life and Environmental Science, Guilin University of Electronic Technology, Guilin, China
- School of Electronic Engineering and Automation, Guilin University of Electronic Technology, Guilin, China
- Guangxi Colleges and Universities Key Laboratory of Biomedical Sensors and Intelligent Instruments, Guilin, China
- Guangxi Engineering Technology Research Center of Human Physiological Information Noninvasive Detection, Guilin, China
| |
Collapse
|
21
|
Osman YBM, Li C, Huang W, Wang S. Sparse annotation learning for dense volumetric MR image segmentation with uncertainty estimation. Phys Med Biol 2023; 69:015009. [PMID: 38035374 DOI: 10.1088/1361-6560/ad111b] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Accepted: 11/30/2023] [Indexed: 12/02/2023]
Abstract
Objective.Training neural networks for pixel-wise or voxel-wise image segmentation is a challenging task that requires a considerable amount of training samples with highly accurate and densely delineated ground truth maps. This challenge becomes especially prominent in the medical imaging domain, where obtaining reliable annotations for training samples is a difficult, time-consuming, and expert-dependent process. Therefore, developing models that can perform well under the conditions of limited annotated training data is desirable.Approach.In this study, we propose an innovative framework called the extremely sparse annotation neural network (ESA-Net) that learns with only the single central slice label for 3D volumetric segmentation which explores both intra-slice pixel dependencies and inter-slice image correlations with uncertainty estimation. Specifically, ESA-Net consists of four specially designed distinct components: (1) an intra-slice pixel dependency-guided pseudo-label generation module that exploits uncertainty in network predictions while generating pseudo-labels for unlabeled slices with temporal ensembling; (2) an inter-slice image correlation-constrained pseudo-label propagation module which propagates labels from the labeled central slice to unlabeled slices by self-supervised registration with rotation ensembling; (3) a pseudo-label fusion module that fuses the two sets of generated pseudo-labels with voxel-wise uncertainty guidance; and (4) a final segmentation network optimization module to make final predictions with scoring-based label quantification.Main results.Extensive experimental validations have been performed on two popular yet challenging magnetic resonance image segmentation tasks and compared to five state-of-the-art methods.Significance.Results demonstrate that our proposed ESA-Net can consistently achieve better segmentation performances even under the extremely sparse annotation setting, highlighting its effectiveness in exploiting information from unlabeled data.
Collapse
Affiliation(s)
- Yousuf Babiker M Osman
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- University of Chinese Academy of Sciences, Beijing 100049, People's Republic of China
| | - Cheng Li
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application, Guangzhou 510080, People's Republic of China
| | - Weijian Huang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- University of Chinese Academy of Sciences, Beijing 100049, People's Republic of China
- Peng Cheng Laboratory, Shenzhen 518066, People's Republic of China
| | - Shanshan Wang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application, Guangzhou 510080, People's Republic of China
- Peng Cheng Laboratory, Shenzhen 518066, People's Republic of China
| |
Collapse
|
22
|
Chen F, Chen L, Kong W, Zhang W, Zheng P, Sun L, Zhang D, Liao H. Deep Semi-Supervised Ultrasound Image Segmentation by Using a Shadow Aware Network With Boundary Refinement. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3779-3793. [PMID: 37695964 DOI: 10.1109/tmi.2023.3309249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/13/2023]
Abstract
Accurate ultrasound (US) image segmentation is crucial for the screening and diagnosis of diseases. However, it faces two significant challenges: 1) pixel-level annotation is a time-consuming and laborious process; 2) the presence of shadow artifacts leads to missing anatomy and ambiguous boundaries, which negatively impact reliable segmentation results. To address these challenges, we propose a novel semi-supervised shadow aware network with boundary refinement (SABR-Net). Specifically, we add shadow imitation regions to the original US, and design shadow-masked transformer blocks to perceive missing anatomy of shadow regions. Shadow-masked transformer block contains an adaptive shadow attention mechanism that introduces an adaptive mask, which is updated automatically to promote the network training. Additionally, we utilize unlabeled US images to train a missing structure inpainting path with shadow-masked transformer, which further facilitates semi-supervised segmentation. Experiments on two public US datasets demonstrate the superior performance of the SABR-Net over other state-of-the-art semi-supervised segmentation methods. In addition, experiments on a private breast US dataset prove that our method has a good generalization to clinical small-scale US datasets.
Collapse
|
23
|
Xu H, Xie H, Tan Q, Zhang Y. Meta semi-supervised medical image segmentation with label hierarchy. Health Inf Sci Syst 2023; 11:26. [PMID: 37325196 PMCID: PMC10267083 DOI: 10.1007/s13755-023-00222-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Accepted: 04/05/2023] [Indexed: 06/17/2023] Open
Abstract
Semi-supervised learning (SSL) has attracted increasing attention in medical image segmentation, where the mainstream usually explores perturbation-based consistency as a regularization to leverage unlabelled data. However, unlike directly optimizing segmentation task objectives, consistency regularization is a compromise by incorporating invariance towards perturbations, and inevitably suffers from noise in self-predicted targets. The above issues result in a knowledge gap between supervised guidance and unsupervised regularization. To bridge the knowledge gap, this work proposes a meta-based semi-supervised segmentation framework with the exploitation of label hierarchy. Two main prominent components named Divide and Generalize, and Label Hierarchy, are built in this work. Concretely, rather than merging all knowledge indiscriminately, we dynamically divide consistency regularization from supervised guidance as different domains. Then, a domain generalization technique is introduced with a meta-based optimization objective which ensures the update on supervised guidance should generalize to the consistency regularization, thereby bridging the knowledge gap. Furthermore, to alleviate the negative impact of noise in self-predicted targets, we propose to distill the noisy pixel-level consistency by exploiting label hierarchy and extracting hierarchical consistencies. Comprehensive experiments on two public medical segmentation benchmarks demonstrate the superiority of our framework to other semi-supervised segmentation methods, with new state-of-the-art results.
Collapse
Affiliation(s)
- Hai Xu
- School of Information Science and Technology, University of Science and Technology of China, Hefei, 230026 Anhui China
| | - Hongtao Xie
- School of Information Science and Technology, University of Science and Technology of China, Hefei, 230026 Anhui China
| | - Qingfeng Tan
- Cyberspace Institution of Advanced Technology, Guangzhou University, Guangzhou, 511442 Guangdong China
| | - Yongdong Zhang
- School of Information Science and Technology, University of Science and Technology of China, Hefei, 230026 Anhui China
| |
Collapse
|
24
|
Zhang J, Cui Z, Zhou L, Sun Y, Li Z, Liu Z, Shen D. Breast Fibroglandular Tissue Segmentation for Automated BPE Quantification With Iterative Cycle-Consistent Semi-Supervised Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3944-3955. [PMID: 37756174 DOI: 10.1109/tmi.2023.3319646] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/29/2023]
Abstract
Background Parenchymal Enhancement (BPE) quantification in Dynamic Contrast-Enhanced Magnetic Resonance Imaging (DCE-MRI) plays a pivotal role in clinical breast cancer diagnosis and prognosis. However, the emerging deep learning-based breast fibroglandular tissue segmentation, a crucial step in automated BPE quantification, often suffers from limited training samples with accurate annotations. To address this challenge, we propose a novel iterative cycle-consistent semi-supervised framework to leverage segmentation performance by using a large amount of paired pre-/post-contrast images without annotations. Specifically, we design the reconstruction network, cascaded with the segmentation network, to learn a mapping from the pre-contrast images and segmentation predictions to the post-contrast images. Thus, we can implicitly use the reconstruction task to explore the inter-relationship between these two-phase images, which in return guides the segmentation task. Moreover, the reconstructed post-contrast images across multiple auto-context modeling-based iterations can be viewed as new augmentations, facilitating cycle-consistent constraints across each segmentation output. Extensive experiments on two datasets with various data distributions show great segmentation and BPE quantification accuracy compared with other state-of-the-art semi-supervised methods. Importantly, our method achieves 11.80 times of quantification accuracy improvement along with 10 times faster, compared with clinical physicians, demonstrating its potential for automated BPE quantification. The code is available at https://github.com/ZhangJD-ong/Iterative-Cycle-consistent-Semi-supervised-Learning-for-fibroglandular-tissue-segmentation.
Collapse
|
25
|
Zhang X. Image denoising and segmentation model construction based on IWOA-PCNN. Sci Rep 2023; 13:19848. [PMID: 37963960 PMCID: PMC10645996 DOI: 10.1038/s41598-023-47089-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Accepted: 11/08/2023] [Indexed: 11/16/2023] Open
Abstract
The research suggests a method to improve the present pulse coupled neural network (PCNN), which has a complex structure and unsatisfactory performance in image denoising and image segmentation. Then, a multi strategy collaborative improvement whale optimization algorithm (WOA) is proposed, and an improved whale optimization algorithm (IWOA) is constructed. IWOA is used to find the optimal parameter values of PCNN to optimize PCNN. By combining the aforementioned components, the IWOA-PCNN model had the best image denoising performance, and the produced images were crisper and preserve more information. IWOA-PCNN processed pictures have an average PSNR of 35.87 and an average MSE of 0.24. The average processing time for photos with noise is typically 24.80 s, which is 7.30 s and 7.76 s faster than the WTGAN and IGA-NLM models, respectively. Additionally, the average NU value measures 0.947, and the average D value exceeds 1000. The aforementioned findings demonstrate that the suggested method can successfully enhance the PCNN, improving its capability for image denoising and image segmentation. This can, in part, encourage the use and advancement of the PCNN.
Collapse
Affiliation(s)
- Xiaojun Zhang
- College of Software Technology, Henan Finance University, Zhengzhou, 450000, China.
| |
Collapse
|
26
|
Wu H, Li X, Lin Y, Cheng KT. Compete to Win: Enhancing Pseudo Labels for Barely-Supervised Medical Image Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3244-3255. [PMID: 37220039 DOI: 10.1109/tmi.2023.3279110] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
This study investigates barely-supervised medical image segmentation where only few labeled data, i.e., single-digit cases are available. We observe the key limitation of the existing state-of-the-art semi-supervised solution cross pseudo supervision is the unsatisfactory precision of foreground classes, leading to a degenerated result under barely-supervised learning. In this paper, we propose a novel Compete-to-Win method (ComWin) to enhance the pseudo label quality. In contrast to directly using one model's predictions as pseudo labels, our key idea is that high-quality pseudo labels should be generated by comparing multiple confidence maps produced by different networks to select the most confident one (a compete-to-win strategy). To further refine pseudo labels at near-boundary areas, an enhanced version of ComWin, namely, ComWin + , is proposed by integrating a boundary-aware enhancement module. Experiments show that our method can achieve the best performance on three public medical image datasets for cardiac structure segmentation, pancreas segmentation and colon tumor segmentation, respectively. The source code is now available at https://github.com/Huiimin5/comwin.
Collapse
|
27
|
Li M, Zhang W, Yang R, Xu J, Zhao H, Li H. Semi-supervised peripapillary atrophy segmentation with shape constraint. Comput Biol Med 2023; 166:107464. [PMID: 37734355 DOI: 10.1016/j.compbiomed.2023.107464] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2023] [Revised: 08/14/2023] [Accepted: 09/04/2023] [Indexed: 09/23/2023]
Abstract
Peripapillary atrophy (PPA) is a clinical abnormality related to many eye diseases, such as myopia and glaucoma. The shape and area of PPA are essential indicators of disease progression. PPA segmentation is a challenging task due to blurry edge and limited labeled data. In this paper, we propose a novel semi-supervised PPA segmentation method enhanced by prior knowledge. In order to learn shape information in the network, a novel shape constraint module is proposed to restrict the PPA appearance based on active shape model. To further leverage large amount of unlabeled data, a Siamese-like model updated by exponential moving average is introduced to provide pseudo labels. The pseudo labels are further refined by region connectivity correction. Extensive experiments on a clinical dataset demonstrate that our proposed PPA segmentation method provides good qualitative and quantitative performance.
Collapse
Affiliation(s)
- Mengxuan Li
- Beijing Institute of Technology, Beijing 100081, China
| | - Weihang Zhang
- Beijing Institute of Technology, Beijing 100081, China
| | - Ruixiao Yang
- Beijing Institute of Technology, Beijing 100081, China
| | - Jie Xu
- Beijing Institute of Ophthalmology, Beijing Tongren Hospital, Capital Medical University, Beijing 100005, China
| | - He Zhao
- Beijing Institute of Technology, Beijing 100081, China.
| | - Huiqi Li
- Beijing Institute of Technology, Beijing 100081, China.
| |
Collapse
|
28
|
Lou A, Tawfik K, Yao X, Liu Z, Noble J. Min-Max Similarity: A Contrastive Semi-Supervised Deep Learning Network for Surgical Tools Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2832-2841. [PMID: 37037256 PMCID: PMC10597739 DOI: 10.1109/tmi.2023.3266137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
A common problem with segmentation of medical images using neural networks is the difficulty to obtain a significant number of pixel-level annotated data for training. To address this issue, we proposed a semi-supervised segmentation network based on contrastive learning. In contrast to the previous state-of-the-art, we introduce Min-Max Similarity (MMS), a contrastive learning form of dual-view training by employing classifiers and projectors to build all-negative, and positive and negative feature pairs, respectively, to formulate the learning as solving a MMS problem. The all-negative pairs are used to supervise the networks learning from different views and to capture general features, and the consistency of unlabeled predictions is measured by pixel-wise contrastive loss between positive and negative pairs. To quantitatively and qualitatively evaluate our proposed method, we test it on four public endoscopy surgical tool segmentation datasets and one cochlear implant surgery dataset, which we manually annotated. Results indicate that our proposed method consistently outperforms state-of-the-art semi-supervised and fully supervised segmentation algorithms. And our semi-supervised segmentation algorithm can successfully recognize unknown surgical tools and provide good predictions. Also, our MMS approach could achieve inference speeds of about 40 frames per second (fps) and is suitable to deal with the real-time video segmentation.
Collapse
|
29
|
Qiu L, Cheng J, Gao H, Xiong W, Ren H. Federated Semi-Supervised Learning for Medical Image Segmentation via Pseudo-Label Denoising. IEEE J Biomed Health Inform 2023; 27:4672-4683. [PMID: 37155394 DOI: 10.1109/jbhi.2023.3274498] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Distributed big data and digital healthcare technologies have great potential to promote medical services, but challenges arise when it comes to learning predictive model from diverse and complex e-health datasets. Federated Learning (FL), as a collaborative machine learning technique, aims to address the challenges by learning a joint predictive model across multi-site clients, especially for distributed medical institutions or hospitals. However, most existing FL methods assume that clients possess fully labeled data for training, which is often not the case in e-health datasets due to high labeling costs or expertise requirement. Therefore, this work proposes a novel and feasible approach to learn a Federated Semi-Supervised Learning (FSSL) model from distributed medical image domains, where a federated pseudo-labeling strategy for unlabeled clients is developed based on the embedded knowledge learned from labeled clients. This greatly mitigates the annotation deficiency at unlabeled clients and leads to a cost-effective and efficient medical image analysis tool. We demonstrated the effectiveness of our method by achieving significant improvements compared to the state-of-the-art in both fundus image and prostate MRI segmentation tasks, resulting in the highest Dice scores of 89.23% and 91.95% respectively even with only a few labeled clients participating in model training. This reveals the superiority of our method for practical deployment, ultimately facilitating the wider use of FL in healthcare and leading to better patient outcomes.
Collapse
|
30
|
Li W, Lu W, Chu J, Tian Q, Fan F. Confidence-guided mask learning for semi-supervised medical image segmentation. Comput Biol Med 2023; 165:107398. [PMID: 37688993 DOI: 10.1016/j.compbiomed.2023.107398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2023] [Revised: 07/29/2023] [Accepted: 08/26/2023] [Indexed: 09/11/2023]
Abstract
Semi-supervised learning aims to train a high-performance model with a minority of labeled data and a majority of unlabeled data. Existing methods mostly adopt the mechanism of prediction task to obtain precise segmentation maps with the constraints of consistency or pseudo-labels, whereas the mechanism usually fails to overcome confirmation bias. To address this issue, in this paper, we propose a novel Confidence-Guided Mask Learning (CGML) for semi-supervised medical image segmentation. Specifically, on the basis of the prediction task, we further introduce an auxiliary generation task with mask learning, which intends to reconstruct the masked images for extremely facilitating the model capability of learning feature representations. Moreover, a confidence-guided masking strategy is developed to enhance model discrimination in uncertain regions. Besides, we introduce a triple-consistency loss to enforce a consistent prediction of the masked unlabeled image, original unlabeled image and reconstructed unlabeled image for generating more reliable results. Extensive experiments on two datasets demonstrate that our proposed method achieves remarkable performance.
Collapse
Affiliation(s)
- Wenxue Li
- The School of Future Technology, Tianjin University, Tianjin, 300072, China
| | - Wei Lu
- The School of Electrical and Information Engineering, Tianjin University, Tianjin, 300072, China
| | - Jinghui Chu
- The School of Electrical and Information Engineering, Tianjin University, Tianjin, 300072, China
| | - Qi Tian
- Tianjin Children's Hospital, Tianjin, 300204, China
| | - Fugui Fan
- The School of Electrical and Information Engineering, Tianjin University, Tianjin, 300072, China.
| |
Collapse
|
31
|
Irfan M, Malik KM, Ahmad J, Malik G. StrokeNet: An automated approach for segmentation and rupture risk prediction of intracranial aneurysm. Comput Med Imaging Graph 2023; 108:102271. [PMID: 37556901 DOI: 10.1016/j.compmedimag.2023.102271] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Revised: 06/19/2023] [Accepted: 07/05/2023] [Indexed: 08/11/2023]
Abstract
Intracranial Aneurysms (IA) present a complex challenge for neurosurgeons as the risks associated with surgical intervention, such as Subarachnoid Hemorrhage (SAH) mortality and morbidity, may outweigh the benefits of aneurysmal occlusion in some cases. Hence, there is a critical need for developing techniques that assist physicians in assessing the risk of aneurysm rupture to determine which aneurysms require treatment. However, a reliable IA rupture risk prediction technique is currently unavailable. To address this issue, this study proposes a novel approach for aneurysm segmentation and multidisciplinary rupture prediction using 2D Digital Subtraction Angiography (DSA) images. The proposed method involves training a fully connected convolutional neural network (CNN) to segment aneurysm regions in DSA images, followed by extracting and fusing different features using a multidisciplinary approach, including deep features, geometrical features, Fourier descriptor, and shear pressure on the aneurysm wall. The proposed method also adopts a fast correlation-based filter approach to drop highly correlated features from the set of fused features. Finally, the selected fused features are passed through a Decision Tree classifier to predict the rupture severity of the associated aneurysm into four classes: Mild, Moderate, Severe, and Critical. The proposed method is evaluated on a newly developed DSA image dataset and on public datasets to assess its generalizability. The system's performance is also evaluated on DSA images annotated by expert neurosurgeons for the rupture risk assessment of the segmented aneurysm. The proposed system outperforms existing state-of-the-art segmentation methods, achieving an 85 % accuracy against annotated DSA images for the risk assessment of aneurysmal rupture.
Collapse
Affiliation(s)
- Muhammad Irfan
- SMILES LAB, Department of Computer Science and Engineering, Oakland University, Rochester, MI, 48309, USA
| | - Khalid Mahmood Malik
- SMILES LAB, Department of Computer Science and Engineering, Oakland University, Rochester, MI, 48309, USA.
| | - Jamil Ahmad
- Department of Computer Vision, Mohamed Bin Zayed University of Artificial Intelligence (MBZUAI), Abu Dhabi, United Arab Emirates
| | - Ghaus Malik
- Executive Vice-Chair at Department of Neurosurgery, Henry Ford Health System, Detroit, MI, USA
| |
Collapse
|
32
|
Wang F, Xiao C, Jia T, Pan L, Du F, Wang Z. Hepatobiliary surgery based on intelligent image segmentation technology. Open Life Sci 2023; 18:20220674. [PMID: 37671090 PMCID: PMC10476479 DOI: 10.1515/biol-2022-0674] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Revised: 07/01/2023] [Accepted: 07/12/2023] [Indexed: 09/07/2023] Open
Abstract
Liver disease is an important disease that seriously threatens human health. It accounts for the highest proportion in various malignant tumors, and its incidence rate and mortality are on the rise, seriously affecting human health. Modern imaging has developed rapidly, but the application of image segmentation in liver tumor surgery is still rare. The application of image processing technology represented by artificial intelligence (AI) in surgery can greatly improve the efficiency of surgery, reduce surgical complications, and reduce the cost of surgery. Hepatocellular carcinoma is the most common malignant tumor in the world, and its mortality is second only to lung cancer. The resection rate of liver cancer surgery is high, and it is a multidisciplinary surgery, so it is necessary to explore the possibility of effective switching between different disciplines. Resection of hepatobiliary and pancreatic tumors is one of the most challenging and lethal surgical procedures. The operation requires a high level of doctors' experience and understanding of anatomical structures. The surgical segmentation is slow and there may be obvious complications. Therefore, the surgical system needs to make full use of the relevant functions of AI technology and computer vision analysis software, and combine the processing strategy based on image processing algorithm and computer vision analysis model. Intelligent optimization algorithm, also known as modern heuristic algorithm, is an algorithm with global optimization performance, strong universality, and suitable for parallel processing. This algorithm generally has a strict theoretical basis, rather than relying solely on expert experience. In theory, the optimal solution or approximate optimal solution can be found in a certain time. This work studies the hepatobiliary surgery through intelligent image segmentation technology, and analyzes them through intelligent optimization algorithm. The research results showed that when other conditions were the same, there were three patients who had adverse reactions in hepatobiliary surgery through intelligent image segmentation technology, accounting for 10%. The number of patients with adverse reactions in hepatobiliary surgery by conventional methods was nine, accounting for 30%, which was significantly higher than the former, indicating a positive relationship between intelligent image segmentation technology and hepatobiliary surgery.
Collapse
Affiliation(s)
- Fuchuan Wang
- Faculty of Hepatology Medicine, Chinese People’s Liberation Army (PLA) General Hospital, Beijing100039, China
| | - Chaohui Xiao
- Faculty of Hepato-Biliary-Pancreatic Surgery, Chinese People’s Liberation Army (PLA) General Hospital, Beijing100853, China
| | - Tianye Jia
- Department of Laboratory, Fifth Medical Center, Chinese People’s Liberation Army (PLA) General Hospital, Beijing100039, China
| | - Liru Pan
- Faculty of Hepato-Biliary-Pancreatic Surgery, Chinese People’s Liberation Army (PLA) General Hospital, Beijing100853, China
| | - Fengxia Du
- Faculty of Hepatology Medicine, Chinese People’s Liberation Army (PLA) General Hospital, Beijing100039, China
| | - Zhaohai Wang
- Faculty of Hepato-Biliary-Pancreatic Surgery, Chinese People’s Liberation Army (PLA) General Hospital, Beijing100853, China
| |
Collapse
|
33
|
Du J, Zhang X, Liu P, Wang T. Coarse-Refined Consistency Learning Using Pixel-Level Features for Semi-Supervised Medical Image Segmentation. IEEE J Biomed Health Inform 2023; 27:3970-3981. [PMID: 37220034 DOI: 10.1109/jbhi.2023.3278741] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Pixel-level annotations are extremely expensive for medical image segmentation tasks as both expertise and time are needed to generate accurate annotations. Semi-supervised learning (SSL) for medical image segmentation has recently attracted growing attention because it can alleviate the exhausting manual annotations for clinicians by leveraging unlabeled data. However, most of the existing SSL methods do not take pixel-level information (e.g., pixel-level features) of labeled data into account, i.e., the labeled data are underutilized. Hence, in this work, an innovative Coarse-Refined Network with pixel-wise Intra-patch ranked loss and patch-wise Inter-patch ranked loss (CRII-Net) is proposed. It provides three advantages: i) it can produce stable targets for unlabeled data, as a simple yet effective coarse-refined consistency constraint is designed; ii) it is very effective for the extreme case where very scarce labeled data are available, as the pixel-level and patch-level features are extracted by our CRII-Net; and iii) it can output fine-grained segmentation results for hard regions (e.g., blurred object boundaries and low-contrast lesions), as the proposed Intra-Patch Ranked Loss (Intra-PRL) focuses on object boundaries and Inter-Patch Ranked loss (Inter-PRL) mitigates the adverse impact of low-contrast lesions. Experimental results on two common SSL tasks for medical image segmentation demonstrate the superiority of our CRII-Net. Specifically, when there are only 4% labeled data, our CRII-Net improves the Dice similarity coefficient (DSC) score by at least 7.49% when compared to five classical or state-of-the-art (SOTA) SSL methods. For hard samples/regions, our CRII-Net also significantly outperforms other compared methods in both quantitative and visualization results.
Collapse
|
34
|
Xu Z, Wang Y, Lu D, Luo X, Yan J, Zheng Y, Tong RKY. Ambiguity-selective consistency regularization for mean-teacher semi-supervised medical image segmentation. Med Image Anal 2023; 88:102880. [PMID: 37413792 DOI: 10.1016/j.media.2023.102880] [Citation(s) in RCA: 8] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/24/2022] [Revised: 05/17/2023] [Accepted: 06/22/2023] [Indexed: 07/08/2023]
Abstract
Semi-supervised learning has greatly advanced medical image segmentation since it effectively alleviates the need of acquiring abundant annotations from experts, wherein the mean-teacher model, known as a milestone of perturbed consistency learning, commonly serves as a standard and simple baseline. Inherently, learning from consistency can be regarded as learning from stability under perturbations. Recent improvement leans toward more complex consistency learning frameworks, yet, little attention is paid to the consistency target selection. Considering that the ambiguous regions from unlabeled data contain more informative complementary clues, in this paper, we improve the mean-teacher model to a novel ambiguity-consensus mean-teacher (AC-MT) model. Particularly, we comprehensively introduce and benchmark a family of plug-and-play strategies for ambiguous target selection from the perspectives of entropy, model uncertainty and label noise self-identification, respectively. Then, the estimated ambiguity map is incorporated into the consistency loss to encourage consensus between the two models' predictions in these informative regions. In essence, our AC-MT aims to find out the most worthwhile voxel-wise targets from the unlabeled data, and the model especially learns from the perturbed stability of these informative regions. The proposed methods are extensively evaluated on left atrium segmentation and brain tumor segmentation. Encouragingly, our strategies bring substantial improvement over recent state-of-the-art methods. The ablation study further demonstrates our hypothesis and shows impressive results under various extreme annotation conditions.
Collapse
Affiliation(s)
- Zhe Xu
- Department of Biomedical Engineering, The Chinese University of Hong Kong, Shatin, NT, Hong Kong, China.
| | - Yixin Wang
- Department of Bioengineering, Stanford University, Stanford, CA, USA
| | | | - Xiangde Luo
- Department of Mechanical Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Jiangpeng Yan
- Department of Automation, Tsinghua University, Beijing, China
| | | | - Raymond Kai-Yu Tong
- Department of Biomedical Engineering, The Chinese University of Hong Kong, Shatin, NT, Hong Kong, China.
| |
Collapse
|
35
|
Chen QQ, Sun ZH, Wei CF, Wu EQ, Ming D. Semi-Supervised 3D Medical Image Segmentation Based on Dual-Task Consistent Joint Learning and Task-Level Regularization. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2023; 20:2457-2467. [PMID: 35061590 DOI: 10.1109/tcbb.2022.3144428] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Semi-supervised learning has attracted wide attention from many researchers since its ability to utilize a few data with labels and relatively more data without labels to learn information. Some existing semi-supervised methods for medical image segmentation enforce the regularization of training by implicitly perturbing data or networks to perform the consistency. Most consistency regularization methods focus on data level or network structure level, and rarely of them focus on the task level. It may not directly lead to an improvement in task accuracy. To overcome the problem, this work proposes a semi-supervised dual-task consistent joint learning framework with task-level regularization for 3D medical image segmentation. Two branches are utilized to simultaneously predict the segmented and signed distance maps, and they can learn useful information from each other by constructing a consistency loss function between the two tasks. The segmentation branch learns rich information from both labeled and unlabeled data to strengthen the constraints on the geometric structure of the target. Experimental results on two benchmark datasets show that the proposed method can achieve better performance compared with other state-of-the-art works. It illustrates our method improves segmentation performance by utilizing unlabeled data and consistent regularization.
Collapse
|
36
|
Shen H, Yang Q, Chen Z, Ye Z, Dai P, Duan X. Semi-supervised OCT lesion segmentation via transformation-consistent with uncertainty and self-deep supervision. BIOMEDICAL OPTICS EXPRESS 2023; 14:3828-3840. [PMID: 37497513 PMCID: PMC10368041 DOI: 10.1364/boe.492680] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 06/25/2023] [Accepted: 06/26/2023] [Indexed: 07/28/2023]
Abstract
Optical coherence tomography (OCT) is a non-invasive, high-resolution ocular imaging technique with important implications for the diagnosis and management of retinal diseases. Automatic segmentation of lesions in OCT images is critical for assessing disease progression and treatment outcomes. However, existing methods for lesion segmentation require numerous pixel-wise annotations, which are difficult and time-consuming to obtain. To address this challenge, we propose a novel framework for semi-supervised OCT lesion segmentation, termed transformation-consistent with uncertainty and self-deep supervision (TCUS). To address the issue of lesion area blurring in OCT images and unreliable predictions from the teacher network for unlabeled images, an uncertainty-guided transformation-consistent strategy is proposed. Transformation-consistent is used to enhance the unsupervised regularization effect. The student network gradually learns from meaningful and reliable targets by utilizing the uncertainty information from the teacher network, to alleviate the performance degradation caused by potential errors in the teacher network's prediction results. Additionally, self-deep supervision is used to acquire multi-scale information from labeled and unlabeled OCT images, enabling accurate segmentation of lesions of various sizes and shapes. Self-deep supervision significantly improves the accuracy of lesion segmentation in terms of the Dice coefficient. Experimental results on two OCT datasets demonstrate that the proposed TCUS outperforms state-of-the-art semi-supervised segmentation methods.
Collapse
Affiliation(s)
- Hailan Shen
- School of Computer Science and Engineering, Central South University, Changsha 410083, China
| | - Qiao Yang
- School of Computer Science and Engineering, Central South University, Changsha 410083, China
| | - Zailiang Chen
- School of Computer Science and Engineering, Central South University, Changsha 410083, China
| | - Ziyu Ye
- XJTLU Entrepreneur College, Xi’an Jiaotong-liverpool University, Suzhou 215123, China
| | - Peishan Dai
- School of Computer Science and Engineering, Central South University, Changsha 410083, China
| | - Xuanchu Duan
- Changsha Aier Eye Hospital, Changsha 410015, China
| |
Collapse
|
37
|
Mirikharaji Z, Abhishek K, Bissoto A, Barata C, Avila S, Valle E, Celebi ME, Hamarneh G. A survey on deep learning for skin lesion segmentation. Med Image Anal 2023; 88:102863. [PMID: 37343323 DOI: 10.1016/j.media.2023.102863] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2022] [Revised: 02/01/2023] [Accepted: 05/31/2023] [Indexed: 06/23/2023]
Abstract
Skin cancer is a major public health problem that could benefit from computer-aided diagnosis to reduce the burden of this common disease. Skin lesion segmentation from images is an important step toward achieving this goal. However, the presence of natural and artificial artifacts (e.g., hair and air bubbles), intrinsic factors (e.g., lesion shape and contrast), and variations in image acquisition conditions make skin lesion segmentation a challenging task. Recently, various researchers have explored the applicability of deep learning models to skin lesion segmentation. In this survey, we cross-examine 177 research papers that deal with deep learning-based segmentation of skin lesions. We analyze these works along several dimensions, including input data (datasets, preprocessing, and synthetic data generation), model design (architecture, modules, and losses), and evaluation aspects (data annotation requirements and segmentation performance). We discuss these dimensions both from the viewpoint of select seminal works, and from a systematic viewpoint, examining how those choices have influenced current trends, and how their limitations should be addressed. To facilitate comparisons, we summarize all examined works in a comprehensive table as well as an interactive table available online3.
Collapse
Affiliation(s)
- Zahra Mirikharaji
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Kumar Abhishek
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada
| | - Alceu Bissoto
- RECOD.ai Lab, Institute of Computing, University of Campinas, Av. Albert Einstein 1251, Campinas 13083-852, Brazil
| | - Catarina Barata
- Institute for Systems and Robotics, Instituto Superior Técnico, Avenida Rovisco Pais, Lisbon 1049-001, Portugal
| | - Sandra Avila
- RECOD.ai Lab, Institute of Computing, University of Campinas, Av. Albert Einstein 1251, Campinas 13083-852, Brazil
| | - Eduardo Valle
- RECOD.ai Lab, School of Electrical and Computing Engineering, University of Campinas, Av. Albert Einstein 400, Campinas 13083-952, Brazil
| | - M Emre Celebi
- Department of Computer Science and Engineering, University of Central Arkansas, 201 Donaghey Ave., Conway, AR 72035, USA.
| | - Ghassan Hamarneh
- Medical Image Analysis Lab, School of Computing Science, Simon Fraser University, Burnaby V5A 1S6, Canada.
| |
Collapse
|
38
|
Wang J, Chang R, Zhao Z, Pahwa RS. Robust Detection, Segmentation, and Metrology of High Bandwidth Memory 3D Scans Using an Improved Semi-Supervised Deep Learning Approach. SENSORS (BASEL, SWITZERLAND) 2023; 23:5470. [PMID: 37420637 DOI: 10.3390/s23125470] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Revised: 05/21/2023] [Accepted: 05/27/2023] [Indexed: 07/09/2023]
Abstract
Recent advancements in 3D deep learning have led to significant progress in improving accuracy and reducing processing time, with applications spanning various domains such as medical imaging, robotics, and autonomous vehicle navigation for identifying and segmenting different structures. In this study, we employ the latest developments in 3D semi-supervised learning to create cutting-edge models for the 3D object detection and segmentation of buried structures in high-resolution X-ray semiconductors scans. We illustrate our approach to locating the region of interest of the structures, their individual components, and their void defects. We showcase how semi-supervised learning is utilized to capitalize on the vast amounts of available unlabeled data to enhance both detection and segmentation performance. Additionally, we explore the benefit of contrastive learning in the data pre-selection step for our detection model and multi-scale Mean Teacher training paradigm in 3D semantic segmentation to achieve better performance compared with the state of the art. Our extensive experiments have shown that our method achieves competitive performance and is able to outperform by up to 16% on object detection and 7.8% on semantic segmentation. Additionally, our automated metrology package shows a mean error of less than 2 μm for key features such as Bond Line Thickness and pad misalignment.
Collapse
Affiliation(s)
- Jie Wang
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #21-01, Connexis South Tower, Singapore 138632, Singapore
| | - Richard Chang
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #21-01, Connexis South Tower, Singapore 138632, Singapore
| | - Ziyuan Zhao
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #21-01, Connexis South Tower, Singapore 138632, Singapore
| | - Ramanpreet Singh Pahwa
- Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), 1 Fusionopolis Way, #21-01, Connexis South Tower, Singapore 138632, Singapore
| |
Collapse
|
39
|
Zou J, Wang Z, Du X. A Double-Teacher Model Capable of Exploiting Isomorphic and Heterogeneous Discrepancy Information for Medical Image Segmentation. Diagnostics (Basel) 2023; 13:diagnostics13111971. [PMID: 37296823 DOI: 10.3390/diagnostics13111971] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2023] [Revised: 05/20/2023] [Accepted: 05/31/2023] [Indexed: 06/12/2023] Open
Abstract
Deep learning, with continuous development, has achieved relatively good results in the field of left atrial segmentation, and numerous semi-supervised methods in this field have been implemented based on consistency regularization to obtain high-performance 3D models by training. However, most semi-supervised methods focus on inter-model consistency and ignore inter-model discrepancy. Therefore, we designed an improved double-teacher framework with discrepancy information. Herein, one teacher learns 2D information, another learns both 2D and 3D information, and the two models jointly guide the student model for learning. Simultaneously, we extract the isomorphic/heterogeneous discrepancy information between the predictions of the student and teacher model to optimize the whole framework. Unlike other semi-supervised methods based on 3D models, ours only uses 3D information to assist 2D models, and does not have a fully 3D model, thus addressing the large memory consumption and limited training data of 3D models to some extent. Our approach shows excellent performance on the left atrium (LA) dataset, similar to that of the best performing 3D semi-supervised methods available, compared to existing techniques.
Collapse
Affiliation(s)
- Junguo Zou
- School of Information Engineering, Chuzhou Polytechnic, Chuzhou 239000, China
| | - Zhaohe Wang
- School of Computer Science and Technology, Anhui University, Hefei 230601, China
| | - Xiuquan Du
- School of Computer Science and Technology, Anhui University, Hefei 230601, China
| |
Collapse
|
40
|
Huang M, Zhou S, Chen X, Lai H, Feng Q. Semi-supervised hybrid spine network for segmentation of spine MR images. Comput Med Imaging Graph 2023; 107:102245. [PMID: 37245416 DOI: 10.1016/j.compmedimag.2023.102245] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2023] [Revised: 04/25/2023] [Accepted: 05/11/2023] [Indexed: 05/30/2023]
Abstract
Automatic segmentation of vertebral bodies (VBs) and intervertebral discs (IVDs) in 3D magnetic resonance (MR) images is vital in diagnosing and treating spinal diseases. However, segmenting the VBs and IVDs simultaneously is not trivial. Moreover, problems exist, including blurry segmentation caused by anisotropy resolution, high computational cost, inter-class similarity and intra-class variability, and data imbalances. We proposed a two-stage algorithm, named semi-supervised hybrid spine network (SSHSNet), to address these problems by achieving accurate simultaneous VB and IVD segmentation. In the first stage, we constructed a 2D semi-supervised DeepLabv3+ by using cross pseudo supervision to obtain intra-slice features and coarse segmentation. In the second stage, a 3D full-resolution patch-based DeepLabv3+ was built. This model can be used to extract inter-slice information and combine the coarse segmentation and intra-slice features provided from the first stage. Moreover, a cross tri-attention module was applied to compensate for the loss of inter-slice and intra-slice information separately generated from 2D and 3D networks, thereby improving feature representation ability and achieving satisfactory segmentation results. The proposed SSHSNet was validated on a publicly available spine MR image dataset, and remarkable segmentation performance was achieved. Moreover, results show that the proposed method has great potential in dealing with the data imbalance problem. Based on previous reports, few studies have incorporated a semi-supervised learning strategy with a cross attention mechanism for spine segmentation. Therefore, the proposed method may provide a useful tool for spine segmentation and aid clinically in spinal disease diagnoses and treatments. Codes are publicly available at: https://github.com/Meiyan88/SSHSNet.
Collapse
Affiliation(s)
- Meiyan Huang
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China.
| | - Shuoling Zhou
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China
| | - Xiumei Chen
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China
| | - Haoran Lai
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China.
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China.
| |
Collapse
|
41
|
Tang Y, Wang S, Qu Y, Cui Z, Zhang W. Consistency and adversarial semi-supervised learning for medical image segmentation. Comput Biol Med 2023; 161:107018. [PMID: 37216776 DOI: 10.1016/j.compbiomed.2023.107018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2022] [Revised: 04/15/2023] [Accepted: 05/05/2023] [Indexed: 05/24/2023]
Abstract
Medical image segmentation based on deep learning has made enormous progress in recent years. However, the performance of existing methods generally heavily relies on a large amount of labeled data, which are commonly expensive and time-consuming to obtain. To settle above issue, in this paper, a novel semi-supervised medical image segmentation method is proposed, in which the adversarial training mechanism and the collaborative consistency learning strategy are introduced into the mean teacher model. With the adversarial training mechanism, the discriminator can generate confidence maps for unlabeled data, such that more reliable supervised information for the student network is exploited. In the process of adversarial training, we further propose a collaborative consistency learning strategy by which the auxiliary discriminator can assist the primary discriminator in achieving supervised information with higher quality. We extensively evaluate our method on three representative yet challenging medical image segmentation tasks: (1) skin lesion segmentation from dermoscopy images in the International Skin Imaging Collaboration (ISIC) 2017 dataset; (2) optic cup and optic disk (OC/OD) segmentation from fundus images in the Retinal Fundus Glaucoma Challenge (REFUGE) dataset; and (3) tumor segmentation from lower-grade glioma (LGG) tumors images. The experimental results validate the superiority and effectiveness of our proposal when compared with the state-of-the-art semi-supervised medical image segmentation methods.
Collapse
Affiliation(s)
- Yongqiang Tang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China.
| | - Shilei Wang
- School of Computer Science and Technology, Taiyuan University of Science and Technology, Taiyuan, Shanxi, 030024, China.
| | - Yuxun Qu
- College of Computer Science, Nankai University, Tianjin, 300350, China.
| | - Zhihua Cui
- School of Computer Science and Technology, Taiyuan University of Science and Technology, Taiyuan, Shanxi, 030024, China.
| | - Wensheng Zhang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, 100190, China.
| |
Collapse
|
42
|
Dai W, Li X, Ding X, Cheng KT. Cyclical Self-Supervision for Semi-Supervised Ejection Fraction Prediction From Echocardiogram Videos. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:1446-1461. [PMID: 37015560 DOI: 10.1109/tmi.2022.3229136] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Left-ventricular ejection fraction (LVEF) is an important indicator of heart failure. Existing methods for LVEF estimation from video require large amounts of annotated data to achieve high performance, e.g. using 10,030 labeled echocardiogram videos to achieve mean absolute error (MAE) of 4.10. Labeling these videos is time-consuming however and limits potential downstream applications to other heart diseases. This paper presents the first semi-supervised approach for LVEF prediction. Unlike general video prediction tasks, LVEF prediction is specifically related to changes in the left ventricle (LV) in echocardiogram videos. By incorporating knowledge learned from predicting LV segmentations into LVEF regression, we can provide additional context to the model for better predictions. To this end, we propose a novel Cyclical Self-Supervision (CSS) method for learning video-based LV segmentation, which is motivated by the observation that the heartbeat is a cyclical process with temporal repetition. Prediction masks from our segmentation model can then be used as additional input for LVEF regression to provide spatial context for the LV region. We also introduce teacher-student distillation to distill the information from LV segmentation masks into an end-to-end LVEF regression model that only requires video inputs. Results show our method outperforms alternative semi-supervised methods and can achieve MAE of 4.17, which is competitive with state-of-the-art supervised performance, using half the number of labels. Validation on an external dataset also shows improved generalization ability from using our method.
Collapse
|
43
|
Qiao P, Li H, Song G, Han H, Gao Z, Tian Y, Liang Y, Li X, Zhou SK, Chen J. Semi-Supervised CT Lesion Segmentation Using Uncertainty-Based Data Pairing and SwapMix. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:1546-1562. [PMID: 37015649 DOI: 10.1109/tmi.2022.3232572] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Semi-supervised learning (SSL) methods show their powerful performance to deal with the issue of data shortage in the field of medical image segmentation. However, existing SSL methods still suffer from the problem of unreliable predictions on unannotated data due to the lack of manual annotations for them. In this paper, we propose an unreliability-diluted consistency training (UDiCT) mechanism to dilute the unreliability in SSL by assembling reliable annotated data into unreliable unannotated data. Specifically, we first propose an uncertainty-based data pairing module to pair annotated data with unannotated data based on a complementary uncertainty pairing rule, which avoids two hard samples being paired off. Secondly, we develop SwapMix, a mixed sample data augmentation method, to integrate annotated data into unannotated data for training our model in a low-unreliability manner. Finally, UDiCT is trained by minimizing a supervised loss and an unreliability-diluted consistency loss, which makes our model robust to diverse backgrounds. Extensive experiments on three chest CT datasets show the effectiveness of our method for semi-supervised CT lesion segmentation.
Collapse
|
44
|
Lei T, Zhang D, Du X, Wang X, Wan Y, Nandi AK. Semi-Supervised Medical Image Segmentation Using Adversarial Consistency Learning and Dynamic Convolution Network. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:1265-1277. [PMID: 36449588 DOI: 10.1109/tmi.2022.3225687] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Popular semi-supervised medical image segmentation networks often suffer from error supervision from unlabeled data since they usually use consistency learning under different data perturbations to regularize model training. These networks ignore the relationship between labeled and unlabeled data, and only compute single pixel-level consistency leading to uncertain prediction results. Besides, these networks often require a large number of parameters since their backbone networks are designed depending on supervised image segmentation tasks. Moreover, these networks often face a high over-fitting risk since a small number of training samples are popular for semi-supervised image segmentation. To address the above problems, in this paper, we propose a novel adversarial self-ensembling network using dynamic convolution (ASE-Net) for semi-supervised medical image segmentation. First, we use an adversarial consistency training strategy (ACTS) that employs two discriminators based on consistency learning to obtain prior relationships between labeled and unlabeled data. The ACTS can simultaneously compute pixel-level and image-level consistency of unlabeled data under different data perturbations to improve the prediction quality of labels. Second, we design a dynamic convolution-based bidirectional attention component (DyBAC) that can be embedded in any segmentation network, aiming at adaptively adjusting the weights of ASE-Net based on the structural information of input samples. This component effectively improves the feature representation ability of ASE-Net and reduces the overfitting risk of the network. The proposed ASE-Net has been extensively tested on three publicly available datasets, and experiments indicate that ASE-Net is superior to state-of-the-art networks, and reduces computational costs and memory overhead. The code is available at: https://github.com/SUST-reynole/ASE-Nethttps://github.com/SUST-reynole/ASE-Net.
Collapse
|
45
|
Zhu Y, Yin X, Meijering E. A Compound Loss Function With Shape Aware Weight Map for Microscopy Cell Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:1278-1288. [PMID: 36455082 DOI: 10.1109/tmi.2022.3226226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Microscopy cell segmentation is a crucial step in biological image analysis and a challenging task. In recent years, deep learning has been widely used to tackle this task, with promising results. A critical aspect of training complex neural networks for this purpose is the selection of the loss function, as it affects the learning process. In the field of cell segmentation, most of the recent research in improving the loss function focuses on addressing the problem of inter-class imbalance. Despite promising achievements, more work is needed, as the challenge of cell segmentation is not only the inter-class imbalance but also the intra-class imbalance (the cost imbalance between the false positives and false negatives of the inference model), the segmentation of cell minutiae, and the missing annotations. To deal with these challenges, in this paper, we propose a new compound loss function employing a shape aware weight map. The proposed loss function is inspired by Youden's J index to handle the problem of inter-class imbalance and uses a focal cross-entropy term to penalize the intra-class imbalance and weight easy/hard samples. The proposed shape aware weight map can handle the problem of missing annotations and facilitate valid segmentation of cell minutiae. Results of evaluations on all ten 2D+time datasets from the public cell tracking challenge demonstrate 1) the superiority of the proposed loss function with the shape aware weight map, and 2) that the performance of recent deep learning-based cell segmentation methods can be improved by using the proposed compound loss function.
Collapse
|
46
|
Chen Z, Hou Y, Liu H, Ye Z, Zhao R, Shen H. FDCT: Fusion-Guided Dual-View Consistency Training for semi-supervised tissue segmentation on MRI. Comput Biol Med 2023; 160:106908. [PMID: 37120986 DOI: 10.1016/j.compbiomed.2023.106908] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2022] [Revised: 03/30/2023] [Accepted: 04/10/2023] [Indexed: 05/02/2023]
Abstract
Accurate tissue segmentation on MRI is important for physicians to make diagnosis and treatment for patients. However, most of the models are only designed for single-task tissue segmentation, and tend to lack generality to other MRI tissue segmentation tasks. Not only that, the acquisition of labels is time-consuming and laborious, which remains a challenge to be solved. In this study, we propose the universal Fusion-Guided Dual-View Consistency Training(FDCT) for semi-supervised tissue segmentation on MRI. It can obtain accurate and robust tissue segmentation for multiple tasks, and alleviates the problem of insufficient labeled data. Especially, for building bidirectional consistency, we feed dual-view images into a single-encoder dual-decoder structure to obtain view-level predictions, then put them into a fusion module to generate image-level pseudo-label. Moreover, to improve boundary segmentation quality, we propose the Soft-label Boundary Optimization Module(SBOM). We have conducted extensive experiments on three MRI datasets to evaluate the effectiveness of our method. Experimental results demonstrate that our method outperforms the state-of-the-art semi-supervised medical image segmentation methods.
Collapse
Affiliation(s)
- Zailiang Chen
- Central South University, No. 932 Lushan South Road, Changsha, 410000, China.
| | - Yazheng Hou
- Central South University, No. 932 Lushan South Road, Changsha, 410000, China.
| | - Hui Liu
- Xiangya Hospital of Central South University, No. 87 Xiangya Road, 410000, China.
| | - Ziyu Ye
- Xi'an Jiaotong-liverpool University, No. 111 Renai Road, 215000, China.
| | - Rongchang Zhao
- Central South University, No. 932 Lushan South Road, Changsha, 410000, China.
| | - Hailan Shen
- Central South University, No. 932 Lushan South Road, Changsha, 410000, China.
| |
Collapse
|
47
|
Ou Y, Huang SX, Wong KK, Cummock J, Volpi J, Wang JZ, Wong STC. BBox-Guided Segmentor: Leveraging expert knowledge for accurate stroke lesion segmentation using weakly supervised bounding box prior. Comput Med Imaging Graph 2023; 107:102236. [PMID: 37146318 DOI: 10.1016/j.compmedimag.2023.102236] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2022] [Revised: 02/17/2023] [Accepted: 04/06/2023] [Indexed: 05/07/2023]
Abstract
Stroke is one of the leading causes of death and disability in the world. Despite intensive research on automatic stroke lesion segmentation from non-invasive imaging modalities including diffusion-weighted imaging (DWI), challenges remain such as a lack of sufficient labeled data for training deep learning models and failure in detecting small lesions. In this paper, we propose BBox-Guided Segmentor, a method that significantly improves the accuracy of stroke lesion segmentation by leveraging expert knowledge. Specifically, our model uses a very coarse bounding box label provided by the expert and then performs accurate segmentation automatically. The small overhead of having the expert provide a rough bounding box leads to large performance improvement in segmentation, which is paramount to accurate stroke diagnosis. To train our model, we employ a weakly-supervised approach that uses a large number of weakly-labeled images with only bounding boxes and a small number of fully labeled images. The scarce fully labeled images are used to train a generator segmentation network, while adversarial training is used to leverage the large number of weakly-labeled images to provide additional learning signals. We evaluate our method extensively using a unique clinical dataset of 99 fully labeled cases (i.e., with full segmentation map labels) and 831 weakly labeled cases (i.e., with only bounding box labels), and the results demonstrate the superior performance of our approach over state-of-the-art stroke lesion segmentation models. We also achieve competitive performance as a SOTA fully supervised method using less than one-tenth of the complete labels. Our proposed approach has the potential to improve stroke diagnosis and treatment planning, which may lead to better patient outcomes.
Collapse
Affiliation(s)
- Yanglan Ou
- Data Science and Artificial Intelligence Area, College of Information Sciences and Technology, The Pennsylvania State University, University Park, PA 16802, USA.
| | - Sharon X Huang
- Data Science and Artificial Intelligence Area, College of Information Sciences and Technology, The Pennsylvania State University, University Park, PA 16802, USA.
| | - Kelvin K Wong
- T.T. and W.F. Chao Center for BRAIN & Houston Methodist Cancer Center, Houston Methodist Hospital, Houston, TX 77030, USA.
| | - Jonathon Cummock
- T.T. and W.F. Chao Center for BRAIN & Houston Methodist Cancer Center, Houston Methodist Hospital, Houston, TX 77030, USA
| | - John Volpi
- Eddy Scurlock Comprehensive Stroke Center, Department of Neurology, Houston Methodist Hospital, Houston, TX 77030, USA
| | - James Z Wang
- Data Science and Artificial Intelligence Area, College of Information Sciences and Technology, The Pennsylvania State University, University Park, PA 16802, USA
| | - Stephen T C Wong
- T.T. and W.F. Chao Center for BRAIN & Houston Methodist Cancer Center, Houston Methodist Hospital, Houston, TX 77030, USA
| |
Collapse
|
48
|
Semi-supervised segmentation of coronary DSA using mixed networks and multi-strategies. Comput Biol Med 2023; 156:106493. [PMID: 36893708 DOI: 10.1016/j.compbiomed.2022.106493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Revised: 12/11/2022] [Accepted: 12/27/2022] [Indexed: 12/31/2022]
Abstract
The coronary arteries supply blood to the myocardium, which originate from the root of the aorta and mainly branch into the left and right. X-ray digital subtraction angiography (DSA) is a technique for evaluating coronary artery plaques and narrowing, that is widely used because of its time efficiency and cost-effectiveness. However, automated coronary vessel classification and segmentation remains challenging using a little data. Therefore, the purpose of this study is twofold: one is to propose a more robust method for vessel segmentation, the other is to provide a solution that is feasible with a small amount of labeled data. Currently, there are three main types of vessel segmentation methods, i.e., graphical- and statistical-based; clustering theory based, and deep learning-based methods for pixel-by-pixel probabilistic prediction, among which the last method is the mainstream with high accuracy and automation. Under this trend, an Inception-SwinUnet (ISUnet) network combining the convolutional neural network and Transformer basic module was proposed in this paper. Considering that data-driven fully supervised learning (FSL) segmentation methods require a large set of paired data with high-quality pixel-level annotation, which is expertise-demanding and time-consuming, we proposed a Semi-supervised Learning (SSL) method to achieve better performance with a small amount of labeled and unlabeled data. Different from the classical SSL method, i.e., Mean-Teacher, our method used two different networks for cross-teaching as the backbone. Meanwhile, inspired by deep supervision and confidence learning (CL), two effective strategies for SSL were adopted, which were denominated Pyramid-consistency Learning (PL) and Confidence Learning (CL), respectively. Both were designed to filter the noise and improve the credibility of pseudo labels generated by unlabeled data. Compared with existing methods, ours achieved superior segmentation performance over other FSL and SSL ones by using data with a small equal number of labels. Code is available in https://github.com/Allenem/SSL4DSA.
Collapse
|
49
|
Lu Y, Shen Y, Xing X, Ye C, Meng MQH. Boundary-enhanced semi-supervised retinal layer segmentation in optical coherence tomography images using fewer labels. Comput Med Imaging Graph 2023; 105:102199. [PMID: 36805709 DOI: 10.1016/j.compmedimag.2023.102199] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2022] [Revised: 01/24/2023] [Accepted: 02/06/2023] [Indexed: 02/13/2023]
Abstract
Automatic segmentation of multiple layers in retinal optical coherence tomography (OCT) images is crucial for eye disease diagnosis and treatment. Despite the success of deep learning algorithms, it still remains a challenge due to the blurry layer boundaries and lack of adequate pixel-wise annotations. To tackle these issues, we propose a Boundary-Enhanced Semi-supervised Network (BE-SemiNet) that exploits an auxiliary distance map regression task to improve retinal layer segmentation with scarce labeled data and abundant unlabeled data. Specifically, a novel Unilaterally Truncated Distance Map (UTDM) is firstly introduced to alleviate the class imbalance problem and enhance the layer boundary learning in the regression task. Then for the pixel-wise segmentation and UTDM regression branches, we impose task-level and data-level consistency regularization on unlabeled data to enrich the diversity of unsupervised information and improve the regularization effects. Pseudo supervision is incorporated in consistency regularization to bridge the task prediction spaces for consistency and expand training labeled data. Experiments on two public retinal OCT datasets show that our method can greatly improve the supervised baseline performance with only 5 annotations and outperform the state-of-the-art methods. Since it is difficult and labor-expensive to obtain adequate pixel-wise annotations in practice, our method has a promising application future in clinical retinal OCT image analysis.
Collapse
Affiliation(s)
- Ye Lu
- Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | - Yutian Shen
- Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | - Xiaohan Xing
- Department of Electrical Engineering, The City University of Hong Kong, Hong Kong, China
| | - Chengwei Ye
- Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | - Max Q-H Meng
- Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong, China; Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, China; Shenzhen Research Institute of The Chinese University of Hong Kong, Shenzhen, China.
| |
Collapse
|
50
|
Chen J, Zhang J, Debattista K, Han J. Semi-Supervised Unpaired Medical Image Segmentation Through Task-Affinity Consistency. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:594-605. [PMID: 36219664 DOI: 10.1109/tmi.2022.3213372] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
Deep learning-based semi-supervised learning (SSL) algorithms are promising in reducing the cost of manual annotation of clinicians by using unlabelled data, when developing medical image segmentation tools. However, to date, most existing semi-supervised learning (SSL) algorithms treat the labelled images and unlabelled images separately and ignore the explicit connection between them; this disregards essential shared information and thus hinders further performance improvements. To mine the shared information between the labelled and unlabelled images, we introduce a class-specific representation extraction approach, in which a task-affinity module is specifically designed for representation extraction. We further cast the representation into two different views of feature maps; one is focusing on low-level context, while the other concentrates on structural information. The two views of feature maps are incorporated into the task-affinity module, which then extracts the class-specific representations to aid the knowledge transfer from the labelled images to the unlabelled images. In particular, a task-affinity consistency loss between the labelled images and unlabelled images based on the multi-scale class-specific representations is formulated, leading to a significant performance improvement. Experimental results on three datasets show that our method consistently outperforms existing state-of-the-art methods. Our findings highlight the potential of consistency between class-specific knowledge for semi-supervised medical image segmentation. The code and models are to be made publicly available at https://github.com/jingkunchen/TAC.
Collapse
|