1
|
Migliorelli G, Fiorentino MC, Di Cosmo M, Villani FP, Mancini A, Moccia S. On the use of contrastive learning for standard-plane classification in fetal ultrasound imaging. Comput Biol Med 2024; 174:108430. [PMID: 38613892 DOI: 10.1016/j.compbiomed.2024.108430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Revised: 03/06/2024] [Accepted: 04/07/2024] [Indexed: 04/15/2024]
Abstract
BACKGROUND To investigate the effectiveness of contrastive learning, in particular SimClr, in reducing the need for large annotated ultrasound (US) image datasets for fetal standard plane identification. METHODS We explore SimClr advantage in the cases of both low and high inter-class variability, considering at the same time how classification performance varies according to different amounts of labels used. This evaluation is performed by exploiting contrastive learning through different training strategies. We apply both quantitative and qualitative analyses, using standard metrics (F1-score, sensitivity, and precision), Class Activation Mapping (CAM), and t-Distributed Stochastic Neighbor Embedding (t-SNE). RESULTS When dealing with high inter-class variability classification tasks, contrastive learning does not bring a significant advantage; whereas it results to be relevant for low inter-class variability classification, specifically when initialized with ImageNet weights. CONCLUSIONS Contrastive learning approaches are typically used when a large number of unlabeled data is available, which is not representative of US datasets. We proved that SimClr either as pre-training with backbone initialized via ImageNet weights or used in an end-to-end dual-task may impact positively the performance over standard transfer learning approaches, under a scenario in which the dataset is small and characterized by low inter-class variability.
Collapse
Affiliation(s)
| | | | - Mariachiara Di Cosmo
- Department of Information Engineering, Università Politecnica delle Marche, Ancona, Italy
| | | | - Adriano Mancini
- Department of Information Engineering, Università Politecnica delle Marche, Ancona, Italy
| | - Sara Moccia
- The BioRobotics Institute and Department of Excellence in Robotics and AI, Scuola Superiore Sant'Anna, Pisa, Italy
| |
Collapse
|
2
|
Qi Y, Hossain MS. Semi-supervised Federated Learning for Digital Twin 6G-enabled IIoT: A Bayesian estimated approach. J Adv Res 2024:S2090-1232(24)00074-2. [PMID: 38417576 DOI: 10.1016/j.jare.2024.02.012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Revised: 01/03/2024] [Accepted: 02/17/2024] [Indexed: 03/01/2024] Open
Abstract
INTRODUCTION In recent years, the proliferation of Industrial Internet of Things (IIoT) devices has resulted in a substantial increase in data generation across various domains, including the nascent 6G networks. Digital Twins (DTs), serving as virtual replicas of physical entities, have gained popularity within the realm of IoT due to their capacity to simulate and optimize physical systems in a cost-effective manner. Nonetheless, the security of DTs and the safeguarding of the sensitive data they generate have emerged as paramount concerns. Fortunately, the Federated Fearning (FL) system has emerged as a promising solution to address the challenge of data privacy within DTs. Nonetheless, the requisite acquisition of a significant volume of labeled data for training purposes poses a formidable challenge, particularly in a DT environment that blends real and virtual data. OBJECTIVES To tackle this challenge, this study presents an innovative Semi-supervised FL (SSFL) framework designed to overcome the scarcity of labeled data through the strategic utilization of pseudo-labels. METHODS Specifically, our proposed SSFL algorithm, named SSFL-MBE, introduces a novel approach by combining Mix data augmentation and Bayesian Estimation consistency regularization loss, thereby integrating robust augmentation techniques to enhance model generalization. Furthermore, we introduce a Bayesian-estimated pseudo-label loss that leverages prior probabilistic knowledge to enhance model performance. Our investigation focuses particularly on a demanding scenario where labeled and unlabeled data are segregated across disparate locations, specifically, the server and various clients. RESULTS Comprehensive evaluations conducted on CIFAR-10 and MNIST datasets conclusively demonstrate that our proposed algorithm consistently surpasses mainstream SSFL baseline models, exhibiting an enhancement in model performance ranging from 0.5% to 1.5%. CONCLUSION Overall, this work contributes to the development of more efficient and secure approaches for model training in DT-empowered FL settings, which is crucial for the deployment of IIoTs in 6G-enabled environments.
Collapse
Affiliation(s)
- Yuanhang Qi
- School of Computer Science, University of Electronic Science and Technology of China, Zhongshan Institute, Zhongshan 528402, China.
| | - M Shamim Hossain
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, Riyadh 12372, Saudi Arabia.
| |
Collapse
|
3
|
Ouyang J, Mao D, Guo Z, Liu S, Xu D, Wang W. Contrastive self-supervised learning for diabetic retinopathy early detection. Med Biol Eng Comput 2023; 61:2441-2452. [PMID: 37119374 DOI: 10.1007/s11517-023-02810-5] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2022] [Accepted: 02/22/2023] [Indexed: 05/01/2023]
Abstract
Diabetic Retinopathy (DR) is the major cause of blindness, which seriously threatens the world's vision health. Limited medical resources make early diagnosis and a large-scale screening of DR difficult. Most of the current automatic diagnostic methods are mostly based on deep learning and large-scale labeled data. However, the insufficiency of manual annotations for medical images still is a great challenge of training deep neural networks. Self-supervised learning methods are proposed to learn general features from dataset without manual annotations. Inspired by this, we proposed a deep learning based DR classification model (SimCLR-DR). In this paper, we first use contrastive self-learning algorithm to pre-train the encoder based on convolution network with unlabeled retinal images, then retrain the encoder with classifier on a small annotated training data to detect referable DR. The experimental results on Kaggle dataset show that this proposed method can overcome the training data insufficiency problem and performs better than transfer learning. SimCLR-DR is a good beginning for other deep learning based medical image detection approaches facing the challenge of insufficient annotated data. Figure presents an overview of the proposed framework, which contains three main steps: (i) Data preprocessing; (ii) Pretext task of SimCLR-DR based on contrastive learning; (iii) Downstream Task of SimCLRDR based on CNN.
Collapse
Affiliation(s)
- Jihong Ouyang
- Department of Computer Science and Technology, Jilin University, Qianjin Street, Changchun, 130015, Jilin Province, China
| | - Dong Mao
- Department of Computer Science and Technology, Jilin University, Qianjin Street, Changchun, 130015, Jilin Province, China
| | - Zeqi Guo
- Department of Computer Science and Technology, Jilin University, Qianjin Street, Changchun, 130015, Jilin Province, China
| | - Siguang Liu
- Department of Computer Science and Technology, Jilin University, Qianjin Street, Changchun, 130015, Jilin Province, China.
| | - Dong Xu
- University of Missouri, Columbia, MO, USA
| | - Wenting Wang
- Department of Computer Science, University of Texas at Dallas, Richardson, TX, USA
| |
Collapse
|
4
|
Kokilepersaud K, Corona ST, Prabhushankar M, AlRegib G, Wykoff C. Clinically Labeled Contrastive Learning for OCT Biomarker Classification. IEEE J Biomed Health Inform 2023; 27:4397-4408. [PMID: 37216249 DOI: 10.1109/jbhi.2023.3277789] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
This article presents a novel positive and negative set selection strategy for contrastive learning of medical images based on labels that can be extracted from clinical data. In the medical field, there exists a variety of labels for data that serve different purposes at different stages of a diagnostic and treatment process. Clinical labels and biomarker labels are two examples. In general, clinical labels are easier to obtain in larger quantities because they are regularly collected during routine clinical care, while biomarker labels require expert analysis and interpretation to obtain. Within the field of ophthalmology, previous work has shown that clinical values exhibit correlations with biomarker structures that manifest within optical coherence tomography (OCT) scans. We exploit this relationship by using the clinical data as pseudo-labels for our data without biomarker labels in order to choose positive and negative instances for training a backbone network with a supervised contrastive loss. In this way, a backbone network learns a representation space that aligns with the clinical data distribution available. Afterwards, we fine-tune the network trained in this manner with the smaller amount of biomarker labeled data with a cross-entropy loss in order to classify these key indicators of disease directly from OCT scans. We also expand on this concept by proposing a method that uses a linear combination of clinical contrastive losses. We benchmark our methods against state of the art self-supervised methods in a novel setting with biomarkers of varying granularity. We show performance improvements by as much as 5% in total biomarker detection AUROC.
Collapse
|
5
|
Wang Q, Chen K, Dou W, Ma Y. Cross-Attention Based Multi-Resolution Feature Fusion Model for Self-Supervised Cervical OCT Image Classification. IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2023; 20:2541-2554. [PMID: 37027657 DOI: 10.1109/tcbb.2023.3246979] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Cervical cancer seriously endangers the health of the female reproductive system and even risks women's life in severe cases. Optical coherence tomography (OCT) is a non-invasive, real-time, high-resolution imaging technology for cervical tissues. However, since the interpretation of cervical OCT images is a knowledge-intensive, time-consuming task, it is tough to acquire a large number of high-quality labeled images quickly, which is a big challenge for supervised learning. In this study, we introduce the vision Transformer (ViT) architecture, which has recently achieved impressive results in natural image analysis, into the classification task of cervical OCT images. Our work aims to develop a computer-aided diagnosis (CADx) approach based on a self-supervised ViT-based model to classify cervical OCT images effectively. We leverage masked autoencoders (MAE) to perform self-supervised pre-training on cervical OCT images, so the proposed classification model has a better transfer learning ability. In the fine-tuning process, the ViT-based classification model extracts multi-scale features from OCT images of different resolutions and fuses them with the cross-attention module. The ten-fold cross-validation results on an OCT image dataset from a multi-center clinical study of 733 patients in China indicate that our model achieved an AUC value of 0.9963 ± 0.0069 with a 95.89 ± 3.30% sensitivity and 98.23 ± 1.36 % specificity, outperforming some state-of-the-art classification models based on Transformers and convolutional neural networks (CNNs) in the binary classification task of detecting high-risk cervical diseases, including high-grade squamous intraepithelial lesion (HSIL) and cervical cancer. Furthermore, our model with the cross-shaped voting strategy achieved a sensitivity of 92.06% and specificity of 95.56% on an external validation dataset containing 288 three-dimensional (3D) OCT volumes from 118 Chinese patients in a different new hospital. This result met or exceeded the average of four medical experts who have used OCT for over one year. In addition to promising classification performance, our model has a remarkable ability to detect and visualize local lesions using the attention map of the standard ViT model, providing good interpretability for gynecologists to locate and diagnose possible cervical diseases.
Collapse
|
6
|
Chen Z, Wei Q. Developing an Improved Survival Prediction Model for Disease Prognosis. Biomolecules 2022; 12:biom12121751. [PMID: 36551179 PMCID: PMC9775036 DOI: 10.3390/biom12121751] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2022] [Revised: 11/18/2022] [Accepted: 11/23/2022] [Indexed: 11/27/2022] Open
Abstract
Machine learning has become an important research field in genetics and molecular biology. Survival analysis using machine learning can provide an important computed-aid clinical research scheme for evaluating tumor treatment options. However, the genomic features are high-dimensional, which limits the prediction performance of the survival learning model. Therefore, in this paper, we propose an improved survival prediction model using a deep forest and self-supervised learning. It uses a deep survival forest to perform adaptive learning of high-dimensional genomic data and ensure robustness. In addition, self-supervised learning, as a semi-supervised learning style, is designed to utilize unlabeled samples to improve model performance. Based on four cancer datasets from The Cancer Genome Atlas (TCGA), the experimental results show that our proposed method outperforms four advanced survival analysis methods in terms of the C-index and brier score. The developed prediction model will help doctors rethink patient characteristics' relevance to survival time and personalize treatment decisions.
Collapse
|
7
|
Semi-supervised structure attentive temporal mixup coherence for medical image segmentation. Biocybern Biomed Eng 2022. [DOI: 10.1016/j.bbe.2022.09.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/22/2022]
|
8
|
Liu Y, He Q, Duan H, Shi H, Han A, He Y. Using Sparse Patch Annotation for Tumor Segmentation in Histopathological Images. SENSORS (BASEL, SWITZERLAND) 2022; 22:6053. [PMID: 36015814 PMCID: PMC9414209 DOI: 10.3390/s22166053] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/19/2022] [Revised: 08/05/2022] [Accepted: 08/10/2022] [Indexed: 06/15/2023]
Abstract
Tumor segmentation is a fundamental task in histopathological image analysis. Creating accurate pixel-wise annotations for such segmentation tasks in a fully-supervised training framework requires significant effort. To reduce the burden of manual annotation, we propose a novel weakly supervised segmentation framework based on sparse patch annotation, i.e., only small portions of patches in an image are labeled as 'tumor' or 'normal'. The framework consists of a patch-wise segmentation model called PSeger, and an innovative semi-supervised algorithm. PSeger has two branches for patch classification and image classification, respectively. This two-branch structure enables the model to learn more general features and thus reduce the risk of overfitting when learning sparsely annotated data. We incorporate the idea of consistency learning and self-training into the semi-supervised training strategy to take advantage of the unlabeled images. Trained on the BCSS dataset with only 25% of the images labeled (five patches for each labeled image), our proposed method achieved competitive performance compared to the fully supervised pixel-wise segmentation models. Experiments demonstrate that the proposed solution has the potential to reduce the burden of labeling histopathological images.
Collapse
Affiliation(s)
- Yiqing Liu
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| | - Qiming He
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| | - Hufei Duan
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| | - Huijuan Shi
- Department of Pathology, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou 510080, China
| | - Anjia Han
- Department of Pathology, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou 510080, China
| | - Yonghong He
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| |
Collapse
|
9
|
Peng X, Li H, Yuan F, Razul SG, Chen Z, Lin Z. An extreme learning machine for unsupervised online anomaly detection in multivariate time series. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.06.042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
10
|
Baffour AA, Qin Z, Geng J, Ding Y, Deng F, Qin Z. Generic network for domain adaptation based on self-supervised learning and deep clustering. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.12.099] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/01/2022]
|
11
|
Shurrab S, Duwairi R. Self-supervised learning methods and applications in medical imaging analysis: a survey. PeerJ Comput Sci 2022; 8:e1045. [PMID: 36091989 PMCID: PMC9455147 DOI: 10.7717/peerj-cs.1045] [Citation(s) in RCA: 28] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2021] [Accepted: 06/27/2022] [Indexed: 05/12/2023]
Abstract
The scarcity of high-quality annotated medical imaging datasets is a major problem that collides with machine learning applications in the field of medical imaging analysis and impedes its advancement. Self-supervised learning is a recent training paradigm that enables learning robust representations without the need for human annotation which can be considered an effective solution for the scarcity of annotated medical data. This article reviews the state-of-the-art research directions in self-supervised learning approaches for image data with a concentration on their applications in the field of medical imaging analysis. The article covers a set of the most recent self-supervised learning methods from the computer vision field as they are applicable to the medical imaging analysis and categorize them as predictive, generative, and contrastive approaches. Moreover, the article covers 40 of the most recent research papers in the field of self-supervised learning in medical imaging analysis aiming at shedding the light on the recent innovation in the field. Finally, the article concludes with possible future research directions in the field.
Collapse
Affiliation(s)
- Saeed Shurrab
- Department of Computer Information Systems, Jordan University of Science and Technology, Irbid, Jordan
| | - Rehab Duwairi
- Department of Computer Information Systems, Jordan University of Science and Technology, Irbid, Jordan
| |
Collapse
|