Online pseudo labeling for polyp segmentation with momentum networks
- URL: http://arxiv.org/abs/2209.14599v1
- Date: Thu, 29 Sep 2022 07:33:54 GMT
- Title: Online pseudo labeling for polyp segmentation with momentum networks
- Authors: Toan Pham Van, Linh Bao Doan, Thanh Tung Nguyen, Duc Trung Tran, Quan
Van Nguyen, Dinh Viet Sang
- Abstract summary: In semi-supervised learning, the quality of labels plays a crucial role in model performance.
We present a new pseudo labeling strategy that enhances the quality of pseudo labels used for training student networks.
Our results surpass common practice by 3% and even approach fully-supervised results on some datasets.
- Score: 5.920947681019466
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Semantic segmentation is an essential task in developing medical image
diagnosis systems. However, building an annotated medical dataset is expensive.
Thus, semi-supervised methods are significant in this circumstance. In
semi-supervised learning, the quality of labels plays a crucial role in model
performance. In this work, we present a new pseudo labeling strategy that
enhances the quality of pseudo labels used for training student networks. We
follow the multi-stage semi-supervised training approach, which trains a
teacher model on a labeled dataset and then uses the trained teacher to render
pseudo labels for student training. By doing so, the pseudo labels will be
updated and more precise as training progress. The key difference between
previous and our methods is that we update the teacher model during the student
training process. So the quality of pseudo labels is improved during the
student training process. We also propose a simple but effective strategy to
enhance the quality of pseudo labels using a momentum model -- a slow copy
version of the original model during training. By applying the momentum model
combined with re-rendering pseudo labels during student training, we achieved
an average of 84.1% Dice Score on five datasets (i.e., Kvarsir, CVC-ClinicDB,
ETIS-LaribPolypDB, CVC-ColonDB, and CVC-300) with only 20% of the dataset used
as labeled data. Our results surpass common practice by 3% and even approach
fully-supervised results on some datasets. Our source code and pre-trained
models are available at https://github.com/sun-asterisk-research/online
learning ssl
Related papers
- Pseudo Labels for Single Positive Multi-Label Learning [0.0]
Single positive multi-label (SPML) learning is a cost-effective solution, where models are trained on a single positive label per image.
In this work, we propose a method to turn single positive data into fully-labeled data: Pseudo Multi-Labels.
arXiv Detail & Related papers (2023-06-01T17:21:42Z) - Doubly Robust Self-Training [46.168395767948965]
We introduce doubly robust self-training, a novel semi-supervised algorithm.
We demonstrate the superiority of the doubly robust loss over the standard self-training baseline.
arXiv Detail & Related papers (2023-06-01T00:57:16Z) - You Only Need One Thing One Click: Self-Training for Weakly Supervised
3D Scene Understanding [107.06117227661204]
We propose One Thing One Click'', meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our model can be compatible to 3D instance segmentation equipped with a point-clustering strategy.
arXiv Detail & Related papers (2023-03-26T13:57:00Z) - Large-Scale Pre-training for Person Re-identification with Noisy Labels [125.49696935852634]
We develop a large-scale Pre-training framework utilizing Noisy Labels (PNL)
In principle, joint learning of these three modules not only clusters similar examples to one prototype, but also rectifies noisy labels based on the prototype assignment.
This simple pre-training task provides a scalable way to learn SOTA Re-ID representations from scratch on "LUPerson-NL" without bells and whistles.
arXiv Detail & Related papers (2022-03-30T17:59:58Z) - Cross-Model Pseudo-Labeling for Semi-Supervised Action Recognition [98.25592165484737]
We propose a more effective pseudo-labeling scheme, called Cross-Model Pseudo-Labeling (CMPL)
CMPL achieves $17.6%$ and $25.1%$ Top-1 accuracy on Kinetics-400 and UCF-101 using only the RGB modality and $1%$ labeled data, respectively.
arXiv Detail & Related papers (2021-12-17T18:59:41Z) - Weakly Supervised Pseudo-Label assisted Learning for ALS Point Cloud
Semantic Segmentation [1.4620086904601473]
Competitive point cloud results usually rely on a large amount of labeled data.
In this study, we propose a pseudo-labeling strategy to obtain accurate results with limited ground truth.
arXiv Detail & Related papers (2021-05-05T08:07:21Z) - SLADE: A Self-Training Framework For Distance Metric Learning [75.54078592084217]
We present a self-training framework, SLADE, to improve retrieval performance by leveraging additional unlabeled data.
We first train a teacher model on the labeled data and use it to generate pseudo labels for the unlabeled data.
We then train a student model on both labels and pseudo labels to generate final feature embeddings.
arXiv Detail & Related papers (2020-11-20T08:26:10Z) - Self-training For Pre-training Language Models [0.5139874302398955]
In industry NLP applications, we have large amounts of data produced by users or customers.
Our learning framework is based on this large amounts of unlabel data.
arXiv Detail & Related papers (2020-11-18T01:35:01Z) - Uncertainty-aware Self-training for Text Classification with Few Labels [54.13279574908808]
We study self-training as one of the earliest semi-supervised learning approaches to reduce the annotation bottleneck.
We propose an approach to improve self-training by incorporating uncertainty estimates of the underlying neural network.
We show our methods leveraging only 20-30 labeled samples per class for each task for training and for validation can perform within 3% of fully supervised pre-trained language models.
arXiv Detail & Related papers (2020-06-27T08:13:58Z) - Improving Semantic Segmentation via Self-Training [75.07114899941095]
We show that we can obtain state-of-the-art results using a semi-supervised approach, specifically a self-training paradigm.
We first train a teacher model on labeled data, and then generate pseudo labels on a large set of unlabeled data.
Our robust training framework can digest human-annotated and pseudo labels jointly and achieve top performances on Cityscapes, CamVid and KITTI datasets.
arXiv Detail & Related papers (2020-04-30T17:09:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.