SemiReward: A General Reward Model for Semi-supervised Learning
- URL: http://arxiv.org/abs/2310.03013v2
- Date: Tue, 20 Feb 2024 16:02:18 GMT
- Title: SemiReward: A General Reward Model for Semi-supervised Learning
- Authors: Siyuan Li, Weiyang Jin, Zedong Wang, Fang Wu, Zicheng Liu, Cheng Tan,
Stan Z. Li
- Abstract summary: Semi-supervised learning (SSL) has witnessed great progress with various improvements in the self-training framework with pseudo labeling.
Main challenge is how to distinguish high-quality pseudo labels against the confirmation bias.
We propose a Semi-supervised Reward framework (SemiReward) that predicts reward scores to evaluate and filter out high-quality pseudo labels.
- Score: 58.47299780978101
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Semi-supervised learning (SSL) has witnessed great progress with various
improvements in the self-training framework with pseudo labeling. The main
challenge is how to distinguish high-quality pseudo labels against the
confirmation bias. However, existing pseudo-label selection strategies are
limited to pre-defined schemes or complex hand-crafted policies specially
designed for classification, failing to achieve high-quality labels, fast
convergence, and task versatility simultaneously. To these ends, we propose a
Semi-supervised Reward framework (SemiReward) that predicts reward scores to
evaluate and filter out high-quality pseudo labels, which is pluggable to
mainstream SSL methods in wide task types and scenarios. To mitigate
confirmation bias, SemiReward is trained online in two stages with a generator
model and subsampling strategy. With classification and regression tasks on 13
standard SSL benchmarks across three modalities, extensive experiments verify
that SemiReward achieves significant performance gains and faster convergence
speeds upon Pseudo Label, FlexMatch, and Free/SoftMatch. Code and models are
available at https://github.com/Westlake-AI/SemiReward.
Related papers
- SAM as the Guide: Mastering Pseudo-Label Refinement in Semi-Supervised Referring Expression Segmentation [66.92696817276288]
SemiRES is a semi-supervised framework that effectively leverages a combination of labeled and unlabeled data to perform RES.
SemiRES incorporates the Segment Anything Model (SAM), renowned for its precise boundary demarcation.
In instances where a precise mask cannot be matched from the available candidates, we develop the Pixel-Wise Adjustment (PWA) strategy.
arXiv Detail & Related papers (2024-06-03T15:42:30Z) - Semi-Supervised Class-Agnostic Motion Prediction with Pseudo Label
Regeneration and BEVMix [59.55173022987071]
We study the potential of semi-supervised learning for class-agnostic motion prediction.
Our framework adopts a consistency-based self-training paradigm, enabling the model to learn from unlabeled data.
Our method exhibits comparable performance to weakly and some fully supervised methods.
arXiv Detail & Related papers (2023-12-13T09:32:50Z) - RankMatch: A Novel Approach to Semi-Supervised Label Distribution
Learning Leveraging Inter-label Correlations [52.549807652527306]
This paper introduces RankMatch, an innovative approach for Semi-Supervised Label Distribution Learning (SSLDL)
RankMatch effectively utilizes a small number of labeled examples in conjunction with a larger quantity of unlabeled data.
We establish a theoretical generalization bound for RankMatch, and through extensive experiments, demonstrate its superiority in performance against existing SSLDL methods.
arXiv Detail & Related papers (2023-12-11T12:47:29Z) - Improving Semi-Supervised Semantic Segmentation with Dual-Level Siamese Structure Network [7.438140196173472]
We propose a dual-level Siamese structure network (DSSN) for pixel-wise contrastive learning.
We introduce a novel class-aware pseudo-label selection strategy for weak-to-strong supervision.
Our proposed method achieves state-of-the-art results on two datasets.
arXiv Detail & Related papers (2023-07-26T03:30:28Z) - On Non-Random Missing Labels in Semi-Supervised Learning [114.62655062520425]
Semi-Supervised Learning (SSL) is fundamentally a missing label problem.
We explicitly incorporate "class" into SSL.
Our method not only significantly outperforms existing baselines but also surpasses other label bias removal SSL methods.
arXiv Detail & Related papers (2022-06-29T22:01:29Z) - An Iterative Co-Training Transductive Framework for Zero Shot Learning [24.401200814880124]
We introduce an iterative co-training framework which contains two different base ZSL models and an exchanging module.
At each iteration, the two different ZSL models are co-trained to separately predict pseudo labels for the unseen-class samples.
Our framework can gradually boost the ZSL performance by fully exploiting the potential complementarity of the two models' classification capabilities.
arXiv Detail & Related papers (2022-03-30T04:08:44Z) - CLS: Cross Labeling Supervision for Semi-Supervised Learning [9.929229055862491]
Cross Labeling Supervision ( CLS) is a framework that generalizes the typical pseudo-labeling process.
CLS allows the creation of both pseudo and complementary labels to support both positive and negative learning.
arXiv Detail & Related papers (2022-02-17T08:09:40Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.