Debiased Contrastive Learning
- URL: http://arxiv.org/abs/2007.00224v3
- Date: Wed, 21 Oct 2020 06:39:24 GMT
- Title: Debiased Contrastive Learning
- Authors: Ching-Yao Chuang, Joshua Robinson, Lin Yen-Chen, Antonio Torralba,
Stefanie Jegelka
- Abstract summary: We develop a debiased contrastive objective that corrects for the sampling of same-label datapoints.
Empirically, the proposed objective consistently outperforms the state-of-the-art for representation learning in vision, language, and reinforcement learning benchmarks.
- Score: 64.98602526764599
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A prominent technique for self-supervised representation learning has been to
contrast semantically similar and dissimilar pairs of samples. Without access
to labels, dissimilar (negative) points are typically taken to be randomly
sampled datapoints, implicitly accepting that these points may, in reality,
actually have the same label. Perhaps unsurprisingly, we observe that sampling
negative examples from truly different labels improves performance, in a
synthetic setting where labels are available. Motivated by this observation, we
develop a debiased contrastive objective that corrects for the sampling of
same-label datapoints, even without knowledge of the true labels. Empirically,
the proposed objective consistently outperforms the state-of-the-art for
representation learning in vision, language, and reinforcement learning
benchmarks. Theoretically, we establish generalization bounds for the
downstream classification task.
Related papers
- Virtual Category Learning: A Semi-Supervised Learning Method for Dense
Prediction with Extremely Limited Labels [63.16824565919966]
This paper proposes to use confusing samples proactively without label correction.
A Virtual Category (VC) is assigned to each confusing sample in such a way that it can safely contribute to the model optimisation.
Our intriguing findings highlight the usage of VC learning in dense vision tasks.
arXiv Detail & Related papers (2023-12-02T16:23:52Z) - Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective [89.5370481649529]
We propose a label distribution perspective for PU learning in this paper.
Motivated by this, we propose to pursue the label distribution consistency between predicted and ground-truth label distributions.
Experiments on three benchmark datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2022-12-06T07:38:29Z) - A Theory-Driven Self-Labeling Refinement Method for Contrastive
Representation Learning [111.05365744744437]
Unsupervised contrastive learning labels crops of the same image as positives, and other image crops as negatives.
In this work, we first prove that for contrastive learning, inaccurate label assignment heavily impairs its generalization for semantic instance discrimination.
Inspired by this theory, we propose a novel self-labeling refinement approach for contrastive learning.
arXiv Detail & Related papers (2021-06-28T14:24:52Z) - Disentangling Sampling and Labeling Bias for Learning in Large-Output
Spaces [64.23172847182109]
We show that different negative sampling schemes implicitly trade-off performance on dominant versus rare labels.
We provide a unified means to explicitly tackle both sampling bias, arising from working with a subset of all labels, and labeling bias, which is inherent to the data due to label imbalance.
arXiv Detail & Related papers (2021-05-12T15:40:13Z) - MatchGAN: A Self-Supervised Semi-Supervised Conditional Generative
Adversarial Network [51.84251358009803]
We present a novel self-supervised learning approach for conditional generative adversarial networks (GANs) under a semi-supervised setting.
We perform augmentation by randomly sampling sensible labels from the label space of the few labelled examples available.
Our method surpasses the baseline with only 20% of the labelled examples used to train the baseline.
arXiv Detail & Related papers (2020-06-11T17:14:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.