Improving Self-supervised Learning with Automated Unsupervised Outlier
Arbitration
- URL: http://arxiv.org/abs/2112.08132v1
- Date: Wed, 15 Dec 2021 14:05:23 GMT
- Title: Improving Self-supervised Learning with Automated Unsupervised Outlier
Arbitration
- Authors: Yu Wang and Jingyang Lin and Jingjing Zou and Yingwei Pan and Ting Yao
and Tao Mei
- Abstract summary: We introduce a lightweight latent variable model UOTA, targeting the view sampling issue for self-supervised learning.
Our method directly generalizes to many mainstream self-supervised learning approaches.
- Score: 83.29856873525674
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Our work reveals a structured shortcoming of the existing mainstream
self-supervised learning methods. Whereas self-supervised learning frameworks
usually take the prevailing perfect instance level invariance hypothesis for
granted, we carefully investigate the pitfalls behind. Particularly, we argue
that the existing augmentation pipeline for generating multiple positive views
naturally introduces out-of-distribution (OOD) samples that undermine the
learning of the downstream tasks. Generating diverse positive augmentations on
the input does not always pay off in benefiting downstream tasks. To overcome
this inherent deficiency, we introduce a lightweight latent variable model
UOTA, targeting the view sampling issue for self-supervised learning. UOTA
adaptively searches for the most important sampling region to produce views,
and provides viable choice for outlier-robust self-supervised learning
approaches. Our method directly generalizes to many mainstream self-supervised
learning approaches, regardless of the loss's nature contrastive or not. We
empirically show UOTA's advantage over the state-of-the-art self-supervised
paradigms with evident margin, which well justifies the existence of the OOD
sample issue embedded in the existing approaches. Especially, we theoretically
prove that the merits of the proposal boil down to guaranteed estimator
variance and bias reduction. Code is available: at
https://github.com/ssl-codelab/uota.
Related papers
- RODD: A Self-Supervised Approach for Robust Out-of-Distribution
Detection [12.341250124228859]
We propose a simple yet effective generalized OOD detection method independent of out-of-distribution datasets.
Our approach relies on self-supervised feature learning of the training samples, where the embeddings lie on a compact low-dimensional space.
We empirically show that a pre-trained model with self-supervised contrastive learning yields a better model for uni-dimensional feature learning in the latent space.
arXiv Detail & Related papers (2022-04-06T03:05:58Z) - Chaos is a Ladder: A New Theoretical Understanding of Contrastive
Learning via Augmentation Overlap [64.60460828425502]
We propose a new guarantee on the downstream performance of contrastive learning.
Our new theory hinges on the insight that the support of different intra-class samples will become more overlapped under aggressive data augmentations.
We propose an unsupervised model selection metric ARC that aligns well with downstream accuracy.
arXiv Detail & Related papers (2022-03-25T05:36:26Z) - Agree to Disagree: Diversity through Disagreement for Better
Transferability [54.308327969778155]
We propose D-BAT (Diversity-By-disAgreement Training), which enforces agreement among the models on the training data.
We show how D-BAT naturally emerges from the notion of generalized discrepancy.
arXiv Detail & Related papers (2022-02-09T12:03:02Z) - A Low Rank Promoting Prior for Unsupervised Contrastive Learning [108.91406719395417]
We construct a novel probabilistic graphical model that effectively incorporates the low rank promoting prior into the framework of contrastive learning.
Our hypothesis explicitly requires that all the samples belonging to the same instance class lie on the same subspace with small dimension.
Empirical evidences show that the proposed algorithm clearly surpasses the state-of-the-art approaches on multiple benchmarks.
arXiv Detail & Related papers (2021-08-05T15:58:25Z) - Unsupervised Embedding Learning from Uncertainty Momentum Modeling [37.674449317054716]
We propose a novel solution to explicitly model and explore the uncertainty of the given unlabeled learning samples.
We leverage such uncertainty modeling momentum to the learning which is helpful to tackle the outliers.
arXiv Detail & Related papers (2021-07-19T14:06:19Z) - Evading the Simplicity Bias: Training a Diverse Set of Models Discovers
Solutions with Superior OOD Generalization [93.8373619657239]
Neural networks trained with SGD were recently shown to rely preferentially on linearly-predictive features.
This simplicity bias can explain their lack of robustness out of distribution (OOD)
We demonstrate that the simplicity bias can be mitigated and OOD generalization improved.
arXiv Detail & Related papers (2021-05-12T12:12:24Z) - Joint Generative and Contrastive Learning for Unsupervised Person
Re-identification [15.486689594217273]
Recent self-supervised contrastive learning provides an effective approach for unsupervised person re-identification (ReID)
In this paper, we incorporate a Generative Adversarial Network (GAN) and a contrastive learning module into one joint training framework.
arXiv Detail & Related papers (2020-12-16T16:49:57Z) - Learning the Truth From Only One Side of the Story [58.65439277460011]
We focus on generalized linear models and show that without adjusting for this sampling bias, the model may converge suboptimally or even fail to converge to the optimal solution.
We propose an adaptive approach that comes with theoretical guarantees and show that it outperforms several existing methods empirically.
arXiv Detail & Related papers (2020-06-08T18:20:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.