Amortised Invariance Learning for Contrastive Self-Supervision
- URL: http://arxiv.org/abs/2302.12712v2
- Date: Mon, 3 Apr 2023 18:08:18 GMT
- Title: Amortised Invariance Learning for Contrastive Self-Supervision
- Authors: Ruchika Chavhan, Henry Gouk, Jan Stuehmer, Calum Heggan, Mehrdad
Yaghoobi, Timothy Hospedales
- Abstract summary: We introduce the notion of amortised invariance learning for contrastive self supervision.
We show that our amortised features provide a reliable way to learn diverse downstream tasks with different invariance requirements.
This provides an exciting perspective that opens up new horizons in the field of general purpose representation learning.
- Score: 11.042648980854485
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Contrastive self-supervised learning methods famously produce high quality
transferable representations by learning invariances to different data
augmentations. Invariances established during pre-training can be interpreted
as strong inductive biases. However these may or may not be helpful, depending
on if they match the invariance requirements of downstream tasks or not. This
has led to several attempts to learn task-specific invariances during
pre-training, however, these methods are highly compute intensive and tedious
to train. We introduce the notion of amortised invariance learning for
contrastive self supervision. In the pre-training stage, we parameterize the
feature extractor by differentiable invariance hyper-parameters that control
the invariances encoded by the representation. Then, for any downstream task,
both linear readout and task-specific invariance requirements can be
efficiently and effectively learned by gradient-descent. We evaluate the notion
of amortised invariances for contrastive learning over two different
modalities: vision and audio, on two widely-used contrastive learning methods
in vision: SimCLR and MoCo-v2 with popular architectures like ResNets and
Vision Transformers, and SimCLR with ResNet-18 for audio. We show that our
amortised features provide a reliable way to learn diverse downstream tasks
with different invariance requirements, while using a single feature and
avoiding task-specific pre-training. This provides an exciting perspective that
opens up new horizons in the field of general purpose representation learning.
Related papers
- Understanding the Role of Invariance in Transfer Learning [9.220104991339104]
Transfer learning is a powerful technique for knowledge-sharing between different tasks.
Recent work has found that the representations of models with certain invariances, such as to adversarial input perturbations, achieve higher performance on downstream tasks.
arXiv Detail & Related papers (2024-07-05T07:53:52Z) - MT-SLVR: Multi-Task Self-Supervised Learning for Transformation
In(Variant) Representations [2.94944680995069]
We propose a multi-task self-supervised framework (MT-SLVR) that learns both variant and invariant features in a parameter-efficient manner.
We evaluate our approach on few-shot classification tasks drawn from a variety of audio domains and demonstrate improved classification performance.
arXiv Detail & Related papers (2023-05-29T09:10:50Z) - CIPER: Combining Invariant and Equivariant Representations Using
Contrastive and Predictive Learning [6.117084972237769]
We introduce Contrastive Invariant and Predictive Equivariant Representation learning (CIPER)
CIPER comprises both invariant and equivariant learning objectives using one shared encoder and two different output heads on top of the encoder.
We evaluate our method on static image tasks and time-augmented image datasets.
arXiv Detail & Related papers (2023-02-05T07:50:46Z) - Multi-task Bias-Variance Trade-off Through Functional Constraints [102.64082402388192]
Multi-task learning aims to acquire a set of functions that perform well for diverse tasks.
In this paper we draw intuition from the two extreme learning scenarios -- a single function for all tasks, and a task-specific function that ignores the other tasks.
We introduce a constrained learning formulation that enforces domain specific solutions to a central function.
arXiv Detail & Related papers (2022-10-27T16:06:47Z) - Task Formulation Matters When Learning Continually: A Case Study in
Visual Question Answering [58.82325933356066]
Continual learning aims to train a model incrementally on a sequence of tasks without forgetting previous knowledge.
We present a detailed study of how different settings affect performance for Visual Question Answering.
arXiv Detail & Related papers (2022-09-30T19:12:58Z) - HyperInvariances: Amortizing Invariance Learning [10.189246340672245]
Invariance learning is expensive and data intensive for popular neural architectures.
We introduce the notion of amortizing invariance learning.
This framework can identify appropriate invariances in different downstream tasks and lead to comparable or better test performance.
arXiv Detail & Related papers (2022-07-17T21:40:37Z) - Agree to Disagree: Diversity through Disagreement for Better
Transferability [54.308327969778155]
We propose D-BAT (Diversity-By-disAgreement Training), which enforces agreement among the models on the training data.
We show how D-BAT naturally emerges from the notion of generalized discrepancy.
arXiv Detail & Related papers (2022-02-09T12:03:02Z) - Why Do Self-Supervised Models Transfer? Investigating the Impact of
Invariance on Downstream Tasks [79.13089902898848]
Self-supervised learning is a powerful paradigm for representation learning on unlabelled images.
We show that different tasks in computer vision require features to encode different (in)variances.
arXiv Detail & Related papers (2021-11-22T18:16:35Z) - What Should Not Be Contrastive in Contrastive Learning [110.14159883496859]
We introduce a contrastive learning framework which does not require prior knowledge of specific, task-dependent invariances.
Our model learns to capture varying and invariant factors for visual representations by constructing separate embedding spaces.
We use a multi-head network with a shared backbone which captures information across each augmentation and alone outperforms all baselines on downstream tasks.
arXiv Detail & Related papers (2020-08-13T03:02:32Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.