What Makes for Good Views for Contrastive Learning?
- URL: http://arxiv.org/abs/2005.10243v3
- Date: Fri, 18 Dec 2020 10:01:34 GMT
- Title: What Makes for Good Views for Contrastive Learning?
- Authors: Yonglong Tian, Chen Sun, Ben Poole, Dilip Krishnan, Cordelia Schmid,
Phillip Isola
- Abstract summary: We argue that we should reduce the mutual information (MI) between views while keeping task-relevant information intact.
We devise unsupervised and semi-supervised frameworks that learn effective views by aiming to reduce their MI.
As a by-product, we achieve a new state-of-the-art accuracy on unsupervised pre-training for ImageNet classification.
- Score: 90.49736973404046
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contrastive learning between multiple views of the data has recently achieved
state of the art performance in the field of self-supervised representation
learning. Despite its success, the influence of different view choices has been
less studied. In this paper, we use theoretical and empirical analysis to
better understand the importance of view selection, and argue that we should
reduce the mutual information (MI) between views while keeping task-relevant
information intact. To verify this hypothesis, we devise unsupervised and
semi-supervised frameworks that learn effective views by aiming to reduce their
MI. We also consider data augmentation as a way to reduce MI, and show that
increasing data augmentation indeed leads to decreasing MI and improves
downstream classification accuracy. As a by-product, we achieve a new
state-of-the-art accuracy on unsupervised pre-training for ImageNet
classification ($73\%$ top-1 linear readout with a ResNet-50). In addition,
transferring our models to PASCAL VOC object detection and COCO instance
segmentation consistently outperforms supervised pre-training.
Code:http://github.com/HobbitLong/PyContrast
Related papers
- Regularized Contrastive Partial Multi-view Outlier Detection [76.77036536484114]
We propose a novel method named Regularized Contrastive Partial Multi-view Outlier Detection (RCPMOD)
In this framework, we utilize contrastive learning to learn view-consistent information and distinguish outliers by the degree of consistency.
Experimental results on four benchmark datasets demonstrate that our proposed approach could outperform state-of-the-art competitors.
arXiv Detail & Related papers (2024-08-02T14:34:27Z) - MVEB: Self-Supervised Learning with Multi-View Entropy Bottleneck [53.44358636312935]
Self-supervised approaches regard two views of an image as both the input and the self-supervised signals.
Recent studies show that discarding superfluous information not shared between the views can improve generalization.
We propose an objective multi-view entropy bottleneck (MVEB) to learn minimal sufficient representation effectively.
arXiv Detail & Related papers (2024-03-28T00:50:02Z) - Meta Co-Training: Two Views are Better than One [4.050257210426548]
We present Meta Co-Training which is an extension of the successful Meta Pseudo Labels approach to two views.
Our method achieves new state-of-the-art performance on ImageNet-10% with very few training resources.
arXiv Detail & Related papers (2023-11-29T21:11:58Z) - Masked Unsupervised Self-training for Zero-shot Image Classification [98.23094305347709]
Masked Unsupervised Self-Training (MUST) is a new approach which leverages two different and complimentary sources of supervision: pseudo-labels and raw images.
MUST improves upon CLIP by a large margin and narrows the performance gap between unsupervised and supervised classification.
arXiv Detail & Related papers (2022-06-07T02:03:06Z) - Contrastive Learning with Boosted Memorization [36.957895270908324]
Self-supervised learning has achieved a great success in the representation learning of visual and textual data.
Recent attempts to consider self-supervised long-tailed learning are made by rebalancing in the loss perspective or the model perspective.
We propose a novel Boosted Contrastive Learning (BCL) method to enhance the long-tailed learning in the label-unaware context.
arXiv Detail & Related papers (2022-05-25T11:54:22Z) - Revisiting Contrastive Methods for Unsupervised Learning of Visual
Representations [78.12377360145078]
Contrastive self-supervised learning has outperformed supervised pretraining on many downstream tasks like segmentation and object detection.
In this paper, we first study how biases in the dataset affect existing methods.
We show that current contrastive approaches work surprisingly well across: (i) object- versus scene-centric, (ii) uniform versus long-tailed and (iii) general versus domain-specific datasets.
arXiv Detail & Related papers (2021-06-10T17:59:13Z) - SEED: Self-supervised Distillation For Visual Representation [34.63488756535054]
We propose a new learning paradigm, named SElf-SupErvised Distillation (SEED), to transfer its representational knowledge into a smaller architecture (as Student) in a self-supervised fashion.
We show that SEED dramatically boosts the performance of small networks on downstream tasks.
arXiv Detail & Related papers (2021-01-12T20:04:50Z) - Memory-Augmented Relation Network for Few-Shot Learning [114.47866281436829]
In this work, we investigate a new metric-learning method, Memory-Augmented Relation Network (MRN)
In MRN, we choose the samples that are visually similar from the working context, and perform weighted information propagation to attentively aggregate helpful information from chosen ones to enhance its representation.
We empirically demonstrate that MRN yields significant improvement over its ancestor and achieves competitive or even better performance when compared with other few-shot learning approaches.
arXiv Detail & Related papers (2020-05-09T10:09:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.