Pushing the limits of self-supervised ResNets: Can we outperform
supervised learning without labels on ImageNet?
- URL: http://arxiv.org/abs/2201.05119v1
- Date: Thu, 13 Jan 2022 18:23:30 GMT
- Title: Pushing the limits of self-supervised ResNets: Can we outperform
supervised learning without labels on ImageNet?
- Authors: Nenad Tomasev, Ioana Bica, Brian McWilliams, Lars Buesing, Razvan
Pascanu, Charles Blundell, Jovana Mitrovic
- Abstract summary: ReLICv2 is first representation learning method to consistently outperform the supervised baseline in a like-for-like comparison.
We show that despite using ResNet encoders, ReLICv2 is comparable to state-of-the-art self-supervised vision transformers.
- Score: 35.98841834512082
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite recent progress made by self-supervised methods in representation
learning with residual networks, they still underperform supervised learning on
the ImageNet classification benchmark, limiting their applicability in
performance-critical settings. Building on prior theoretical insights from
Mitrovic et al., 2021, we propose ReLICv2 which combines an explicit invariance
loss with a contrastive objective over a varied set of appropriately
constructed data views. ReLICv2 achieves 77.1% top-1 classification accuracy on
ImageNet using linear evaluation with a ResNet50 architecture and 80.6% with
larger ResNet models, outperforming previous state-of-the-art self-supervised
approaches by a wide margin. Most notably, ReLICv2 is the first representation
learning method to consistently outperform the supervised baseline in a
like-for-like comparison using a range of standard ResNet architectures.
Finally we show that despite using ResNet encoders, ReLICv2 is comparable to
state-of-the-art self-supervised vision transformers.
Related papers
- ConvNeXt V2: Co-designing and Scaling ConvNets with Masked Autoencoders [104.05133094625137]
We propose a fully convolutional masked autoencoder framework and a new Global Response Normalization layer.
This co-design of self-supervised learning techniques and architectural improvement results in a new model family called ConvNeXt V2, which significantly improves the performance of pure ConvNets.
arXiv Detail & Related papers (2023-01-02T18:59:31Z) - Masked Unsupervised Self-training for Zero-shot Image Classification [98.23094305347709]
Masked Unsupervised Self-Training (MUST) is a new approach which leverages two different and complimentary sources of supervision: pseudo-labels and raw images.
MUST improves upon CLIP by a large margin and narrows the performance gap between unsupervised and supervised classification.
arXiv Detail & Related papers (2022-06-07T02:03:06Z) - m-RevNet: Deep Reversible Neural Networks with Momentum [25.609808975649624]
We propose a reversible neural network, termed as m-RevNet, that is characterized by inserting momentum update to residual blocks.
For certain learning scenarios, we analytically and empirically reveal that our m-RevNet succeeds while standard ResNet fails.
arXiv Detail & Related papers (2021-08-12T17:14:32Z) - Revisiting ResNets: Improved Training and Scaling Strategies [54.0162571976267]
Training and scaling strategies may matter more than architectural changes, and the resulting ResNets match recent state-of-the-art models.
We show that the best performing scaling strategy depends on the training regime.
We design a family of ResNet architectures, ResNet-RS, which are 1.7x - 2.7x faster than EfficientNets on TPUs, while achieving similar accuracies on ImageNet.
arXiv Detail & Related papers (2021-03-13T00:18:19Z) - Supervised Contrastive Learning [42.27949000093086]
We extend the self-supervised batch contrastive approach to the fully-supervised setting.
We analyze two possible versions of the supervised contrastive (SupCon) loss, identifying the best-performing formulation of the loss.
On ResNet-200, we achieve top-1 accuracy of 81.4% on the ImageNet dataset, which is 0.8% above the best number reported for this architecture.
arXiv Detail & Related papers (2020-04-23T17:58:56Z) - ReActNet: Towards Precise Binary Neural Network with Generalized
Activation Functions [76.05981545084738]
We propose several ideas for enhancing a binary network to close its accuracy gap from real-valued networks without incurring any additional computational cost.
We first construct a baseline network by modifying and binarizing a compact real-valued network with parameter-free shortcuts.
We show that the proposed ReActNet outperforms all the state-of-the-arts by a large margin.
arXiv Detail & Related papers (2020-03-07T02:12:02Z) - A Simple Framework for Contrastive Learning of Visual Representations [116.37752766922407]
This paper presents SimCLR: a simple framework for contrastive learning of visual representations.
We show that composition of data augmentations plays a critical role in defining effective predictive tasks.
We are able to considerably outperform previous methods for self-supervised and semi-supervised learning on ImageNet.
arXiv Detail & Related papers (2020-02-13T18:50:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.