Self-Supervised Visual Representation Learning via Residual Momentum
- URL: http://arxiv.org/abs/2211.09861v2
- Date: Mon, 21 Nov 2022 18:34:25 GMT
- Title: Self-Supervised Visual Representation Learning via Residual Momentum
- Authors: Trung X. Pham, Axi Niu, Zhang Kang, Sultan Rizky Madjid, Ji Woo Hong,
Daehyeok Kim, Joshua Tian Jin Tee, Chang D. Yoo
- Abstract summary: Self-supervised learning (SSL) approaches have shown promising capabilities in learning the representation from unlabeled data.
momentum-based SSL frameworks suffer from a large gap in representation between the online encoder (student) and the momentum encoder (teacher)
This paper is the first to investigate and identify this invisible gap as a bottleneck that has been overlooked in the existing SSL frameworks.
We propose "residual momentum" to directly reduce this gap to encourage the student to learn the representation as close to that of the teacher as possible.
- Score: 15.515169550346517
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Self-supervised learning (SSL) approaches have shown promising capabilities
in learning the representation from unlabeled data. Amongst them,
momentum-based frameworks have attracted significant attention. Despite being a
great success, these momentum-based SSL frameworks suffer from a large gap in
representation between the online encoder (student) and the momentum encoder
(teacher), which hinders performance on downstream tasks. This paper is the
first to investigate and identify this invisible gap as a bottleneck that has
been overlooked in the existing SSL frameworks, potentially preventing the
models from learning good representation. To solve this problem, we propose
"residual momentum" to directly reduce this gap to encourage the student to
learn the representation as close to that of the teacher as possible, narrow
the performance gap with the teacher, and significantly improve the existing
SSL. Our method is straightforward, easy to implement, and can be easily
plugged into other SSL frameworks. Extensive experimental results on numerous
benchmark datasets and diverse network architectures have demonstrated the
effectiveness of our method over the state-of-the-art contrastive learning
baselines.
Related papers
- Self-supervised contrastive learning performs non-linear system identification [2.393499494583001]
We show that self-supervised learning can perform system identification in latent space.
We propose DynCL, a framework to uncover linear, switching linear and non-linear dynamics under a non-linear observation model.
arXiv Detail & Related papers (2024-10-18T17:59:25Z) - A Survey of the Self Supervised Learning Mechanisms for Vision Transformers [5.152455218955949]
The application of self supervised learning (SSL) in vision tasks has gained significant attention.
We develop a comprehensive taxonomy of systematically classifying the SSL techniques.
We discuss the motivations behind SSL, review popular pre-training tasks, and highlight the challenges and advancements in this field.
arXiv Detail & Related papers (2024-08-30T07:38:28Z) - Reverse Engineering Self-Supervised Learning [17.720366509919167]
Self-supervised learning (SSL) is a powerful tool in machine learning.
This paper presents an in-depth empirical analysis of SSL-trained representations.
arXiv Detail & Related papers (2023-05-24T23:15:28Z) - Understanding and Improving the Role of Projection Head in
Self-Supervised Learning [77.59320917894043]
Self-supervised learning (SSL) aims to produce useful feature representations without access to human-labeled data annotations.
Current contrastive learning approaches append a parametrized projection head to the end of some backbone network to optimize the InfoNCE objective.
This raises a fundamental question: Why is a learnable projection head required if we are to discard it after training?
arXiv Detail & Related papers (2022-12-22T05:42:54Z) - Decoupled Adversarial Contrastive Learning for Self-supervised
Adversarial Robustness [69.39073806630583]
Adversarial training (AT) for robust representation learning and self-supervised learning (SSL) for unsupervised representation learning are two active research fields.
We propose a two-stage framework termed Decoupled Adversarial Contrastive Learning (DeACL)
arXiv Detail & Related papers (2022-07-22T06:30:44Z) - On Higher Adversarial Susceptibility of Contrastive Self-Supervised
Learning [104.00264962878956]
Contrastive self-supervised learning (CSL) has managed to match or surpass the performance of supervised learning in image and video classification.
It is still largely unknown if the nature of the representation induced by the two learning paradigms is similar.
We identify the uniform distribution of data representation over a unit hypersphere in the CSL representation space as the key contributor to this phenomenon.
We devise strategies that are simple, yet effective in improving model robustness with CSL training.
arXiv Detail & Related papers (2022-07-22T03:49:50Z) - ReSSL: Relational Self-Supervised Learning with Weak Augmentation [68.47096022526927]
Self-supervised learning has achieved great success in learning visual representations without data annotations.
We introduce a novel relational SSL paradigm that learns representations by modeling the relationship between different instances.
Our proposed ReSSL significantly outperforms the previous state-of-the-art algorithms in terms of both performance and training efficiency.
arXiv Detail & Related papers (2021-07-20T06:53:07Z) - Heterogeneous Contrastive Learning: Encoding Spatial Information for
Compact Visual Representations [183.03278932562438]
This paper presents an effective approach that adds spatial information to the encoding stage to alleviate the learning inconsistency between the contrastive objective and strong data augmentation operations.
We show that our approach achieves higher efficiency in visual representations and thus delivers a key message to inspire the future research of self-supervised visual representation learning.
arXiv Detail & Related papers (2020-11-19T16:26:25Z) - Adversarial Self-Supervised Learning for Semi-Supervised 3D Action
Recognition [123.62183172631443]
We present Adversarial Self-Supervised Learning (ASSL), a novel framework that tightly couples SSL and the semi-supervised scheme.
Specifically, we design an effective SSL scheme to improve the discrimination capability of learned representations for 3D action recognition.
arXiv Detail & Related papers (2020-07-12T08:01:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.