OPERA: Omni-Supervised Representation Learning with Hierarchical
Supervisions
- URL: http://arxiv.org/abs/2210.05557v1
- Date: Tue, 11 Oct 2022 15:51:31 GMT
- Title: OPERA: Omni-Supervised Representation Learning with Hierarchical
Supervisions
- Authors: Chengkun Wang, Wenzhao Zheng, Zheng Zhu, Jie Zhou, Jiwen Lu
- Abstract summary: We propose Omni-suPErvised Representation leArning with hierarchical supervisions (OPERA) as a solution.
We extract a set of hierarchical proxy representations for each image and impose self and full supervisions on the corresponding proxy representations.
Experiments on both convolutional neural networks and vision transformers demonstrate the superiority of OPERA in image classification, segmentation, and object detection.
- Score: 94.31804364707575
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The pretrain-finetune paradigm in modern computer vision facilitates the
success of self-supervised learning, which tends to achieve better
transferability than supervised learning. However, with the availability of
massive labeled data, a natural question emerges: how to train a better model
with both self and full supervision signals? In this paper, we propose
Omni-suPErvised Representation leArning with hierarchical supervisions (OPERA)
as a solution. We provide a unified perspective of supervisions from labeled
and unlabeled data and propose a unified framework of fully supervised and
self-supervised learning. We extract a set of hierarchical proxy
representations for each image and impose self and full supervisions on the
corresponding proxy representations. Extensive experiments on both
convolutional neural networks and vision transformers demonstrate the
superiority of OPERA in image classification, segmentation, and object
detection. Code is available at: https://github.com/wangck20/OPERA.
Related papers
- MOCA: Self-supervised Representation Learning by Predicting Masked Online Codebook Assignments [72.6405488990753]
Self-supervised learning can be used for mitigating the greedy needs of Vision Transformer networks.
We propose a single-stage and standalone method, MOCA, which unifies both desired properties.
We achieve new state-of-the-art results on low-shot settings and strong experimental results in various evaluation protocols.
arXiv Detail & Related papers (2023-07-18T15:46:20Z) - Unsupervised Learning of Structured Representations via Closed-Loop
Transcription [21.78655495464155]
This paper proposes an unsupervised method for learning a unified representation that serves both discriminative and generative purposes.
We show that a unified representation can enjoy the mutual benefits of having both.
These structured representations enable classification close to state-of-the-art unsupervised discriminative representations.
arXiv Detail & Related papers (2022-10-30T09:09:05Z) - Pretraining the Vision Transformer using self-supervised methods for
vision based Deep Reinforcement Learning [0.0]
We study pretraining a Vision Transformer using several state-of-the-art self-supervised methods and assess the quality of the learned representations.
Our results show that all methods are effective in learning useful representations and avoiding representational collapse.
The encoder pretrained with the temporal order verification task shows the best results across all experiments.
arXiv Detail & Related papers (2022-09-22T10:18:59Z) - Masked Unsupervised Self-training for Zero-shot Image Classification [98.23094305347709]
Masked Unsupervised Self-Training (MUST) is a new approach which leverages two different and complimentary sources of supervision: pseudo-labels and raw images.
MUST improves upon CLIP by a large margin and narrows the performance gap between unsupervised and supervised classification.
arXiv Detail & Related papers (2022-06-07T02:03:06Z) - Curious Representation Learning for Embodied Intelligence [81.21764276106924]
Self-supervised representation learning has achieved remarkable success in recent years.
Yet to build truly intelligent agents, we must construct representation learning algorithms that can learn from environments.
We propose a framework, curious representation learning, which jointly learns a reinforcement learning policy and a visual representation model.
arXiv Detail & Related papers (2021-05-03T17:59:20Z) - Can Semantic Labels Assist Self-Supervised Visual Representation
Learning? [194.1681088693248]
We present a new algorithm named Supervised Contrastive Adjustment in Neighborhood (SCAN)
In a series of downstream tasks, SCAN achieves superior performance compared to previous fully-supervised and self-supervised methods.
Our study reveals that semantic labels are useful in assisting self-supervised methods, opening a new direction for the community.
arXiv Detail & Related papers (2020-11-17T13:25:00Z) - SHERLock: Self-Supervised Hierarchical Event Representation Learning [22.19386609894017]
We propose a model that learns temporal representations from long-horizon visual demonstration data.
Our method produces a hierarchy of representations that align more closely with ground-truth human-annotated events.
arXiv Detail & Related papers (2020-10-06T09:04:01Z) - Self-Supervised Learning for Large-Scale Unsupervised Image Clustering [8.142434527938535]
We propose a simple scheme for unsupervised classification based on self-supervised representations.
We evaluate the proposed approach with several recent self-supervised methods showing that it achieves competitive results for ImageNet classification.
We suggest adding the unsupervised evaluation to a set of standard benchmarks for self-supervised learning.
arXiv Detail & Related papers (2020-08-24T10:39:19Z) - Self-Supervised Viewpoint Learning From Image Collections [116.56304441362994]
We propose a novel learning framework which incorporates an analysis-by-synthesis paradigm to reconstruct images in a viewpoint aware manner.
We show that our approach performs competitively to fully-supervised approaches for several object categories like human faces, cars, buses, and trains.
arXiv Detail & Related papers (2020-04-03T22:01:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.