Directional Self-supervised Learning for Risky Image Augmentations
- URL: http://arxiv.org/abs/2110.13555v1
- Date: Tue, 26 Oct 2021 10:33:25 GMT
- Title: Directional Self-supervised Learning for Risky Image Augmentations
- Authors: Yalong Bai, Yifan Yang, Wei Zhang, Tao Mei
- Abstract summary: We propose a directional self-supervised learning paradigm (DSSL) for robust augmentation policies.
DSSL can be easily implemented with a few lines of Pseudocode and is highly flexible to popular self-supervised learning frameworks.
- Score: 54.43314754436954
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Only a few cherry-picked robust augmentation policies are beneficial to
standard self-supervised image representation learning, despite the large
augmentation family. In this paper, we propose a directional self-supervised
learning paradigm (DSSL), which is compatible with significantly more
augmentations. Specifically, we adapt risky augmentation policies after
standard views augmented by robust augmentations, to generate harder risky view
(RV). The risky view usually has a higher deviation from the original image
than the standard robust view (SV). Unlike previous methods equally pairing all
augmented views for symmetrical self-supervised training to maximize their
similarities, DSSL treats augmented views of the same instance as a partially
ordered set (SV$\leftrightarrow $SV, SV$\leftarrow$RV), and then equips
directional objective functions respecting to the derived relationships among
views. DSSL can be easily implemented with a few lines of Pseudocode and is
highly flexible to popular self-supervised learning frameworks, including
SimCLR, SimSiam, BYOL. The extensive experimental results on CIFAR and ImageNet
demonstrated that DSSL can stably improve these frameworks with compatibility
to a wider range of augmentations.
Related papers
- Gen-SIS: Generative Self-augmentation Improves Self-supervised Learning [52.170253590364545]
Gen-SIS is a diffusion-based augmentation technique trained exclusively on unlabeled image data.
We show that these self-augmentations', i.e. generative augmentations based on the vanilla SSL encoder embeddings, facilitate the training of a stronger SSL encoder.
arXiv Detail & Related papers (2024-12-02T16:20:59Z) - Hierarchical Consistent Contrastive Learning for Skeleton-Based Action
Recognition with Growing Augmentations [33.68311764817763]
We propose a general hierarchical consistent contrastive learning framework (HiCLR) for skeleton-based action recognition.
Specifically, we first design a gradual growing augmentation policy to generate multiple ordered positive pairs.
Then, an asymmetric loss is proposed to enforce the hierarchical consistency via a directional clustering operation.
arXiv Detail & Related papers (2022-11-24T08:09:50Z) - RegCLR: A Self-Supervised Framework for Tabular Representation Learning
in the Wild [1.5187073409911216]
We introduce RegCLR, a new self-supervised framework that combines contrastive and regularized methods.
RegCLR is compatible with the standard Vision Transformer architecture.
Several real-world table recognition scenarios have been shown to benefit greatly from the representations learned from this new framework.
arXiv Detail & Related papers (2022-11-02T14:39:56Z) - Does Self-supervised Learning Really Improve Reinforcement Learning from
Pixels? [42.404871049605084]
We extend the contrastive reinforcement learning framework (e.g., CURL) that jointly optimize SSL and RL losses.
Our observations suggest that the existing SSL framework for RL fails to bring meaningful improvement over the baselines.
We evaluate the approach in multiple different environments including a real-world robot environment.
arXiv Detail & Related papers (2022-06-10T17:59:30Z) - Masked Autoencoders are Robust Data Augmentors [90.34825840657774]
Regularization techniques like image augmentation are necessary for deep neural networks to generalize well.
We propose a novel perspective of augmentation to regularize the training process.
We show that utilizing such model-based nonlinear transformation as data augmentation can improve high-level recognition tasks.
arXiv Detail & Related papers (2022-06-10T02:41:48Z) - Residual Relaxation for Multi-view Representation Learning [64.40142301026805]
Multi-view methods learn by aligning multiple views of the same image.
Some useful augmentations, such as image rotation, are harmful for multi-view methods because they cause a semantic shift.
We develop a generic approach, Pretext-aware Residual Relaxation (Prelax), that relaxes the exact alignment.
arXiv Detail & Related papers (2021-10-28T17:57:17Z) - Augmentation Pathways Network for Visual Recognition [61.33084317147437]
This paper introduces Augmentation Pathways (AP) to stabilize training on a much wider range of augmentation policies.
AP tames heavy data augmentations and stably boosts performance without a careful selection among augmentation policies.
Experimental results on ImageNet benchmarks demonstrate the compatibility and effectiveness on a much wider range of augmentations.
arXiv Detail & Related papers (2021-07-26T06:54:53Z) - SECANT: Self-Expert Cloning for Zero-Shot Generalization of Visual
Policies [87.78260740602674]
Generalization has been a long-standing challenge for reinforcement learning (RL)
In this work, we consider robust policy learning which targets zero-shot generalization to unseen visual environments with large distributional shift.
We propose SECANT, a novel self-expert cloning technique that leverages image augmentation in two stages to decouple robust representation learning from policy optimization.
arXiv Detail & Related papers (2021-06-17T17:28:18Z) - Contrastive Learning with Stronger Augmentations [63.42057690741711]
We propose a general framework called Contrastive Learning with Stronger Augmentations(A) to complement current contrastive learning approaches.
Here, the distribution divergence between the weakly and strongly augmented images over the representation bank is adopted to supervise the retrieval of strongly augmented queries.
Experiments showed the information from the strongly augmented images can significantly boost the performance.
arXiv Detail & Related papers (2021-04-15T18:40:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.