Augmentation Pathways Network for Visual Recognition
- URL: http://arxiv.org/abs/2107.11990v1
- Date: Mon, 26 Jul 2021 06:54:53 GMT
- Title: Augmentation Pathways Network for Visual Recognition
- Authors: Yalong Bai, Mohan Zhou, Yuxiang Chen, Wei Zhang, Bowen Zhou, Tao Mei
- Abstract summary: This paper introduces Augmentation Pathways (AP) to stabilize training on a much wider range of augmentation policies.
AP tames heavy data augmentations and stably boosts performance without a careful selection among augmentation policies.
Experimental results on ImageNet benchmarks demonstrate the compatibility and effectiveness on a much wider range of augmentations.
- Score: 61.33084317147437
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data augmentation is practically helpful for visual recognition, especially
at the time of data scarcity. However, such success is only limited to quite a
few light augmentations (e.g., random crop, flip). Heavy augmentations (e.g.,
gray, grid shuffle) are either unstable or show adverse effects during
training, owing to the big gap between the original and augmented images. This
paper introduces a novel network design, noted as Augmentation Pathways (AP),
to systematically stabilize training on a much wider range of augmentation
policies. Notably, AP tames heavy data augmentations and stably boosts
performance without a careful selection among augmentation policies. Unlike
traditional single pathway, augmented images are processed in different neural
paths. The main pathway handles light augmentations, while other pathways focus
on heavy augmentations. By interacting with multiple paths in a dependent
manner, the backbone network robustly learns from shared visual patterns among
augmentations, and suppresses noisy patterns at the same time. Furthermore, we
extend AP to a homogeneous version and a heterogeneous version for high-order
scenarios, demonstrating its robustness and flexibility in practical usage.
Experimental results on ImageNet benchmarks demonstrate the compatibility and
effectiveness on a much wider range of augmentations (e.g., Crop, Gray, Grid
Shuffle, RandAugment), while consuming fewer parameters and lower computational
costs at inference time. Source code:https://github.com/ap-conv/ap-net.
Related papers
- Fourier-basis Functions to Bridge Augmentation Gap: Rethinking Frequency
Augmentation in Image Classification [3.129187821625805]
Auxiliary Fourier-basis Augmentation (AFA) is a technique targeting augmentation in the frequency domain and filling the augmentation gap left by visual augmentations.
Our results show that AFA benefits the robustness of models against common corruptions, OOD generalization, and consistency of performance of models against increasing perturbations, with negligible deficit to the standard performance of models.
arXiv Detail & Related papers (2024-03-04T11:30:02Z) - Boosting Semi-Supervised 2D Human Pose Estimation by Revisiting Data
Augmentation and Consistency Training [25.02026393037821]
We find that SSHPE can be boosted from two cores: advanced data augmentations and concise consistency training ways.
We propose to repeatedly augment unlabeled images with diverse hard augmentations, and generate multi-path predictions sequentially.
Compared to SOTA approaches, our method brings substantial improvements on public datasets.
arXiv Detail & Related papers (2024-02-18T12:27:59Z) - Soft Augmentation for Image Classification [68.71067594724663]
We propose generalizing augmentation with invariant transforms to soft augmentation.
We show that soft targets allow for more aggressive data augmentation.
We also show that soft augmentations generalize to self-supervised classification tasks.
arXiv Detail & Related papers (2022-11-09T01:04:06Z) - Masked Autoencoders are Robust Data Augmentors [90.34825840657774]
Regularization techniques like image augmentation are necessary for deep neural networks to generalize well.
We propose a novel perspective of augmentation to regularize the training process.
We show that utilizing such model-based nonlinear transformation as data augmentation can improve high-level recognition tasks.
arXiv Detail & Related papers (2022-06-10T02:41:48Z) - MSR: Making Self-supervised learning Robust to Aggressive Augmentations [98.6457801252358]
We propose a new SSL paradigm, which counteracts the impact of semantic shift by balancing the role of weak and aggressively augmented pairs.
We show that our model achieves 73.1% top-1 accuracy on ImageNet-1K with ResNet-50 for 200 epochs, which is a 2.5% improvement over BYOL.
arXiv Detail & Related papers (2022-06-04T14:27:29Z) - Improving Transferability of Representations via Augmentation-Aware
Self-Supervision [117.15012005163322]
AugSelf is an auxiliary self-supervised loss that learns the difference of augmentation parameters between two randomly augmented samples.
Our intuition is that AugSelf encourages to preserve augmentation-aware information in learned representations, which could be beneficial for their transferability.
AugSelf can easily be incorporated into recent state-of-the-art representation learning methods with a negligible additional training cost.
arXiv Detail & Related papers (2021-11-18T10:43:50Z) - Contrastive Learning with Stronger Augmentations [63.42057690741711]
We propose a general framework called Contrastive Learning with Stronger Augmentations(A) to complement current contrastive learning approaches.
Here, the distribution divergence between the weakly and strongly augmented images over the representation bank is adopted to supervise the retrieval of strongly augmented queries.
Experiments showed the information from the strongly augmented images can significantly boost the performance.
arXiv Detail & Related papers (2021-04-15T18:40:04Z) - InAugment: Improving Classifiers via Internal Augmentation [14.281619356571724]
We present a novel augmentation operation, that exploits image internal statistics.
We show improvement over state-of-the-art augmentation techniques.
We also demonstrate an increase for ResNet50 and EfficientNet-B3 top-1's accuracy on the ImageNet dataset.
arXiv Detail & Related papers (2021-04-08T15:37:21Z) - MetaAugment: Sample-Aware Data Augmentation Policy Learning [20.988767360529362]
We learn a sample-aware data augmentation policy efficiently by formulating it as a sample reweighting problem.
An augmentation policy network takes a transformation and the corresponding augmented image as inputs, and outputs a weight to adjust the augmented image loss computed by a task network.
At training stage, the task network minimizes the weighted losses of augmented training images, while the policy network minimizes the loss of the task network on a validation set via meta-learning.
arXiv Detail & Related papers (2020-12-22T15:19:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.