SAGE: Saliency-Guided Mixup with Optimal Rearrangements
- URL: http://arxiv.org/abs/2211.00113v1
- Date: Mon, 31 Oct 2022 19:45:21 GMT
- Title: SAGE: Saliency-Guided Mixup with Optimal Rearrangements
- Authors: Avery Ma, Nikita Dvornik, Ran Zhang, Leila Pishdad, Konstantinos G.
Derpanis, Afsaneh Fazly
- Abstract summary: Saliency-Guided Mixup with Optimal Rearrangements (SAGE)
SAGE creates new training examples by rearranging and mixing image pairs using visual saliency as guidance.
We demonstrate on CIFAR-10 and CIFAR-100 that SAGE achieves better or comparable performance to the state of the art while being more efficient.
- Score: 22.112463794733188
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data augmentation is a key element for training accurate models by reducing
overfitting and improving generalization. For image classification, the most
popular data augmentation techniques range from simple photometric and
geometrical transformations, to more complex methods that use visual saliency
to craft new training examples. As augmentation methods get more complex, their
ability to increase the test accuracy improves, yet, such methods become
cumbersome, inefficient and lead to poor out-of-domain generalization, as we
show in this paper. This motivates a new augmentation technique that allows for
high accuracy gains while being simple, efficient (i.e., minimal computation
overhead) and generalizable. To this end, we introduce Saliency-Guided Mixup
with Optimal Rearrangements (SAGE), which creates new training examples by
rearranging and mixing image pairs using visual saliency as guidance. By
explicitly leveraging saliency, SAGE promotes discriminative foreground objects
and produces informative new images useful for training. We demonstrate on
CIFAR-10 and CIFAR-100 that SAGE achieves better or comparable performance to
the state of the art while being more efficient. Additionally, evaluations in
the out-of-distribution setting, and few-shot learning on mini-ImageNet, show
that SAGE achieves improved generalization performance without trading off
robustness.
Related papers
- Improving Visual Representation Learning through Perceptual
Understanding [0.0]
We present an extension to masked autoencoders (MAE) which improves on the representations learnt by the model by explicitly encouraging the learning of higher scene-level features.
We achieve 78.1% top-1 accuracy linear probing on ImageNet-1K and up to 88.1% when fine-tuning, with similar results for other downstream tasks.
arXiv Detail & Related papers (2022-12-30T00:59:46Z) - Soft Augmentation for Image Classification [68.71067594724663]
We propose generalizing augmentation with invariant transforms to soft augmentation.
We show that soft targets allow for more aggressive data augmentation.
We also show that soft augmentations generalize to self-supervised classification tasks.
arXiv Detail & Related papers (2022-11-09T01:04:06Z) - Masked Autoencoders are Robust Data Augmentors [90.34825840657774]
Regularization techniques like image augmentation are necessary for deep neural networks to generalize well.
We propose a novel perspective of augmentation to regularize the training process.
We show that utilizing such model-based nonlinear transformation as data augmentation can improve high-level recognition tasks.
arXiv Detail & Related papers (2022-06-10T02:41:48Z) - Contrastive Learning Rivals Masked Image Modeling in Fine-tuning via
Feature Distillation [42.37533586611174]
Masked image modeling (MIM) learns representations with remarkably good fine-tuning performances.
In this paper, we show that the inferior fine-tuning performance of pre-training approaches can be significantly improved by a simple post-processing.
arXiv Detail & Related papers (2022-05-27T17:59:36Z) - Pure Noise to the Rescue of Insufficient Data: Improving Imbalanced
Classification by Training on Random Noise Images [12.91269560135337]
We present a surprisingly simple yet highly effective method to mitigate this limitation.
Unlike the common use of additive noise or adversarial noise for data augmentation, we propose directly training on pure random noise images.
We present a new Distribution-Aware Routing Batch Normalization layer (DAR-BN), which enables training on pure noise images in addition to natural images within the same network.
arXiv Detail & Related papers (2021-12-16T11:51:35Z) - Revisiting Consistency Regularization for Semi-Supervised Learning [80.28461584135967]
We propose an improved consistency regularization framework by a simple yet effective technique, FeatDistLoss.
Experimental results show that our model defines a new state of the art for various datasets and settings.
arXiv Detail & Related papers (2021-12-10T20:46:13Z) - A Generic Approach for Enhancing GANs by Regularized Latent Optimization [79.00740660219256]
We introduce a generic framework called em generative-model inference that is capable of enhancing pre-trained GANs effectively and seamlessly.
Our basic idea is to efficiently infer the optimal latent distribution for the given requirements using Wasserstein gradient flow techniques.
arXiv Detail & Related papers (2021-12-07T05:22:50Z) - InAugment: Improving Classifiers via Internal Augmentation [14.281619356571724]
We present a novel augmentation operation, that exploits image internal statistics.
We show improvement over state-of-the-art augmentation techniques.
We also demonstrate an increase for ResNet50 and EfficientNet-B3 top-1's accuracy on the ImageNet dataset.
arXiv Detail & Related papers (2021-04-08T15:37:21Z) - FeatMatch: Feature-Based Augmentation for Semi-Supervised Learning [64.32306537419498]
We propose a novel learned feature-based refinement and augmentation method that produces a varied set of complex transformations.
These transformations also use information from both within-class and across-class representations that we extract through clustering.
We demonstrate that our method is comparable to current state of art for smaller datasets while being able to scale up to larger datasets.
arXiv Detail & Related papers (2020-07-16T17:55:31Z) - Image Augmentations for GAN Training [57.65145659417266]
We provide insights and guidelines on how to augment images for both vanilla GANs and GANs with regularizations.
Surprisingly, we find that vanilla GANs attain generation quality on par with recent state-of-the-art results.
arXiv Detail & Related papers (2020-06-04T00:16:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.