Improving Model Generalization by Agreement of Learned Representations
from Data Augmentation
- URL: http://arxiv.org/abs/2110.10536v1
- Date: Wed, 20 Oct 2021 12:44:52 GMT
- Title: Improving Model Generalization by Agreement of Learned Representations
from Data Augmentation
- Authors: Rowel Atienza
- Abstract summary: In computer vision, data augmentation techniques such as CutOut, MixUp, and CutMix demonstrated state-of-the-art (SOTA) results.
We call our proposed method Agreement Maximization or simply AgMax.
We show that AgMax can take advantage of the data augmentation to consistently improve model generalization by a significant margin.
- Score: 19.286766429954174
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Data augmentation reduces the generalization error by forcing a model to
learn invariant representations given different transformations of the input
image. In computer vision, on top of the standard image processing functions,
data augmentation techniques based on regional dropout such as CutOut, MixUp,
and CutMix and policy-based selection such as AutoAugment demonstrated
state-of-the-art (SOTA) results. With an increasing number of data augmentation
algorithms being proposed, the focus is always on optimizing the input-output
mapping while not realizing that there might be an untapped value in the
transformed images with the same label. We hypothesize that by forcing the
representations of two transformations to agree, we can further reduce the
model generalization error. We call our proposed method Agreement Maximization
or simply AgMax. With this simple constraint applied during training, empirical
results show that data augmentation algorithms can further improve the
classification accuracy of ResNet50 on ImageNet by up to 1.5%, WideResNet40-2
on CIFAR10 by up to 0.7%, WideResNet40-2 on CIFAR100 by up to 1.6%, and LeNet5
on Speech Commands Dataset by up to 1.4%. Experimental results further show
that unlike other regularization terms such as label smoothing, AgMax can take
advantage of the data augmentation to consistently improve model generalization
by a significant margin. On downstream tasks such as object detection and
segmentation on PascalVOC and COCO, AgMax pre-trained models outperforms other
data augmentation methods by as much as 1.0mAP (box) and 0.5mAP (mask). Code is
available at https://github.com/roatienza/agmax.
Related papers
- Reinforce Data, Multiply Impact: Improved Model Accuracy and Robustness
with Dataset Reinforcement [68.44100784364987]
We propose a strategy to improve a dataset once such that the accuracy of any model architecture trained on the reinforced dataset is improved at no additional training cost for users.
We create a reinforced version of the ImageNet training dataset, called ImageNet+, as well as reinforced datasets CIFAR-100+, Flowers-102+, and Food-101+.
Models trained with ImageNet+ are more accurate, robust, and calibrated, and transfer well to downstream tasks.
arXiv Detail & Related papers (2023-03-15T23:10:17Z) - Improving Zero-shot Generalization and Robustness of Multi-modal Models [70.14692320804178]
Multi-modal image-text models such as CLIP and LiT have demonstrated impressive performance on image classification benchmarks.
We investigate the reasons for this performance gap and find that many of the failure cases are caused by ambiguity in the text prompts.
We propose a simple and efficient way to improve accuracy on such uncertain images by making use of the WordNet hierarchy.
arXiv Detail & Related papers (2022-12-04T07:26:24Z) - Soft Augmentation for Image Classification [68.71067594724663]
We propose generalizing augmentation with invariant transforms to soft augmentation.
We show that soft targets allow for more aggressive data augmentation.
We also show that soft augmentations generalize to self-supervised classification tasks.
arXiv Detail & Related papers (2022-11-09T01:04:06Z) - ParticleAugment: Sampling-Based Data Augmentation [80.44268663372233]
We propose a particle filtering formulation to find optimal augmentation policies and their schedules during model training.
We show that our formulation for automated augmentation reaches promising results on CIFAR-10, CIFAR-100, and ImageNet datasets.
arXiv Detail & Related papers (2021-06-16T10:56:02Z) - InAugment: Improving Classifiers via Internal Augmentation [14.281619356571724]
We present a novel augmentation operation, that exploits image internal statistics.
We show improvement over state-of-the-art augmentation techniques.
We also demonstrate an increase for ResNet50 and EfficientNet-B3 top-1's accuracy on the ImageNet dataset.
arXiv Detail & Related papers (2021-04-08T15:37:21Z) - Learning Representational Invariances for Data-Efficient Action
Recognition [52.23716087656834]
We show that our data augmentation strategy leads to promising performance on the Kinetics-100, UCF-101, and HMDB-51 datasets.
We also validate our data augmentation strategy in the fully supervised setting and demonstrate improved performance.
arXiv Detail & Related papers (2021-03-30T17:59:49Z) - Augmentation Inside the Network [1.5260179407438161]
We present augmentation inside the network, a method that simulates data augmentation techniques for computer vision problems.
We validate our method on the ImageNet-2012 and CIFAR-100 datasets for image classification.
arXiv Detail & Related papers (2020-12-19T20:07:03Z) - FeatMatch: Feature-Based Augmentation for Semi-Supervised Learning [64.32306537419498]
We propose a novel learned feature-based refinement and augmentation method that produces a varied set of complex transformations.
These transformations also use information from both within-class and across-class representations that we extract through clustering.
We demonstrate that our method is comparable to current state of art for smaller datasets while being able to scale up to larger datasets.
arXiv Detail & Related papers (2020-07-16T17:55:31Z) - On the Generalization Effects of Linear Transformations in Data
Augmentation [32.01435459892255]
Data augmentation is a powerful technique to improve performance in applications such as image and text classification tasks.
We study a family of linear transformations and study their effects on the ridge estimator in an over-parametrized linear regression setting.
We propose an augmentation scheme that searches over the space of transformations by how uncertain the model is about the transformed data.
arXiv Detail & Related papers (2020-05-02T04:10:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.