AutoMix: Unveiling the Power of Mixup
- URL: http://arxiv.org/abs/2103.13027v1
- Date: Wed, 24 Mar 2021 07:21:53 GMT
- Title: AutoMix: Unveiling the Power of Mixup
- Authors: Zicheng Liu, Siyuan Li, Di Wu, Zhiyuan Chen, Lirong Wu, Jianzhu Guo,
Stan Z. Li
- Abstract summary: We present a flexible, general Automatic Mixup framework which utilizes discriminative features to learn a sample mixing policy adaptively.
We regard mixup as a pretext task and split it into two sub-problems: mixed samples generation and mixup classification.
Experiments on six popular classification benchmarks show that AutoMix consistently outperforms other leading mixup methods.
- Score: 34.623943038648164
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Mixup-based data augmentation has achieved great success as regularizer for
deep neural networks. However, existing mixup methods require explicitly
designed mixup policies. In this paper, we present a flexible, general
Automatic Mixup (AutoMix) framework which utilizes discriminative features to
learn a sample mixing policy adaptively. We regard mixup as a pretext task and
split it into two sub-problems: mixed samples generation and mixup
classification. To this end, we design a lightweight mix block to generate
synthetic samples based on feature maps and mix labels. Since the two
sub-problems are in the nature of Expectation-Maximization (EM), we also
propose a momentum training pipeline to optimize the mixup process and mixup
classification process alternatively in an end-to-end fashion. Extensive
experiments on six popular classification benchmarks show that AutoMix
consistently outperforms other leading mixup methods and improves
generalization abilities to downstream tasks. We hope AutoMix will motivate the
community to rethink the role of mixup in representation learning. The code
will be released soon.
Related papers
- SUMix: Mixup with Semantic and Uncertain Information [41.99721365685618]
Mixup data augmentation approaches have been applied for various tasks of deep learning.
We propose a novel approach named SUMix to learn the mixing ratio as well as the uncertainty for the mixed samples during the training process.
arXiv Detail & Related papers (2024-07-10T16:25:26Z) - SynerMix: Synergistic Mixup Solution for Enhanced Intra-Class Cohesion and Inter-Class Separability in Image Classification [5.86619806485267]
We propose a novel mixup method named SynerMix-Intra and, building upon this, introduce a synergistic mixup solution named SynerMix.
SynerMix-Intra targets intra-class mixup to bolster intra-class cohesion, a feature not addressed by current mixup methods.
Experiments show that SynerMix achieves a 0.1% to 3.43% higher accuracy than the best of either MixUp or SynerMix-Intra alone, averaging a 1.16% gain.
arXiv Detail & Related papers (2024-03-21T05:13:12Z) - PowMix: A Versatile Regularizer for Multimodal Sentiment Analysis [71.8946280170493]
This paper introduces PowMix, a versatile embedding space regularizer that builds upon the strengths of unimodal mixing-based regularization approaches.
PowMix is integrated before the fusion stage of multimodal architectures and facilitates intra-modal mixing, such as mixing text with text, to act as a regularizer.
arXiv Detail & Related papers (2023-12-19T17:01:58Z) - Adversarial AutoMixup [50.1874436169571]
We propose AdAutomixup, an adversarial automatic mixup augmentation approach.
It generates challenging samples to train a robust classifier for image classification.
Our approach outperforms the state of the art in various classification scenarios.
arXiv Detail & Related papers (2023-12-19T08:55:00Z) - Infinite Class Mixup [26.48101652432502]
Mixup is a strategy for training deep networks where additional samples are augmented by interpolating inputs and labels of training pairs.
This paper seeks to address this cornerstone by mixing the classifiers directly instead of mixing the labels for each mixed pair.
We show that Infinite Class Mixup outperforms standard Mixup and variants such as RegMixup and Remix on balanced, long-tailed, and data-constrained benchmarks.
arXiv Detail & Related papers (2023-05-17T15:27:35Z) - Expeditious Saliency-guided Mix-up through Random Gradient Thresholding [89.59134648542042]
Mix-up training approaches have proven to be effective in improving the generalization ability of Deep Neural Networks.
In this paper, inspired by the superior qualities of each direction over one another, we introduce a novel method that lies at the junction of the two routes.
We name our method R-Mix following the concept of "Random Mix-up"
In order to address the question of whether there exists a better decision protocol, we train a Reinforcement Learning agent that decides the mix-up policies.
arXiv Detail & Related papers (2022-12-09T14:29:57Z) - C-Mixup: Improving Generalization in Regression [71.10418219781575]
Mixup algorithm improves generalization by linearly interpolating a pair of examples and their corresponding labels.
We propose C-Mixup, which adjusts the sampling probability based on the similarity of the labels.
C-Mixup achieves 6.56%, 4.76%, 5.82% improvements in in-distribution generalization, task generalization, and out-of-distribution robustness, respectively.
arXiv Detail & Related papers (2022-10-11T20:39:38Z) - Harnessing Hard Mixed Samples with Decoupled Regularizer [69.98746081734441]
Mixup is an efficient data augmentation approach that improves the generalization of neural networks by smoothing the decision boundary with mixed data.
In this paper, we propose an efficient mixup objective function with a decoupled regularizer named Decoupled Mixup (DM)
DM can adaptively utilize hard mixed samples to mine discriminative features without losing the original smoothness of mixup.
arXiv Detail & Related papers (2022-03-21T07:12:18Z) - MixMo: Mixing Multiple Inputs for Multiple Outputs via Deep Subnetworks [97.08677678499075]
We introduce MixMo, a new framework for learning multi-input multi-output deepworks.
We show that binary mixing in features - particularly with patches from CutMix - enhances results by makingworks stronger and more diverse.
In addition to being easy to implement and adding no cost at inference, our models outperform much costlier data augmented deep ensembles.
arXiv Detail & Related papers (2021-03-10T15:31:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.