Semantic Equivariant Mixup
- URL: http://arxiv.org/abs/2308.06451v1
- Date: Sat, 12 Aug 2023 03:05:53 GMT
- Title: Semantic Equivariant Mixup
- Authors: Zongbo Han, Tianchi Xie, Bingzhe Wu, Qinghua Hu, Changqing Zhang
- Abstract summary: Mixup is a well-established data augmentation technique, which can extend the training distribution and regularize the neural networks.
Previous mixup variants tend to over-focus on the label-related information.
We propose a semantic equivariant mixup (sem) to preserve richer semantic information in the input.
- Score: 54.734054770032934
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Mixup is a well-established data augmentation technique, which can extend the
training distribution and regularize the neural networks by creating ''mixed''
samples based on the label-equivariance assumption, i.e., a proportional mixup
of the input data results in the corresponding labels being mixed in the same
proportion. However, previous mixup variants may fail to exploit the
label-independent information in mixed samples during training, which usually
contains richer semantic information. To further release the power of mixup, we
first improve the previous label-equivariance assumption by the
semantic-equivariance assumption, which states that the proportional mixup of
the input data should lead to the corresponding representation being mixed in
the same proportion. Then a generic mixup regularization at the representation
level is proposed, which can further regularize the model with the semantic
information in mixed samples. At a high level, the proposed semantic
equivariant mixup (sem) encourages the structure of the input data to be
preserved in the representation space, i.e., the change of input will result in
the obtained representation information changing in the same way. Different
from previous mixup variants, which tend to over-focus on the label-related
information, the proposed method aims to preserve richer semantic information
in the input with semantic-equivariance assumption, thereby improving the
robustness of the model against distribution shifts. We conduct extensive
empirical studies and qualitative analyzes to demonstrate the effectiveness of
our proposed method. The code of the manuscript is in the supplement.
Related papers
- SUMix: Mixup with Semantic and Uncertain Information [41.99721365685618]
Mixup data augmentation approaches have been applied for various tasks of deep learning.
We propose a novel approach named SUMix to learn the mixing ratio as well as the uncertainty for the mixed samples during the training process.
arXiv Detail & Related papers (2024-07-10T16:25:26Z) - Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - Selective Mixup Helps with Distribution Shifts, But Not (Only) because
of Mixup [26.105340203096596]
We show that non-random selection of pairs affects the training distribution and improve generalization by means completely unrelated to the mixing.
We have found a new equivalence between two successful methods: selective mixup and resampling.
arXiv Detail & Related papers (2023-05-26T10:56:22Z) - Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot
Text Classification Tasks [75.42002070547267]
We propose a self evolution learning (SE) based mixup approach for data augmentation in text classification.
We introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up.
arXiv Detail & Related papers (2023-05-22T23:43:23Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - C-Mixup: Improving Generalization in Regression [71.10418219781575]
Mixup algorithm improves generalization by linearly interpolating a pair of examples and their corresponding labels.
We propose C-Mixup, which adjusts the sampling probability based on the similarity of the labels.
C-Mixup achieves 6.56%, 4.76%, 5.82% improvements in in-distribution generalization, task generalization, and out-of-distribution robustness, respectively.
arXiv Detail & Related papers (2022-10-11T20:39:38Z) - Harnessing Hard Mixed Samples with Decoupled Regularizer [69.98746081734441]
Mixup is an efficient data augmentation approach that improves the generalization of neural networks by smoothing the decision boundary with mixed data.
In this paper, we propose an efficient mixup objective function with a decoupled regularizer named Decoupled Mixup (DM)
DM can adaptively utilize hard mixed samples to mine discriminative features without losing the original smoothness of mixup.
arXiv Detail & Related papers (2022-03-21T07:12:18Z) - A similarity-based Bayesian mixture-of-experts model [0.5156484100374058]
We present a new non-parametric mixture-of-experts model for multivariate regression problems.
Using a conditionally specified model, predictions for out-of-sample inputs are based on similarities to each observed data point.
Posterior inference is performed on the parameters of the mixture as well as the distance metric.
arXiv Detail & Related papers (2020-12-03T18:08:30Z) - On Mixup Regularization [16.748910388577308]
Mixup is a data augmentation technique that creates new examples as convex combinations of training points and labels.
We show how the random perturbation of the new interpretation of Mixup induces multiple known regularization schemes.
arXiv Detail & Related papers (2020-06-10T20:11:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.