LUMix: Improving Mixup by Better Modelling Label Uncertainty
- URL: http://arxiv.org/abs/2211.15846v1
- Date: Tue, 29 Nov 2022 00:47:55 GMT
- Title: LUMix: Improving Mixup by Better Modelling Label Uncertainty
- Authors: Shuyang Sun, Jie-Neng Chen, Ruifei He, Alan Yuille, Philip Torr, Song
Bai
- Abstract summary: Deep networks can be better generalized when trained with noisy samples and regularization techniques.
Previous Mixup-based methods linearly combine images and labels to generate additional training data.
We propose LUMix, which models such uncertainty by adding label perturbation during training.
- Score: 33.56660038646426
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Modern deep networks can be better generalized when trained with noisy
samples and regularization techniques. Mixup and CutMix have been proven to be
effective for data augmentation to help avoid overfitting. Previous Mixup-based
methods linearly combine images and labels to generate additional training
data. However, this is problematic if the object does not occupy the whole
image as we demonstrate in Figure 1. Correctly assigning the label weights is
hard even for human beings and there is no clear criterion to measure it. To
tackle this problem, in this paper, we propose LUMix, which models such
uncertainty by adding label perturbation during training. LUMix is simple as it
can be implemented in just a few lines of code and can be universally applied
to any deep networks \eg CNNs and Vision Transformers, with minimal
computational cost. Extensive experiments show that our LUMix can consistently
boost the performance for networks with a wide range of diversity and capacity
on ImageNet, \eg $+0.7\%$ for a small model DeiT-S and $+0.6\%$ for a large
variant XCiT-L. We also demonstrate that LUMix can lead to better robustness
when evaluated on ImageNet-O and ImageNet-A. The source code can be found
\href{https://github.com/kevin-ssy/LUMix}{here}
Related papers
- GeMix: Conditional GAN-Based Mixup for Improved Medical Image Augmentation [20.792833454137625]
We propose GeMix, a two-stage framework that replaces blending with a learned, label-aware powered by class-conditional GANs.<n>We benchmark GeMix on the large-scale COVIDx-CT-3 dataset using three backbones (ResNet-50, ResNet-101, EfficientNet-B0). When combined with real data, our method increases macro-F1 over traditional mixup for all backbones, reducing the false negative rate for COVID-19 detection.
arXiv Detail & Related papers (2025-07-21T12:58:05Z) - SUMix: Mixup with Semantic and Uncertain Information [41.99721365685618]
Mixup data augmentation approaches have been applied for various tasks of deep learning.
We propose a novel approach named SUMix to learn the mixing ratio as well as the uncertainty for the mixed samples during the training process.
arXiv Detail & Related papers (2024-07-10T16:25:26Z) - SMMix: Self-Motivated Image Mixing for Vision Transformers [65.809376136455]
CutMix is a vital augmentation strategy that determines the performance and generalization ability of vision transformers (ViTs)
Existing CutMix variants tackle this problem by generating more consistent mixed images or more precise mixed labels.
We propose an efficient and effective Self-Motivated image Mixing method (SMMix) which motivates both image and label enhancement by the model under training itself.
arXiv Detail & Related papers (2022-12-26T00:19:39Z) - RecursiveMix: Mixed Learning with History [21.865332756486314]
"RecursiveMix" (RM) is a mixed-sample learning paradigm that leverages the historical input-prediction-label triplets.
Based on ResNet-50, RM largely improves classification accuracy by $sim$3.2% on CIFAR100 and $sim$2.8% on ImageNet with negligible extra computation/storage costs.
arXiv Detail & Related papers (2022-03-14T03:59:47Z) - Enhanced Performance of Pre-Trained Networks by Matched Augmentation
Distributions [10.74023489125222]
We propose a simple solution to address the train-test distributional shift.
We combine results for multiple random crops for a test image.
This not only matches the train time augmentation but also provides the full coverage of the input image.
arXiv Detail & Related papers (2022-01-19T22:33:00Z) - ReMix: Towards Image-to-Image Translation with Limited Data [154.71724970593036]
We propose a data augmentation method (ReMix) to tackle this issue.
We interpolate training samples at the feature level and propose a novel content loss based on the perceptual relations among samples.
The proposed approach effectively reduces the ambiguity of generation and renders content-preserving results.
arXiv Detail & Related papers (2021-03-31T06:24:10Z) - MixMo: Mixing Multiple Inputs for Multiple Outputs via Deep Subnetworks [97.08677678499075]
We introduce MixMo, a new framework for learning multi-input multi-output deepworks.
We show that binary mixing in features - particularly with patches from CutMix - enhances results by makingworks stronger and more diverse.
In addition to being easy to implement and adding no cost at inference, our models outperform much costlier data augmented deep ensembles.
arXiv Detail & Related papers (2021-03-10T15:31:02Z) - SnapMix: Semantically Proportional Mixing for Augmenting Fine-grained
Data [124.95585891086894]
Proposal is called Semantically Proportional Mixing (SnapMix)
It exploits class activation map (CAM) to lessen the label noise in augmenting fine-grained data.
Our method consistently outperforms existing mixed-based approaches.
arXiv Detail & Related papers (2020-12-09T03:37:30Z) - Puzzle Mix: Exploiting Saliency and Local Statistics for Optimal Mixup [19.680580983094323]
Puzzle Mix is a mixup method for explicitly utilizing the saliency information and the underlying statistics of the natural examples.
Our experiments show Puzzle Mix achieves the state of the art generalization and the adversarial robustness results.
arXiv Detail & Related papers (2020-09-15T10:10:23Z) - Un-Mix: Rethinking Image Mixtures for Unsupervised Visual Representation
Learning [108.999497144296]
Recently advanced unsupervised learning approaches use the siamese-like framework to compare two "views" from the same image for learning representations.
This work aims to involve the distance concept on label space in the unsupervised learning and let the model be aware of the soft degree of similarity between positive or negative pairs.
Despite its conceptual simplicity, we show empirically that with the solution -- Unsupervised image mixtures (Un-Mix), we can learn subtler, more robust and generalized representations from the transformed input and corresponding new label space.
arXiv Detail & Related papers (2020-03-11T17:59:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.