Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification
- URL: http://arxiv.org/abs/2208.02951v1
- Date: Fri, 5 Aug 2022 01:23:54 GMT
- Title: Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification
- Authors: Dandan Guo, Zhuo Li, Meixi Zheng, He Zhao, Mingyuan Zhou, Hongyuan Zha
- Abstract summary: Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
- Score: 74.62203971625173
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Imbalanced data pose challenges for deep learning based classification
models. One of the most widely-used approaches for tackling imbalanced data is
re-weighting, where training samples are associated with different weights in
the loss function. Most of existing re-weighting approaches treat the example
weights as the learnable parameter and optimize the weights on the meta set,
entailing expensive bilevel optimization. In this paper, we propose a novel
re-weighting method based on optimal transport (OT) from a distributional point
of view. Specifically, we view the training set as an imbalanced distribution
over its samples, which is transported by OT to a balanced distribution
obtained from the meta set. The weights of the training samples are the
probability mass of the imbalanced distribution and learned by minimizing the
OT distance between the two distributions. Compared with existing methods, our
proposed one disengages the dependence of the weight learning on the concerned
classifier at each iteration. Experiments on image, text and point cloud
datasets demonstrate that our proposed re-weighting method has excellent
performance, achieving state-of-the-art results in many cases and providing a
promising tool for addressing the imbalanced classification issue.
Related papers
- Optimizing importance weighting in the presence of sub-population shifts [0.0]
A distribution shift between the training and test data can severely harm performance of machine learning models.
We argue that existing weightings for determining the weights are suboptimal, as they neglect the increase of the variance of the estimated model due to the finite sample size of the training data.
We propose a bi-level optimization procedure in which the weights and model parameters are optimized simultaneously.
arXiv Detail & Related papers (2024-10-18T09:21:10Z) - Rethinking Classifier Re-Training in Long-Tailed Recognition: A Simple
Logits Retargeting Approach [102.0769560460338]
We develop a simple logits approach (LORT) without the requirement of prior knowledge of the number of samples per class.
Our method achieves state-of-the-art performance on various imbalanced datasets, including CIFAR100-LT, ImageNet-LT, and iNaturalist 2018.
arXiv Detail & Related papers (2024-03-01T03:27:08Z) - Adaptive Distribution Calibration for Few-Shot Learning with
Hierarchical Optimal Transport [78.9167477093745]
We propose a novel distribution calibration method by learning the adaptive weight matrix between novel samples and base classes.
Experimental results on standard benchmarks demonstrate that our proposed plug-and-play model outperforms competing approaches.
arXiv Detail & Related papers (2022-10-09T02:32:57Z) - Residual-Quantile Adjustment for Adaptive Training of Physics-informed
Neural Network [2.5769426017309915]
In this paper, we show that the bottleneck in the adaptive choice of samples for training efficiency is the behavior of the tail distribution of the numerical residual.
We propose the Residual-Quantile Adjustment (RQA) method for a better weight choice for each training sample.
Experiment results show that the proposed method can outperform several adaptive methods on various partial differential equation (PDE) problems.
arXiv Detail & Related papers (2022-09-09T12:39:38Z) - CMW-Net: Learning a Class-Aware Sample Weighting Mapping for Robust Deep
Learning [55.733193075728096]
Modern deep neural networks can easily overfit to biased training data containing corrupted labels or class imbalance.
Sample re-weighting methods are popularly used to alleviate this data bias issue.
We propose a meta-model capable of adaptively learning an explicit weighting scheme directly from data.
arXiv Detail & Related papers (2022-02-11T13:49:51Z) - Continuous Weight Balancing [0.0]
We propose a simple method by which to choose sample weights for problems with highly imbalanced or skewed traits.
We derive sample weights from the transfer function between an estimated source and specified target distributions.
Our method outperforms both unweighted and discretely-weighted models on both regression and classification tasks.
arXiv Detail & Related papers (2021-03-30T18:03:12Z) - Jo-SRC: A Contrastive Approach for Combating Noisy Labels [58.867237220886885]
We propose a noise-robust approach named Jo-SRC (Joint Sample Selection and Model Regularization based on Consistency)
Specifically, we train the network in a contrastive learning manner. Predictions from two different views of each sample are used to estimate its "likelihood" of being clean or out-of-distribution.
arXiv Detail & Related papers (2021-03-24T07:26:07Z) - Attentional-Biased Stochastic Gradient Descent [74.49926199036481]
We present a provable method (named ABSGD) for addressing the data imbalance or label noise problem in deep learning.
Our method is a simple modification to momentum SGD where we assign an individual importance weight to each sample in the mini-batch.
ABSGD is flexible enough to combine with other robust losses without any additional cost.
arXiv Detail & Related papers (2020-12-13T03:41:52Z) - Imbalanced Data Learning by Minority Class Augmentation using Capsule
Adversarial Networks [31.073558420480964]
We propose a method to restore the balance in imbalanced images, by coalescing two concurrent methods.
In our model, generative and discriminative networks play a novel competitive game.
The coalescing of capsule-GAN is effective at recognizing highly overlapping classes with much fewer parameters compared with the convolutional-GAN.
arXiv Detail & Related papers (2020-04-05T12:36:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.