Transfer Knowledge from Head to Tail: Uncertainty Calibration under
Long-tailed Distribution
- URL: http://arxiv.org/abs/2304.06537v1
- Date: Thu, 13 Apr 2023 13:48:18 GMT
- Title: Transfer Knowledge from Head to Tail: Uncertainty Calibration under
Long-tailed Distribution
- Authors: Jiahao Chen, Bing Su
- Abstract summary: Current calibration techniques treat different classes equally and implicitly assume that the distribution of training data is balanced.
We propose a novel knowledge-transferring-based calibration method by estimating the importance weights for samples of tail classes.
- Score: 24.734851889816206
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: How to estimate the uncertainty of a given model is a crucial problem.
Current calibration techniques treat different classes equally and thus
implicitly assume that the distribution of training data is balanced, but
ignore the fact that real-world data often follows a long-tailed distribution.
In this paper, we explore the problem of calibrating the model trained from a
long-tailed distribution. Due to the difference between the imbalanced training
distribution and balanced test distribution, existing calibration methods such
as temperature scaling can not generalize well to this problem. Specific
calibration methods for domain adaptation are also not applicable because they
rely on unlabeled target domain instances which are not available. Models
trained from a long-tailed distribution tend to be more overconfident to head
classes. To this end, we propose a novel knowledge-transferring-based
calibration method by estimating the importance weights for samples of tail
classes to realize long-tailed calibration. Our method models the distribution
of each class as a Gaussian distribution and views the source statistics of
head classes as a prior to calibrate the target distributions of tail classes.
We adaptively transfer knowledge from head classes to get the target
probability density of tail classes. The importance weight is estimated by the
ratio of the target probability density over the source probability density.
Extensive experiments on CIFAR-10-LT, MNIST-LT, CIFAR-100-LT, and ImageNet-LT
datasets demonstrate the effectiveness of our method.
Related papers
- Universality in Transfer Learning for Linear Models [18.427215139020625]
We study the problem of transfer learning in linear models for both regression and binary classification.
We provide an exact and rigorous analysis and relate generalization errors (in regression) and classification errors (in binary classification) for the pretrained and fine-tuned models.
arXiv Detail & Related papers (2024-10-03T03:09:09Z) - Uncertainty-aware Sampling for Long-tailed Semi-supervised Learning [89.98353600316285]
We introduce uncertainty into the modeling process for pseudo-label sampling, taking into account that the model performance on the tailed classes varies over different training stages.
This approach allows the model to perceive the uncertainty of pseudo-labels at different training stages, thereby adaptively adjusting the selection thresholds for different classes.
Compared to other methods such as the baseline method FixMatch, UDTS achieves an increase in accuracy of at least approximately 5.26%, 1.75%, 9.96%, and 1.28% on the natural scene image datasets.
arXiv Detail & Related papers (2024-01-09T08:59:39Z) - Class-Balancing Diffusion Models [57.38599989220613]
Class-Balancing Diffusion Models (CBDM) are trained with a distribution adjustment regularizer as a solution.
Our method benchmarked the generation results on CIFAR100/CIFAR100LT dataset and shows outstanding performance on the downstream recognition task.
arXiv Detail & Related papers (2023-04-30T20:00:14Z) - Adaptive Distribution Calibration for Few-Shot Learning with
Hierarchical Optimal Transport [78.9167477093745]
We propose a novel distribution calibration method by learning the adaptive weight matrix between novel samples and base classes.
Experimental results on standard benchmarks demonstrate that our proposed plug-and-play model outperforms competing approaches.
arXiv Detail & Related papers (2022-10-09T02:32:57Z) - Learnable Distribution Calibration for Few-Shot Class-Incremental
Learning [122.2241120474278]
Few-shot class-incremental learning (FSCIL) faces challenges of memorizing old class distributions and estimating new class distributions given few training samples.
We propose a learnable distribution calibration (LDC) approach, with the aim to systematically solve these two challenges using a unified framework.
arXiv Detail & Related papers (2022-10-01T09:40:26Z) - Learning Muti-expert Distribution Calibration for Long-tailed Video
Classification [88.12433458277168]
We propose an end-to-end multi-experts distribution calibration method based on two-level distribution information.
By modeling this two-level distribution information, the model can consider the head classes and the tail classes.
Our method achieves state-of-the-art performance on the long-tailed video classification task.
arXiv Detail & Related papers (2022-05-22T09:52:34Z) - Label-Aware Distribution Calibration for Long-tailed Classification [25.588323749920324]
We propose a label-Aware Distribution LADC approach to calibrate the distribution of tail classes.
Experiments on both image and text long-tailed datasets demonstrate that LADC significantly outperforms existing methods.
arXiv Detail & Related papers (2021-11-09T01:38:35Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.