Long-Tailed Partial Label Learning via Dynamic Rebalancing
- URL: http://arxiv.org/abs/2302.05080v1
- Date: Fri, 10 Feb 2023 06:43:53 GMT
- Title: Long-Tailed Partial Label Learning via Dynamic Rebalancing
- Authors: Feng Hong, Jiangchao Yao, Zhihan Zhou, Ya Zhang, Yanfeng Wang
- Abstract summary: Real-world data usually couples the label ambiguity and heavy imbalance.
LT methods build upon a given class distribution that is unavailable in, and the performance of is severely influenced in long-tailed context.
We propose a dynamic rebalancing method, termed as RECORDS, without assuming any prior knowledge about the class distribution.
- Score: 30.16563291182992
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real-world data usually couples the label ambiguity and heavy imbalance,
challenging the algorithmic robustness of partial label learning (PLL) and
long-tailed learning (LT). The straightforward combination of LT and PLL, i.e.,
LT-PLL, suffers from a fundamental dilemma: LT methods build upon a given class
distribution that is unavailable in PLL, and the performance of PLL is severely
influenced in long-tailed context. We show that even with the auxiliary of an
oracle class prior, the state-of-the-art methods underperform due to an adverse
fact that the constant rebalancing in LT is harsh to the label disambiguation
in PLL. To overcome this challenge, we thus propose a dynamic rebalancing
method, termed as RECORDS, without assuming any prior knowledge about the class
distribution. Based on a parametric decomposition of the biased output, our
method constructs a dynamic adjustment that is benign to the label
disambiguation process and theoretically converges to the oracle class prior.
Extensive experiments on three benchmark datasets demonstrate the significant
gain of RECORDS compared with a range of baselines. The code is publicly
available.
Related papers
- A Channel-ensemble Approach: Unbiased and Low-variance Pseudo-labels is Critical for Semi-supervised Classification [61.473485511491795]
Semi-supervised learning (SSL) is a practical challenge in computer vision.
Pseudo-label (PL) methods, e.g., FixMatch and FreeMatch, obtain the State Of The Art (SOTA) performances in SSL.
We propose a lightweight channel-based ensemble method to consolidate multiple inferior PLs into the theoretically guaranteed unbiased and low-variance one.
arXiv Detail & Related papers (2024-03-27T09:49:37Z) - Robust Representation Learning for Unreliable Partial Label Learning [86.909511808373]
Partial Label Learning (PLL) is a type of weakly supervised learning where each training instance is assigned a set of candidate labels, but only one label is the ground-truth.
This is known as Unreliable Partial Label Learning (UPLL) that introduces an additional complexity due to the inherent unreliability and ambiguity of partial labels.
We propose the Unreliability-Robust Representation Learning framework (URRL) that leverages unreliability-robust contrastive learning to help the model fortify against unreliable partial labels effectively.
arXiv Detail & Related papers (2023-08-31T13:37:28Z) - Learning in Imperfect Environment: Multi-Label Classification with
Long-Tailed Distribution and Partial Labels [53.68653940062605]
We introduce a novel task, Partial labeling and Long-Tailed Multi-Label Classification (PLT-MLC)
We find that most LT-MLC and PL-MLC approaches fail to solve the degradation-MLC.
We propose an end-to-end learning framework: textbfCOrrection $rightarrow$ textbfModificattextbfIon $rightarrow$ balantextbfCe.
arXiv Detail & Related papers (2023-04-20T20:05:08Z) - Pseudo Labels Regularization for Imbalanced Partial-Label Learning [11.93067260471484]
The challenge of partial-label learning and long-tail learning lies in matching between a decent marginal prior distribution with drawing the pseudo labels.
By punishing the pseudo labels of head classes, our method implements state-of-art under the standardized benchmarks.
arXiv Detail & Related papers (2023-03-06T15:21:55Z) - Controller-Guided Partial Label Consistency Regularization with
Unlabeled Data [49.24911720809604]
We propose a controller-guided consistency regularization at both the label-level and representation-level.
We dynamically adjust the confidence thresholds so that the number of samples of each class participating in consistency regularization remains roughly equal to alleviate the problem of class-imbalance.
arXiv Detail & Related papers (2022-10-20T12:15:13Z) - SoLar: Sinkhorn Label Refinery for Imbalanced Partial-Label Learning [31.535219018410707]
Partial-label learning (PLL) is a peculiar weakly-supervised learning task where the training samples are generally associated with a set of candidate labels instead of single ground truth.
We propose SoLar, a novel framework that allows refine the disambiguated labels towards matching the marginal class prior distribution.
SoLar exhibits substantially superior results on standardized benchmarks compared to the previous state-the-art methods.
arXiv Detail & Related papers (2022-09-21T14:00:16Z) - Progressive Purification for Instance-Dependent Partial Label Learning [37.65717805892473]
Partial label learning (PLL) aims to train multiclass classifiers from the examples each annotated with a set of candidate labels where a fixed but unknown candidate label is correct.
The candidate labels are always instance-dependent in practice and there is no theoretical guarantee that the model trained on the instance-dependent examples can converge to an ideal one.
In this paper, a theoretically grounded and practically effective approach named POP, i.e. PrOgressive Purification, is proposed. Specifically, POP updates the learning model and purifies each candidate label set progressively in every epoch.
arXiv Detail & Related papers (2022-06-02T02:07:12Z) - Relieving Long-tailed Instance Segmentation via Pairwise Class Balance [85.53585498649252]
Long-tailed instance segmentation is a challenging task due to the extreme imbalance of training samples among classes.
It causes severe biases of the head classes (with majority samples) against the tailed ones.
We propose a novel Pairwise Class Balance (PCB) method, built upon a confusion matrix which is updated during training to accumulate the ongoing prediction preferences.
arXiv Detail & Related papers (2022-01-08T07:48:36Z) - Two-stage Training for Learning from Label Proportions [18.78148397471913]
Learning from label proportions (LLP) aims at learning an instance-level classifier with label proportions in grouped training data.
We introduce the mixup strategy and symmetric crossentropy to further reduce the label noise.
Our framework is model-agnostic, and demonstrates compelling performance improvement in extensive experiments.
arXiv Detail & Related papers (2021-05-22T03:55:35Z) - Provably Consistent Partial-Label Learning [120.4734093544867]
Partial-label learning (PLL) is a multi-class classification problem, where each training example is associated with a set of candidate labels.
In this paper, we propose the first generation model of candidate label sets, and develop two novel methods that are guaranteed to be consistent.
Experiments on benchmark and real-world datasets validate the effectiveness of the proposed generation model and two methods.
arXiv Detail & Related papers (2020-07-17T12:19:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.