Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective
- URL: http://arxiv.org/abs/2212.02801v1
- Date: Tue, 6 Dec 2022 07:38:29 GMT
- Title: Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective
- Authors: Yunrui Zhao, Qianqian Xu, Yangbangyan Jiang, Peisong Wen, and Qingming
Huang
- Abstract summary: We propose a label distribution perspective for PU learning in this paper.
Motivated by this, we propose to pursue the label distribution consistency between predicted and ground-truth label distributions.
Experiments on three benchmark datasets validate the effectiveness of the proposed method.
- Score: 89.5370481649529
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Positive-Unlabeled (PU) learning tries to learn binary classifiers from a few
labeled positive examples with many unlabeled ones. Compared with ordinary
semi-supervised learning, this task is much more challenging due to the absence
of any known negative labels. While existing cost-sensitive-based methods have
achieved state-of-the-art performances, they explicitly minimize the risk of
classifying unlabeled data as negative samples, which might result in a
negative-prediction preference of the classifier. To alleviate this issue, we
resort to a label distribution perspective for PU learning in this paper.
Noticing that the label distribution of unlabeled data is fixed when the class
prior is known, it can be naturally used as learning supervision for the model.
Motivated by this, we propose to pursue the label distribution consistency
between predicted and ground-truth label distributions, which is formulated by
aligning their expectations. Moreover, we further adopt the entropy
minimization and Mixup regularization to avoid the trivial solution of the
label distribution consistency on unlabeled data and mitigate the consequent
confirmation bias. Experiments on three benchmark datasets validate the
effectiveness of the proposed method.Code available at:
https://github.com/Ray-rui/Dist-PU-Positive-Unlabeled-Learning-from-a-Label-Distribution-Perspective .
Related papers
- Reduction-based Pseudo-label Generation for Instance-dependent Partial Label Learning [41.345794038968776]
We propose to leverage reduction-based pseudo-labels to alleviate the influence of incorrect candidate labels.
We show that reduction-based pseudo-labels exhibit greater consistency with the Bayes optimal classifier compared to pseudo-labels directly generated from the predictive model.
arXiv Detail & Related papers (2024-10-28T07:32:20Z) - Class-Distribution-Aware Pseudo Labeling for Semi-Supervised Multi-Label
Learning [97.88458953075205]
Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data.
This paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudo-labeling in a class-aware manner.
arXiv Detail & Related papers (2023-05-04T12:52:18Z) - Label distribution learning via label correlation grid [9.340734188957727]
We propose a textbfLabel textbfCorrelation textbfGrid (LCG) to model the uncertainty of label relationships.
Our network learns the LCG to accurately estimate the label distribution for each instance.
arXiv Detail & Related papers (2022-10-15T03:58:15Z) - Learning with Proper Partial Labels [87.65718705642819]
Partial-label learning is a kind of weakly-supervised learning with inexact labels.
We show that this proper partial-label learning framework includes many previous partial-label learning settings.
We then derive a unified unbiased estimator of the classification risk.
arXiv Detail & Related papers (2021-12-23T01:37:03Z) - Instance-Dependent Partial Label Learning [69.49681837908511]
Partial label learning is a typical weakly supervised learning problem.
Most existing approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels.
In this paper, we consider instance-dependent and assume that each example is associated with a latent label distribution constituted by the real number of each label.
arXiv Detail & Related papers (2021-10-25T12:50:26Z) - Distribution-Aware Semantics-Oriented Pseudo-label for Imbalanced
Semi-Supervised Learning [80.05441565830726]
This paper addresses imbalanced semi-supervised learning, where heavily biased pseudo-labels can harm the model performance.
We propose a general pseudo-labeling framework to address the bias motivated by this observation.
We term the novel pseudo-labeling framework for imbalanced SSL as Distribution-Aware Semantics-Oriented (DASO) Pseudo-label.
arXiv Detail & Related papers (2021-06-10T11:58:25Z) - Disentangling Sampling and Labeling Bias for Learning in Large-Output
Spaces [64.23172847182109]
We show that different negative sampling schemes implicitly trade-off performance on dominant versus rare labels.
We provide a unified means to explicitly tackle both sampling bias, arising from working with a subset of all labels, and labeling bias, which is inherent to the data due to label imbalance.
arXiv Detail & Related papers (2021-05-12T15:40:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.