Multiclass Classification using dilute bandit feedback
- URL: http://arxiv.org/abs/2105.08093v1
- Date: Mon, 17 May 2021 18:05:34 GMT
- Title: Multiclass Classification using dilute bandit feedback
- Authors: Gaurav Batra, Naresh Manwani
- Abstract summary: We propose an algorithm for multiclass classification using dilute bandit feedback (MC-DBF)
We show that the proposed algorithm achieves O(T1-frac1m+2) mistake bound if candidate label set size (in each step) is m.
- Score: 8.452237741722726
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces a new online learning framework for multiclass
classification called learning with diluted bandit feedback. At every time
step, the algorithm predicts a candidate label set instead of a single label
for the observed example. It then receives feedback from the environment
whether the actual label lies in this candidate label set or not. This feedback
is called "diluted bandit feedback". Learning in this setting is even more
challenging than the bandit feedback setting, as there is more uncertainty in
the supervision. We propose an algorithm for multiclass classification using
dilute bandit feedback (MC-DBF), which uses the exploration-exploitation
strategy to predict the candidate set in each trial. We show that the proposed
algorithm achieves O(T^{1-\frac{1}{m+2}}) mistake bound if candidate label set
size (in each step) is m. We demonstrate the effectiveness of the proposed
approach with extensive simulations.
Related papers
- Partial-Label Regression [54.74984751371617]
Partial-label learning is a weakly supervised learning setting that allows each training example to be annotated with a set of candidate labels.
Previous studies on partial-label learning only focused on the classification setting where candidate labels are all discrete.
In this paper, we provide the first attempt to investigate partial-label regression, where each training example is annotated with a set of real-valued candidate labels.
arXiv Detail & Related papers (2023-06-15T09:02:24Z) - Class-Distribution-Aware Pseudo Labeling for Semi-Supervised Multi-Label
Learning [97.88458953075205]
Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data.
This paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudo-labeling in a class-aware manner.
arXiv Detail & Related papers (2023-05-04T12:52:18Z) - M-Tuning: Prompt Tuning with Mitigated Label Bias in Open-Set Scenarios [103.6153593636399]
We propose a vision-language prompt tuning method with mitigated label bias (M-Tuning)
It introduces open words from the WordNet to extend the range of words forming the prompt texts from only closed-set label words to more, and thus prompts are tuned in a simulated open-set scenario.
Our method achieves the best performance on datasets with various scales, and extensive ablation studies also validate its effectiveness.
arXiv Detail & Related papers (2023-03-09T09:05:47Z) - Learning with Proper Partial Labels [87.65718705642819]
Partial-label learning is a kind of weakly-supervised learning with inexact labels.
We show that this proper partial-label learning framework includes many previous partial-label learning settings.
We then derive a unified unbiased estimator of the classification risk.
arXiv Detail & Related papers (2021-12-23T01:37:03Z) - Multi-Objective Interpolation Training for Robustness to Label Noise [17.264550056296915]
We show that standard supervised contrastive learning degrades in the presence of label noise.
We propose a novel label noise detection method that exploits the robust feature representations learned via contrastive learning.
Experiments on synthetic and real-world noise benchmarks demonstrate that MOIT/MOIT+ achieves state-of-the-art results.
arXiv Detail & Related papers (2020-12-08T15:01:54Z) - One-bit Supervision for Image Classification [121.87598671087494]
One-bit supervision is a novel setting of learning from incomplete annotations.
We propose a multi-stage training paradigm which incorporates negative label suppression into an off-the-shelf semi-supervised learning algorithm.
arXiv Detail & Related papers (2020-09-14T03:06:23Z) - Reliable Label Bootstrapping for Semi-Supervised Learning [19.841733658911767]
ReLaB is an unsupervised preprossessing algorithm which improves the performance of semi-supervised algorithms in extremely low supervision settings.
We show that the selection of the network architecture and the self-supervised algorithm are important factors to achieve successful label propagation.
We reach average error rates of $boldsymbol22.34$ with 1 random labeled sample per class on CIFAR-10 and lower this error to $boldsymbol8.46$ when the labeled sample in each class is highly representative.
arXiv Detail & Related papers (2020-07-23T08:51:37Z) - Learning Multiclass Classifier Under Noisy Bandit Feedback [6.624726878647541]
We propose a novel approach to deal with noisy bandit feedback based on the unbiased estimator technique.
We show our approach's effectiveness using extensive experiments on several benchmark datasets.
arXiv Detail & Related papers (2020-06-05T16:31:05Z) - Unsupervised Person Re-identification via Multi-label Classification [55.65870468861157]
This paper formulates unsupervised person ReID as a multi-label classification task to progressively seek true labels.
Our method starts by assigning each person image with a single-class label, then evolves to multi-label classification by leveraging the updated ReID model for label prediction.
To boost the ReID model training efficiency in multi-label classification, we propose the memory-based multi-label classification loss (MMCL)
arXiv Detail & Related papers (2020-04-20T12:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.