Trustable Co-label Learning from Multiple Noisy Annotators
- URL: http://arxiv.org/abs/2203.04199v1
- Date: Tue, 8 Mar 2022 16:57:00 GMT
- Title: Trustable Co-label Learning from Multiple Noisy Annotators
- Authors: Shikun Li, Tongliang Liu, Jiyong Tan, Dan Zeng, Shiming Ge
- Abstract summary: Supervised deep learning depends on massive accurately annotated examples.
A typical alternative is learning from multiple noisy annotators.
This paper proposes a data-efficient approach, called emphTrustable Co-label Learning (TCL)
- Score: 68.59187658490804
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Supervised deep learning depends on massive accurately annotated examples,
which is usually impractical in many real-world scenarios. A typical
alternative is learning from multiple noisy annotators. Numerous earlier works
assume that all labels are noisy, while it is usually the case that a few
trusted samples with clean labels are available. This raises the following
important question: how can we effectively use a small amount of trusted data
to facilitate robust classifier learning from multiple annotators? This paper
proposes a data-efficient approach, called \emph{Trustable Co-label Learning}
(TCL), to learn deep classifiers from multiple noisy annotators when a small
set of trusted data is available. This approach follows the coupled-view
learning manner, which jointly learns the data classifier and the label
aggregator. It effectively uses trusted data as a guide to generate trustable
soft labels (termed co-labels). A co-label learning can then be performed by
alternately reannotating the pseudo labels and refining the classifiers. In
addition, we further improve TCL for a special complete data case, where each
instance is labeled by all annotators and the label aggregator is represented
by multilayer neural networks to enhance model capacity. Extensive experiments
on synthetic and real datasets clearly demonstrate the effectiveness and
robustness of the proposed approach. Source code is available at
https://github.com/ShikunLi/TCL
Related papers
- Channel-Wise Contrastive Learning for Learning with Noisy Labels [60.46434734808148]
We introduce channel-wise contrastive learning (CWCL) to distinguish authentic label information from noise.
Unlike conventional instance-wise contrastive learning (IWCL), CWCL tends to yield more nuanced and resilient features aligned with the authentic labels.
Our strategy is twofold: firstly, using CWCL to extract pertinent features to identify cleanly labeled samples, and secondly, progressively fine-tuning using these samples.
arXiv Detail & Related papers (2023-08-14T06:04:50Z) - Multi-Label Knowledge Distillation [86.03990467785312]
We propose a novel multi-label knowledge distillation method.
On one hand, it exploits the informative semantic knowledge from the logits by dividing the multi-label learning problem into a set of binary classification problems.
On the other hand, it enhances the distinctiveness of the learned feature representations by leveraging the structural information of label-wise embeddings.
arXiv Detail & Related papers (2023-08-12T03:19:08Z) - ActiveLab: Active Learning with Re-Labeling by Multiple Annotators [19.84626033109009]
ActiveLab is a method to decide what to label next in batch active learning.
It automatically estimates when it is more informative to re-label examples vs. labeling entirely new ones.
It reliably trains more accurate classifiers with far fewer annotations than a wide variety of popular active learning methods.
arXiv Detail & Related papers (2023-01-27T17:00:11Z) - Label-Noise Learning with Intrinsically Long-Tailed Data [65.41318436799993]
We propose a learning framework for label-noise learning with intrinsically long-tailed data.
Specifically, we propose two-stage bi-dimensional sample selection (TABASCO) to better separate clean samples from noisy samples.
arXiv Detail & Related papers (2022-08-21T07:47:05Z) - Learning from Multiple Annotator Noisy Labels via Sample-wise Label
Fusion [17.427778867371153]
In some real-world applications, accurate labeling might not be viable.
Multiple noisy labels are provided by several annotators for each data sample.
arXiv Detail & Related papers (2022-07-22T20:38:20Z) - One Positive Label is Sufficient: Single-Positive Multi-Label Learning
with Label Enhancement [71.9401831465908]
We investigate single-positive multi-label learning (SPMLL) where each example is annotated with only one relevant label.
A novel method named proposed, i.e., Single-positive MultI-label learning with Label Enhancement, is proposed.
Experiments on benchmark datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2022-06-01T14:26:30Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z) - Active Learning in Incomplete Label Multiple Instance Multiple Label
Learning [17.5720245903743]
We propose a novel bag-class pair based approach for active learning in the MIML setting.
Our approach is based on a discriminative graphical model with efficient and exact inference.
arXiv Detail & Related papers (2021-07-22T17:01:28Z) - Analysis of label noise in graph-based semi-supervised learning [2.4366811507669124]
In machine learning, one must acquire labels to help supervise a model that will be able to generalize to unseen data.
It is often the case that most of our data is unlabeled.
Semi-supervised learning (SSL) alleviates that by making strong assumptions about the relation between the labels and the input data distribution.
arXiv Detail & Related papers (2020-09-27T22:13:20Z) - Instance Credibility Inference for Few-Shot Learning [45.577880041135785]
Few-shot learning aims to recognize new objects with extremely limited training data for each category.
This paper presents a simple statistical approach, dubbed Instance Credibility Inference (ICI) to exploit the distribution support of unlabeled instances for few-shot learning.
Our simple approach can establish new state-of-the-arts on four widely used few-shot learning benchmark datasets.
arXiv Detail & Related papers (2020-03-26T12:01:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.