Statistical Hypothesis Testing for Class-Conditional Label Noise
- URL: http://arxiv.org/abs/2103.02630v1
- Date: Wed, 3 Mar 2021 19:03:06 GMT
- Title: Statistical Hypothesis Testing for Class-Conditional Label Noise
- Authors: Rafael Poyiadzi, Weisong Yang, Niall Twomey, Raul Santos-Rodriguez
- Abstract summary: This work aims to provide machine learning practitioners with tools to answer the question: is there class-conditional flipping noise in my labels?
In particular, we present hypothesis tests to check whether a given dataset of instance-label pairs has been corrupted with class-conditional label noise.
- Score: 3.6895394817068357
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we aim to provide machine learning practitioners with tools to
answer the question: is there class-conditional flipping noise in my labels? In
particular, we present hypothesis tests to reliably check whether a given
dataset of instance-label pairs has been corrupted with class-conditional label
noise. While previous works explore the direct estimation of the noise rates,
this is known to be hard in practice and does not offer a real understanding of
how trustworthy the estimates are. These methods typically require anchor
points - examples whose true posterior is either 0 or 1. Differently, in this
paper we assume we have access to a set of anchor points whose true posterior
is approximately 1/2. The proposed hypothesis tests are built upon the
asymptotic properties of Maximum Likelihood Estimators for Logistic Regression
models and accurately distinguish the presence of class-conditional noise from
uniform noise. We establish the main properties of the tests, including a
theoretical and empirical analysis of the dependence of the power on the test
on the training sample size, the number of anchor points, the difference of the
noise rates and the use of realistic relaxed anchors.
Related papers
- Mitigating LLM Hallucinations via Conformal Abstention [70.83870602967625]
We develop a principled procedure for determining when a large language model should abstain from responding in a general domain.
We leverage conformal prediction techniques to develop an abstention procedure that benefits from rigorous theoretical guarantees on the hallucination rate (error rate)
Experimentally, our resulting conformal abstention method reliably bounds the hallucination rate on various closed-book, open-domain generative question answering datasets.
arXiv Detail & Related papers (2024-04-04T11:32:03Z) - Hypothesis Testing for Class-Conditional Noise Using Local Maximum
Likelihood [1.8798171797988192]
In supervised learning, automatically assessing the quality of the labels before any learning takes place remains an open research question.
In this paper we show how similar procedures can be followed when the underlying model is a product of Local Maximum Likelihood Estimation.
This different view allows for wider applicability of the tests by offering users access to a richer model class.
arXiv Detail & Related papers (2023-12-15T22:14:58Z) - Robust Online Classification: From Estimation to Denoising [14.535583931446807]
We study online classification of features into labels with general hypothesis classes.
Predictions are made using observed noisy labels and noiseless features.
The performance is measured via minimax risk when comparing against true labels.
arXiv Detail & Related papers (2023-09-04T16:17:39Z) - Centrality and Consistency: Two-Stage Clean Samples Identification for
Learning with Instance-Dependent Noisy Labels [87.48541631675889]
We propose a two-stage clean samples identification method.
First, we employ a class-level feature clustering procedure for the early identification of clean samples.
Second, for the remaining clean samples that are close to the ground truth class boundary, we propose a novel consistency-based classification method.
arXiv Detail & Related papers (2022-07-29T04:54:57Z) - Uncertainty-Aware Learning Against Label Noise on Imbalanced Datasets [23.4536532321199]
We propose an Uncertainty-aware Label Correction framework to handle label noise on imbalanced datasets.
Inspired by our observations, we propose an Uncertainty-aware Label Correction framework to handle label noise on imbalanced datasets.
arXiv Detail & Related papers (2022-07-12T11:35:55Z) - The Optimal Noise in Noise-Contrastive Learning Is Not What You Think [80.07065346699005]
We show that deviating from this assumption can actually lead to better statistical estimators.
In particular, the optimal noise distribution is different from the data's and even from a different family.
arXiv Detail & Related papers (2022-03-02T13:59:20Z) - Label noise detection under the Noise at Random model with ensemble
filters [5.994719700262245]
This work investigates the performance of ensemble noise detection under two different noise models.
We investigate the effect of class distribution on noise detection performance since it changes the total noise level observed in a dataset.
arXiv Detail & Related papers (2021-12-02T21:49:41Z) - A Second-Order Approach to Learning with Instance-Dependent Label Noise [58.555527517928596]
The presence of label noise often misleads the training of deep neural networks.
We show that the errors in human-annotated labels are more likely to be dependent on the difficulty levels of tasks.
arXiv Detail & Related papers (2020-12-22T06:36:58Z) - Robustness of Accuracy Metric and its Inspirations in Learning with
Noisy Labels [51.66448070984615]
We show that maximizing training accuracy on sufficiently many noisy samples yields an approximately optimal classifier.
For validation, we prove that a noisy validation set is reliable, addressing the critical demand of model selection.
We show characterizations of models trained with noisy labels, motivated by our theoretical results, and verify the utility of a noisy validation set.
arXiv Detail & Related papers (2020-12-08T03:37:47Z) - Confidence Scores Make Instance-dependent Label-noise Learning Possible [129.84497190791103]
In learning with noisy labels, for every instance, its label can randomly walk to other classes following a transition distribution which is named a noise model.
We introduce confidence-scored instance-dependent noise (CSIDN), where each instance-label pair is equipped with a confidence score.
We find with the help of confidence scores, the transition distribution of each instance can be approximately estimated.
arXiv Detail & Related papers (2020-01-11T16:15:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.