Hitting the Target: Stopping Active Learning at the Cost-Based Optimum
- URL: http://arxiv.org/abs/2110.03802v1
- Date: Thu, 7 Oct 2021 21:33:34 GMT
- Title: Hitting the Target: Stopping Active Learning at the Cost-Based Optimum
- Authors: Zac Pullar-Strecker, Katharina Dost, Eibe Frank, J\"org Wicker
- Abstract summary: Active learning allows machine learning models to be trained using fewer labels while retaining similar performance to traditional fully supervised learning.
This paper is the first to give actionable advice to practitioners on what stopping criteria they should use in a given real-world scenario.
We contribute the first large-scale comparison of stopping criteria, using a cost measure to quantify the accuracy/label trade-off, public implementations of all stopping criteria we evaluate, and an open-source framework for evaluating stopping criteria.
- Score: 1.1756822700775666
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Active learning allows machine learning models to be trained using fewer
labels while retaining similar performance to traditional fully supervised
learning. An active learner selects the most informative data points, requests
their labels, and retrains itself. While this approach is promising, it leaves
an open problem of how to determine when the model is `good enough' without the
additional labels required for traditional evaluation. In the past, different
stopping criteria have been proposed aiming to identify the optimal stopping
point. However, optimality can only be expressed as a domain-dependent
trade-off between accuracy and the number of labels, and no criterion is
superior in all applications. This paper is the first to give actionable advice
to practitioners on what stopping criteria they should use in a given
real-world scenario. We contribute the first large-scale comparison of stopping
criteria, using a cost measure to quantify the accuracy/label trade-off, public
implementations of all stopping criteria we evaluate, and an open-source
framework for evaluating stopping criteria. Our research enables practitioners
to substantially reduce labelling costs by utilizing the stopping criterion
which best suits their domain.
Related papers
- One-bit Supervision for Image Classification: Problem, Solution, and
Beyond [114.95815360508395]
This paper presents one-bit supervision, a novel setting of learning with fewer labels, for image classification.
We propose a multi-stage training paradigm and incorporate negative label suppression into an off-the-shelf semi-supervised learning algorithm.
In multiple benchmarks, the learning efficiency of the proposed approach surpasses that using full-bit, semi-supervised supervision.
arXiv Detail & Related papers (2023-11-26T07:39:00Z) - Fair Classifiers that Abstain without Harm [24.90899074869189]
In critical applications, it is vital for classifiers to defer decision-making to humans.
We propose a post-hoc method that makes existing classifiers selectively abstain from predicting certain samples.
Our framework outperforms existing methods in terms of fairness disparity without sacrificing accuracy at similar abstention rates.
arXiv Detail & Related papers (2023-10-09T23:07:28Z) - Canary in a Coalmine: Better Membership Inference with Ensembled
Adversarial Queries [53.222218035435006]
We use adversarial tools to optimize for queries that are discriminative and diverse.
Our improvements achieve significantly more accurate membership inference than existing methods.
arXiv Detail & Related papers (2022-10-19T17:46:50Z) - Is margin all you need? An extensive empirical study of active learning
on tabular data [66.18464006872345]
We analyze the performance of a variety of active learning algorithms on 69 real-world datasets from the OpenML-CC18 benchmark.
Surprisingly, we find that the classical margin sampling technique matches or outperforms all others, including current state-of-art.
arXiv Detail & Related papers (2022-10-07T21:18:24Z) - Optimizing Partial Area Under the Top-k Curve: Theory and Practice [151.5072746015253]
We develop a novel metric named partial Area Under the top-k Curve (AUTKC)
AUTKC has a better discrimination ability, and its Bayes optimal score function could give a correct top-K ranking with respect to the conditional probability.
We present an empirical surrogate risk minimization framework to optimize the proposed metric.
arXiv Detail & Related papers (2022-09-03T11:09:13Z) - Cost-Effective Online Contextual Model Selection [14.094350329970537]
We formulate this task as an online contextual active model selection problem, where at each round the learner receives an unlabeled data point along with a context.
The goal is to output the best model for any given context without obtaining an excessive amount of labels.
We propose a contextual active model selection algorithm (CAMS), which relies on a novel uncertainty sampling query criterion defined on a given policy class for adaptive model selection.
arXiv Detail & Related papers (2022-07-13T08:22:22Z) - Debiased Pseudo Labeling in Self-Training [77.83549261035277]
Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets.
To mitigate the requirement for labeled data, self-training is widely used in both academia and industry by pseudo labeling on readily-available unlabeled data.
We propose Debiased, in which the generation and utilization of pseudo labels are decoupled by two independent heads.
arXiv Detail & Related papers (2022-02-15T02:14:33Z) - Beyond traditional assumptions in fair machine learning [5.029280887073969]
This thesis scrutinizes common assumptions underlying traditional machine learning approaches to fairness in consequential decision making.
We show that group fairness criteria purely based on statistical properties of observed data are fundamentally limited.
We overcome the assumption that sensitive data is readily available in practice.
arXiv Detail & Related papers (2021-01-29T09:02:15Z) - Stopping criterion for active learning based on deterministic
generalization bounds [4.518012967046983]
We propose a criterion for automatically stopping active learning.
The proposed stopping criterion is based on the difference in the expected generalization errors and hypothesis testing.
We demonstrate the effectiveness of the proposed method via experiments with both artificial and real datasets.
arXiv Detail & Related papers (2020-05-15T08:15:47Z) - Progressive Identification of True Labels for Partial-Label Learning [112.94467491335611]
Partial-label learning (PLL) is a typical weakly supervised learning problem, where each training instance is equipped with a set of candidate labels among which only one is the true label.
Most existing methods elaborately designed as constrained optimizations that must be solved in specific manners, making their computational complexity a bottleneck for scaling up to big data.
This paper proposes a novel framework of classifier with flexibility on the model and optimization algorithm.
arXiv Detail & Related papers (2020-02-19T08:35:15Z) - Active Bayesian Assessment for Black-Box Classifiers [20.668691047355072]
We introduce an active Bayesian approach for assessment of classifier performance to satisfy the desiderata of both reliability and label-efficiency.
We first develop inference strategies to quantify uncertainty for common assessment metrics such as accuracy, misclassification cost, and calibration error.
We then propose a general framework for active Bayesian assessment using inferred uncertainty to guide efficient selection of instances for labeling.
arXiv Detail & Related papers (2020-02-16T08:08:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.