Margin-based sampling in high dimensions: When being active is less
efficient than staying passive
- URL: http://arxiv.org/abs/2212.00772v2
- Date: Fri, 2 Jun 2023 12:45:17 GMT
- Title: Margin-based sampling in high dimensions: When being active is less
efficient than staying passive
- Authors: Alexandru Tifrea, Jacob Clarysse, Fanny Yang
- Abstract summary: Recent empirical evidence suggests that margin-based active learning can sometimes perform even worse than passive learning.
We prove for logistic regression that PL outperforms margin-based AL even for noiseless data.
Insights from our proof indicate that this high-dimensional phenomenon is exacerbated when the separation between the classes is small.
- Score: 76.71565772067113
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is widely believed that given the same labeling budget, active learning
(AL) algorithms like margin-based active learning achieve better predictive
performance than passive learning (PL), albeit at a higher computational cost.
Recent empirical evidence suggests that this added cost might be in vain, as
margin-based AL can sometimes perform even worse than PL. While existing works
offer different explanations in the low-dimensional regime, this paper shows
that the underlying mechanism is entirely different in high dimensions: we
prove for logistic regression that PL outperforms margin-based AL even for
noiseless data and when using the Bayes optimal decision boundary for sampling.
Insights from our proof indicate that this high-dimensional phenomenon is
exacerbated when the separation between the classes is small. We corroborate
this intuition with experiments on 20 high-dimensional datasets spanning a
diverse range of applications, from finance and histology to chemistry and
computer vision.
Related papers
- Uncertainty Aware Learning for Language Model Alignment [97.36361196793929]
We propose uncertainty-aware learning (UAL) to improve the model alignment of different task scenarios.
We implement UAL in a simple fashion -- adaptively setting the label smoothing value of training according to the uncertainty of individual samples.
Experiments on widely used benchmarks demonstrate that our UAL significantly and consistently outperforms standard supervised fine-tuning.
arXiv Detail & Related papers (2024-06-07T11:37:45Z) - Simple Ingredients for Offline Reinforcement Learning [86.1988266277766]
offline reinforcement learning algorithms have proven effective on datasets highly connected to the target downstream task.
We show that existing methods struggle with diverse data: their performance considerably deteriorates as data collected for related but different tasks is simply added to the offline buffer.
We show that scale, more than algorithmic considerations, is the key factor influencing performance.
arXiv Detail & Related papers (2024-03-19T18:57:53Z) - Improve Cost Efficiency of Active Learning over Noisy Dataset [1.3846014191157405]
In this paper, we consider cases of binary classification, where acquiring a positive instance incurs a significantly higher cost compared to that of negative instances.
We propose a shifted normal distribution sampling function that samples from a wider range than typical uncertainty sampling.
Our simulation underscores that our proposed sampling function limits both noisy and positive label selection, delivering between 20% and 32% improved cost efficiency over different test datasets.
arXiv Detail & Related papers (2024-03-02T23:53:24Z) - Direct Acquisition Optimization for Low-Budget Active Learning [15.355195433709717]
Active Learning (AL) has gained prominence in integrating data-intensive machine learning (ML) models into domains with limited labeled data.
In this paper, we first empirically observe the performance degradation of existing AL algorithms in the low-budget settings.
We then introduce Direct Acquisition Optimization (DAO), a novel AL algorithm that optimize sample selections based on expected true loss reduction.
arXiv Detail & Related papers (2024-02-08T20:36:21Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Active learning for reducing labeling effort in text classification
tasks [3.8424737607413153]
Active learning (AL) is a paradigm that aims to reduce labeling effort by only using the data which the used model deems most informative.
We present an empirical study that compares different uncertainty-based algorithms BERT$_base$ as the used classifiers.
Our results show that using uncertainty-based AL with BERT$base$ outperforms random sampling of data.
arXiv Detail & Related papers (2021-09-10T13:00:36Z) - Effective Evaluation of Deep Active Learning on Image Classification
Tasks [10.27095298129151]
We present a unified re-implementation of state-of-the-art active learning algorithms in the context of image classification.
On the positive side, we show that AL techniques are 2x to 4x more label-efficient compared to RS with the use of data augmentation.
arXiv Detail & Related papers (2021-06-16T23:29:39Z) - An Exponential Lower Bound for Linearly-Realizable MDPs with Constant
Suboptimality Gap [66.75488143823337]
We show that an exponential sample complexity lower bound still holds even if a constant suboptimality gap is assumed.
Perhaps surprisingly, this implies an exponential separation between the online RL setting and the generative model setting.
arXiv Detail & Related papers (2021-03-23T17:05:54Z) - Provably Efficient Reward-Agnostic Navigation with Linear Value
Iteration [143.43658264904863]
We show how iteration under a more standard notion of low inherent Bellman error, typically employed in least-square value-style algorithms, can provide strong PAC guarantees on learning a near optimal value function.
We present a computationally tractable algorithm for the reward-free setting and show how it can be used to learn a near optimal policy for any (linear) reward function.
arXiv Detail & Related papers (2020-08-18T04:34:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.