Optimal Sample Selection Through Uncertainty Estimation and Its
Application in Deep Learning
- URL: http://arxiv.org/abs/2309.02476v1
- Date: Tue, 5 Sep 2023 14:06:33 GMT
- Title: Optimal Sample Selection Through Uncertainty Estimation and Its
Application in Deep Learning
- Authors: Yong Lin, Chen Liu, Chenlu Ye, Qing Lian, Yuan Yao, Tong Zhang
- Abstract summary: We present a theoretically optimal solution for addressing both coreset selection and active learning.
Our proposed method, COPS, is designed to minimize the expected loss of a model trained on subsampled data.
- Score: 22.410220040736235
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern deep learning heavily relies on large labeled datasets, which often
comse with high costs in terms of both manual labeling and computational
resources. To mitigate these challenges, researchers have explored the use of
informative subset selection techniques, including coreset selection and active
learning. Specifically, coreset selection involves sampling data with both
input ($\bx$) and output ($\by$), active learning focuses solely on the input
data ($\bx$).
In this study, we present a theoretically optimal solution for addressing
both coreset selection and active learning within the context of linear softmax
regression. Our proposed method, COPS (unCertainty based OPtimal Sub-sampling),
is designed to minimize the expected loss of a model trained on subsampled
data. Unlike existing approaches that rely on explicit calculations of the
inverse covariance matrix, which are not easily applicable to deep learning
scenarios, COPS leverages the model's logits to estimate the sampling ratio.
This sampling ratio is closely associated with model uncertainty and can be
effectively applied to deep learning tasks. Furthermore, we address the
challenge of model sensitivity to misspecification by incorporating a
down-weighting approach for low-density samples, drawing inspiration from
previous works.
To assess the effectiveness of our proposed method, we conducted extensive
empirical experiments using deep neural networks on benchmark datasets. The
results consistently showcase the superior performance of COPS compared to
baseline methods, reaffirming its efficacy.
Related papers
- Annotation-Efficient Polyp Segmentation via Active Learning [45.59503015577479]
We propose a deep active learning framework for annotation-efficient polyp segmentation.
In practice, we measure the uncertainty of each sample by examining the similarity between features masked by the prediction map of the polyp and the background area.
We show that our proposed method achieved state-of-the-art performance compared to other competitors on both a public dataset and a large-scale in-house dataset.
arXiv Detail & Related papers (2024-03-21T12:25:17Z) - Querying Easily Flip-flopped Samples for Deep Active Learning [63.62397322172216]
Active learning is a machine learning paradigm that aims to improve the performance of a model by strategically selecting and querying unlabeled data.
One effective selection strategy is to base it on the model's predictive uncertainty, which can be interpreted as a measure of how informative a sample is.
This paper proposes the it least disagree metric (LDM) as the smallest probability of disagreement of the predicted label.
arXiv Detail & Related papers (2024-01-18T08:12:23Z) - Sample Complexity of Preference-Based Nonparametric Off-Policy
Evaluation with Deep Networks [58.469818546042696]
We study the sample efficiency of OPE with human preference and establish a statistical guarantee for it.
By appropriately selecting the size of a ReLU network, we show that one can leverage any low-dimensional manifold structure in the Markov decision process.
arXiv Detail & Related papers (2023-10-16T16:27:06Z) - Temporal Output Discrepancy for Loss Estimation-based Active Learning [65.93767110342502]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2022-12-20T19:29:37Z) - Mitigating Sampling Bias and Improving Robustness in Active Learning [13.994967246046008]
We introduce supervised contrastive active learning by leveraging the contrastive loss for active learning under a supervised setting.
We propose an unbiased query strategy that selects informative data samples of diverse feature representations.
We empirically demonstrate our proposed methods reduce sampling bias, achieve state-of-the-art accuracy and model calibration in an active learning setup.
arXiv Detail & Related papers (2021-09-13T20:58:40Z) - Semi-Supervised Active Learning with Temporal Output Discrepancy [42.01906895756629]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2021-07-29T16:25:56Z) - Local policy search with Bayesian optimization [73.0364959221845]
Reinforcement learning aims to find an optimal policy by interaction with an environment.
Policy gradients for local search are often obtained from random perturbations.
We develop an algorithm utilizing a probabilistic model of the objective function and its gradient.
arXiv Detail & Related papers (2021-06-22T16:07:02Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - DEALIO: Data-Efficient Adversarial Learning for Imitation from
Observation [57.358212277226315]
In imitation learning from observation IfO, a learning agent seeks to imitate a demonstrating agent using only observations of the demonstrated behavior without access to the control signals generated by the demonstrator.
Recent methods based on adversarial imitation learning have led to state-of-the-art performance on IfO problems, but they typically suffer from high sample complexity due to a reliance on data-inefficient, model-free reinforcement learning algorithms.
This issue makes them impractical to deploy in real-world settings, where gathering samples can incur high costs in terms of time, energy, and risk.
We propose a more data-efficient IfO algorithm
arXiv Detail & Related papers (2021-03-31T23:46:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.