A Robust Optimization Method for Label Noisy Datasets Based on Adaptive
Threshold: Adaptive-k
- URL: http://arxiv.org/abs/2203.14165v1
- Date: Sat, 26 Mar 2022 21:48:12 GMT
- Title: A Robust Optimization Method for Label Noisy Datasets Based on Adaptive
Threshold: Adaptive-k
- Authors: Enes Dedeoglu, Himmet Toprak Kesgin, Mehmet Fatih Amasyali
- Abstract summary: SGD does not produce robust results on datasets with label noise.
In this paper, we recommend using samples with loss less than a threshold value determined during the optimization process, instead of using all samples in the mini-batch.
Our proposed method, Adaptive-k, aims to exclude label noise samples from the optimization process and make the process robust.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: SGD does not produce robust results on datasets with label noise. Because the
gradients calculated according to the losses of the noisy samples cause the
optimization process to go in the wrong direction. In this paper, as an
alternative to SGD, we recommend using samples with loss less than a threshold
value determined during the optimization process, instead of using all samples
in the mini-batch. Our proposed method, Adaptive-k, aims to exclude label noise
samples from the optimization process and make the process robust. On noisy
datasets, we found that using a threshold-based approach, such as Adaptive-k,
produces better results than using all samples or a fixed number of low-loss
samples in the mini-batch. Based on our theoretical analysis and experimental
results, we show that the Adaptive-k method is closest to the performance of
the oracle, in which noisy samples are entirely removed from the dataset.
Adaptive-k is a simple but effective method. It does not require prior
knowledge of the noise ratio of the dataset, does not require additional model
training, and does not increase training time significantly. The code for
Adaptive-k is available at https://github.com/enesdedeoglu-TR/Adaptive-k
Related papers
- ANNE: Adaptive Nearest Neighbors and Eigenvector-based Sample Selection for Robust Learning with Noisy Labels [7.897299759691143]
This paper introduces the Adaptive Nearest Neighbors and Eigenvector-based (ANNE) sample selection methodology.
ANNE integrates loss-based sampling with the feature-based sampling methods FINE and Adaptive KNN to optimize performance across a wide range of noise rate scenarios.
arXiv Detail & Related papers (2024-11-03T15:51:38Z) - Double Correction Framework for Denoising Recommendation [45.98207284259792]
In implicit feedback, noisy samples can affect precise user preference learning.
A popular solution is based on dropping noisy samples in the model training phase.
We propose a Double Correction Framework for Denoising Recommendation.
arXiv Detail & Related papers (2024-05-18T12:15:10Z) - Regroup Median Loss for Combating Label Noise [19.51996047333779]
Deep model training requires large-scale datasets of annotated data.
Due to the difficulty of annotating a large number of samples, label noise caused by incorrect annotations is inevitable.
We propose Regroup Median Loss (RML) to reduce the probability of selecting noisy samples and correct losses of noisy samples.
arXiv Detail & Related papers (2023-12-11T10:19:55Z) - Let's Sample Step by Step: Adaptive-Consistency for Efficient Reasoning
and Coding with LLMs [60.58434523646137]
A popular approach for improving the correctness of output from large language models (LLMs) is Self-Consistency.
We introduce Adaptive-Consistency, a cost-efficient, model-agnostic technique that dynamically adjusts the number of samples per question.
Our experiments show that Adaptive-Consistency reduces sample budget by up to 7.9 times with an average accuracy drop of less than 0.1%.
arXiv Detail & Related papers (2023-05-19T17:49:25Z) - Towards Automated Imbalanced Learning with Deep Hierarchical
Reinforcement Learning [57.163525407022966]
Imbalanced learning is a fundamental challenge in data mining, where there is a disproportionate ratio of training samples in each class.
Over-sampling is an effective technique to tackle imbalanced learning through generating synthetic samples for the minority class.
We propose AutoSMOTE, an automated over-sampling algorithm that can jointly optimize different levels of decisions.
arXiv Detail & Related papers (2022-08-26T04:28:01Z) - Adaptive Sketches for Robust Regression with Importance Sampling [64.75899469557272]
We introduce data structures for solving robust regression through gradient descent (SGD)
Our algorithm effectively runs $T$ steps of SGD with importance sampling while using sublinear space and just making a single pass over the data.
arXiv Detail & Related papers (2022-07-16T03:09:30Z) - Robust Meta-learning with Sampling Noise and Label Noise via
Eigen-Reptile [78.1212767880785]
meta-learner is prone to overfitting since there are only a few available samples.
When handling the data with noisy labels, the meta-learner could be extremely sensitive to label noise.
We present Eigen-Reptile (ER) that updates the meta- parameters with the main direction of historical task-specific parameters.
arXiv Detail & Related papers (2022-06-04T08:48:02Z) - Adaptive Sample Selection for Robust Learning under Label Noise [1.71982924656402]
Deep Neural Networks (DNNs) have been shown to be susceptible to memorization or overfitting in the presence of noisily labelled data.
A prominent class of algorithms rely on sample selection strategies, motivated by curriculum learning.
We propose a data-dependent, adaptive sample selection strategy that relies only on batch statistics.
arXiv Detail & Related papers (2021-06-29T12:10:58Z) - Efficient SVDD Sampling with Approximation Guarantees for the Decision
Boundary [7.251418581794502]
Support Vector Data Description (SVDD) is a popular one-class classifier for anomaly and novelty detection.
Despite its effectiveness, SVDD does not scale well with data size.
In this article, we study how to select a sample considering these points.
Our approach is to frame SVDD sampling as an optimization problem, where constraints guarantee that sampling indeed approximates the original decision boundary.
arXiv Detail & Related papers (2020-09-29T08:28:01Z) - Least Squares Regression with Markovian Data: Fundamental Limits and
Algorithms [69.45237691598774]
We study the problem of least squares linear regression where the data-points are dependent and are sampled from a Markov chain.
We establish sharp information theoretic minimax lower bounds for this problem in terms of $tau_mathsfmix$.
We propose an algorithm based on experience replay--a popular reinforcement learning technique--that achieves a significantly better error rate.
arXiv Detail & Related papers (2020-06-16T04:26:50Z) - Non-Adaptive Adaptive Sampling on Turnstile Streams [57.619901304728366]
We give the first relative-error algorithms for column subset selection, subspace approximation, projective clustering, and volume on turnstile streams that use space sublinear in $n$.
Our adaptive sampling procedure has a number of applications to various data summarization problems that either improve state-of-the-art or have only been previously studied in the more relaxed row-arrival model.
arXiv Detail & Related papers (2020-04-23T05:00:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.