Relabeling Minimal Training Subset to Flip a Prediction
- URL: http://arxiv.org/abs/2305.12809v4
- Date: Sat, 3 Feb 2024 07:28:51 GMT
- Title: Relabeling Minimal Training Subset to Flip a Prediction
- Authors: Jinghan Yang, Linjie Xu, Lequan Yu
- Abstract summary: We find that relabeling fewer than 2% of the training points can always flip a prediction.
We show that $|mathcalS_t|$ is highly related to the noise ratio in the training set and $|mathcalS_t|$ is correlated with but complementary to predicted probabilities.
- Score: 20.708004593740004
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When facing an unsatisfactory prediction from a machine learning model, users
can be interested in investigating the underlying reasons and exploring the
potential for reversing the outcome. We ask: To flip the prediction on a test
point $x_t$, how to identify the smallest training subset $\mathcal{S}_t$ that
we need to relabel? We propose an efficient algorithm to identify and relabel
such a subset via an extended influence function for binary classification
models with convex loss. We find that relabeling fewer than 2% of the training
points can always flip a prediction. This mechanism can serve multiple
purposes: (1) providing an approach to challenge a model prediction by altering
training points; (2) evaluating model robustness with the cardinality of the
subset (i.e., $|\mathcal{S}_t|$); we show that $|\mathcal{S}_t|$ is highly
related to the noise ratio in the training set and $|\mathcal{S}_t|$ is
correlated with but complementary to predicted probabilities; and (3) revealing
training points lead to group attribution bias. To the best of our knowledge,
we are the first to investigate identifying and relabeling the minimal training
subset required to flip a given prediction.
Related papers
- Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Self-Directed Linear Classification [50.659479930171585]
In online classification, a learner aims to predict their labels in an online fashion so as to minimize the total number of mistakes.
Here we study the power of choosing the prediction order and establish the first strong separation between worst-order and random-order learning.
arXiv Detail & Related papers (2023-08-06T15:38:44Z) - How Many and Which Training Points Would Need to be Removed to Flip this
Prediction? [34.9118528281516]
We consider the problem of identifying a minimal subset of training data $mathcalS_t$.
If the instances comprising $mathcalS_t$ had been removed prior to training, the categorization of a given test point $x_t$ would have been different.
We propose comparatively fast approximation methods to find $mathcalS_t$ based on influence functions.
arXiv Detail & Related papers (2023-02-04T13:55:12Z) - Memorize to Generalize: on the Necessity of Interpolation in High
Dimensional Linear Regression [6.594338220264161]
achieve optimal predictive risk in machine learning problems requires interpolating the training data.
We characterize how prediction (test) error necessarily scales with training error in this setting.
optimal performance requires fitting training data to substantially higher accuracy than the inherent noise floor of the problem.
arXiv Detail & Related papers (2022-02-20T18:51:45Z) - Datamodels: Predicting Predictions from Training Data [86.66720175866415]
We present a conceptual framework, datamodeling, for analyzing the behavior of a model class in terms of the training data.
We show that even simple linear datamodels can successfully predict model outputs.
arXiv Detail & Related papers (2022-02-01T18:15:24Z) - Thought Flow Nets: From Single Predictions to Trains of Model Thought [39.619001911390804]
When humans solve complex problems, they rarely come up with a decision right-away.
Instead, they start with an intuitive decision reflecting upon it, spot mistakes, resolve contradictions and jump between different hypotheses.
arXiv Detail & Related papers (2021-07-26T13:56:37Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z) - Towards optimally abstaining from prediction [22.937799541125607]
A common challenge across all areas of machine learning is that training data is not distributed like test data.
We consider a model where one may abstain from predicting, at a fixed cost.
Our work builds on a recent abstention algorithm of Goldwasser, Kalais, and Montasser ( 2020) for transductive binary classification.
arXiv Detail & Related papers (2021-05-28T21:44:48Z) - Outlier-robust sparse/low-rank least-squares regression and robust
matrix completion [1.0878040851637998]
We study high-dimensional least-squares regression within a subgaussian statistical learning framework with heterogeneous noise.
We also present a novel theory of trace-regression with matrix decomposition based on a new application of the product process.
arXiv Detail & Related papers (2020-12-12T07:42:47Z) - Probing Model Signal-Awareness via Prediction-Preserving Input
Minimization [67.62847721118142]
We evaluate models' ability to capture the correct vulnerability signals to produce their predictions.
We measure the signal awareness of models using a new metric we propose- Signal-aware Recall (SAR)
The results show a sharp drop in the model's Recall from the high 90s to sub-60s with the new metric.
arXiv Detail & Related papers (2020-11-25T20:05:23Z) - L2R2: Leveraging Ranking for Abductive Reasoning [65.40375542988416]
The abductive natural language inference task ($alpha$NLI) is proposed to evaluate the abductive reasoning ability of a learning system.
A novel $L2R2$ approach is proposed under the learning-to-rank framework.
Experiments on the ART dataset reach the state-of-the-art in the public leaderboard.
arXiv Detail & Related papers (2020-05-22T15:01:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.