DIVINE: Diverse Influential Training Points for Data Visualization and
Model Refinement
- URL: http://arxiv.org/abs/2107.05978v1
- Date: Tue, 13 Jul 2021 10:50:58 GMT
- Title: DIVINE: Diverse Influential Training Points for Data Visualization and
Model Refinement
- Authors: Umang Bhatt, Isabel Chien, Muhammad Bilal Zafar, Adrian Weller
- Abstract summary: We propose a method to select a set of DIVerse INfluEntial (DIVINE) training points as a useful explanation of model behavior.
Our method can identify unfairness-inducing training points, which can be removed to improve fairness outcomes.
- Score: 32.045420977032926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As the complexity of machine learning (ML) models increases, resulting in a
lack of prediction explainability, several methods have been developed to
explain a model's behavior in terms of the training data points that most
influence the model. However, these methods tend to mark outliers as highly
influential points, limiting the insights that practitioners can draw from
points that are not representative of the training data. In this work, we take
a step towards finding influential training points that also represent the
training data well. We first review methods for assigning importance scores to
training points. Given importance scores, we propose a method to select a set
of DIVerse INfluEntial (DIVINE) training points as a useful explanation of
model behavior. As practitioners might not only be interested in finding data
points influential with respect to model accuracy, but also with respect to
other important metrics, we show how to evaluate training data points on the
basis of group fairness. Our method can identify unfairness-inducing training
points, which can be removed to improve fairness outcomes. Our quantitative
experiments and user studies show that visualizing DIVINE points helps
practitioners understand and explain model behavior better than earlier
approaches.
Related papers
- How to Achieve Higher Accuracy with Less Training Points? [2.1834099301440526]
We propose a technique based on influence functions to determine which training samples should be included in the training set.
Our approach demonstrates performance comparable to that of training on the entire dataset while using only 10% of the data.
arXiv Detail & Related papers (2025-04-18T09:38:26Z) - How to Probe: Simple Yet Effective Techniques for Improving Post-hoc Explanations [69.72654127617058]
Post-hoc importance attribution methods are a popular tool for "explaining" Deep Neural Networks (DNNs)
In this work we bring forward empirical evidence that challenges this very notion.
We discover a strong dependency on and demonstrate that the training details of a pre-trained model's classification layer play a crucial role.
arXiv Detail & Related papers (2025-03-01T22:25:11Z) - Capturing the Temporal Dependence of Training Data Influence [100.91355498124527]
We formalize the concept of trajectory-specific leave-one-out influence, which quantifies the impact of removing a data point during training.
We propose data value embedding, a novel technique enabling efficient approximation of trajectory-specific LOO.
As data value embedding captures training data ordering, it offers valuable insights into model training dynamics.
arXiv Detail & Related papers (2024-12-12T18:28:55Z) - What Do Learning Dynamics Reveal About Generalization in LLM Reasoning? [83.83230167222852]
We find that a model's generalization behavior can be effectively characterized by a training metric we call pre-memorization train accuracy.
By connecting a model's learning behavior to its generalization, pre-memorization train accuracy can guide targeted improvements to training strategies.
arXiv Detail & Related papers (2024-11-12T09:52:40Z) - Distilled Datamodel with Reverse Gradient Matching [74.75248610868685]
We introduce an efficient framework for assessing data impact, comprising offline training and online evaluation stages.
Our proposed method achieves comparable model behavior evaluation while significantly speeding up the process compared to the direct retraining method.
arXiv Detail & Related papers (2024-04-22T09:16:14Z) - Unlearning Traces the Influential Training Data of Language Models [31.33791825286853]
This paper presents UnTrac: unlearning traces the influence of a training dataset on the model's performance.
We propose a more scalable approach, UnTrac-Inv, which unlearns a test dataset and evaluates the unlearned model on training datasets.
arXiv Detail & Related papers (2024-01-26T23:17:31Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Understanding new tasks through the lens of training data via
exponential tilting [43.33775132139584]
We consider the problem of reweighing the training samples to gain insights into the distribution of the target task.
We formulate a distribution shift model based on the exponential tilt assumption and learn train data importance weights.
The learned train data weights can then be used for downstream tasks such as target performance evaluation, fine-tuning, and model selection.
arXiv Detail & Related papers (2022-05-26T18:38:43Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - FairIF: Boosting Fairness in Deep Learning via Influence Functions with
Validation Set Sensitive Attributes [51.02407217197623]
We propose a two-stage training algorithm named FAIRIF.
It minimizes the loss over the reweighted data set where the sample weights are computed.
We show that FAIRIF yields models with better fairness-utility trade-offs against various types of bias.
arXiv Detail & Related papers (2022-01-15T05:14:48Z) - Efficient Estimation of Influence of a Training Instance [56.29080605123304]
We propose an efficient method for estimating the influence of a training instance on a neural network model.
Our method is inspired by dropout, which zero-masks a sub-network and prevents the sub-network from learning each training instance.
We demonstrate that the proposed method can capture training influences, enhance the interpretability of error predictions, and cleanse the training dataset for improving generalization.
arXiv Detail & Related papers (2020-12-08T04:31:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.