Efficient Estimation of Influence of a Training Instance
- URL: http://arxiv.org/abs/2012.04207v1
- Date: Tue, 8 Dec 2020 04:31:38 GMT
- Title: Efficient Estimation of Influence of a Training Instance
- Authors: Sosuke Kobayashi, Sho Yokoi, Jun Suzuki, Kentaro Inui
- Abstract summary: We propose an efficient method for estimating the influence of a training instance on a neural network model.
Our method is inspired by dropout, which zero-masks a sub-network and prevents the sub-network from learning each training instance.
We demonstrate that the proposed method can capture training influences, enhance the interpretability of error predictions, and cleanse the training dataset for improving generalization.
- Score: 56.29080605123304
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding the influence of a training instance on a neural network model
leads to improving interpretability. However, it is difficult and inefficient
to evaluate the influence, which shows how a model's prediction would be
changed if a training instance were not used. In this paper, we propose an
efficient method for estimating the influence. Our method is inspired by
dropout, which zero-masks a sub-network and prevents the sub-network from
learning each training instance. By switching between dropout masks, we can use
sub-networks that learned or did not learn each training instance and estimate
its influence. Through experiments with BERT and VGGNet on classification
datasets, we demonstrate that the proposed method can capture training
influences, enhance the interpretability of error predictions, and cleanse the
training dataset for improving generalization.
Related papers
- Empirical influence functions to understand the logic of fine-tuning [1.9116784879310031]
We use empirical influence measured using fine-tuning to demonstrate how individual training samples affect outputs.
We show that these desiderata are violated for both for simple convolutional networks and for a modern LLM.
Our results suggest that popular models cannot generalize or perform logic in the way they appear.
arXiv Detail & Related papers (2024-06-01T17:31:06Z) - Distilled Datamodel with Reverse Gradient Matching [74.75248610868685]
We introduce an efficient framework for assessing data impact, comprising offline training and online evaluation stages.
Our proposed method achieves comparable model behavior evaluation while significantly speeding up the process compared to the direct retraining method.
arXiv Detail & Related papers (2024-04-22T09:16:14Z) - The Mirrored Influence Hypothesis: Efficient Data Influence Estimation by Harnessing Forward Passes [30.30769701138665]
We introduce and explore the Mirrored Influence Hypothesis, highlighting a reciprocal nature of influence between training and test data.
Specifically, it suggests that evaluating the influence of training data on test predictions can be reformulated as an equivalent, yet inverse problem.
We introduce a new method for estimating the influence of training data, which requires calculating gradients for specific test samples, paired with a forward pass for each training point.
arXiv Detail & Related papers (2024-02-14T03:43:05Z) - Unlearning Traces the Influential Training Data of Language Models [31.33791825286853]
This paper presents UnTrac: unlearning traces the influence of a training dataset on the model's performance.
We propose a more scalable approach, UnTrac-Inv, which unlearns a test dataset and evaluates the unlearned model on training datasets.
arXiv Detail & Related papers (2024-01-26T23:17:31Z) - Agree to Disagree: Diversity through Disagreement for Better
Transferability [54.308327969778155]
We propose D-BAT (Diversity-By-disAgreement Training), which enforces agreement among the models on the training data.
We show how D-BAT naturally emerges from the notion of generalized discrepancy.
arXiv Detail & Related papers (2022-02-09T12:03:02Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - Self-Adaptive Training: Bridging the Supervised and Self-Supervised
Learning [16.765461276790944]
Self-adaptive training is a unified training algorithm that dynamically calibrates and enhances training process by model predictions without incurring extra computational cost.
We analyze the training dynamics of deep networks on training data corrupted by, e.g., random noise and adversarial examples.
Our analysis shows that model predictions are able to magnify useful underlying information in data and this phenomenon occurs broadly even in the absence of emphany label information.
arXiv Detail & Related papers (2021-01-21T17:17:30Z) - Using Cross-Loss Influence Functions to Explain Deep Network
Representations [1.7778609937758327]
We show that influence functions can be extended to handle mismatched training and testing settings.
Our result enables us to compute the influence of unsupervised and self-supervised training examples with respect to a supervised test objective.
arXiv Detail & Related papers (2020-12-03T03:43:26Z) - Influence Functions in Deep Learning Are Fragile [52.31375893260445]
influence functions approximate the effect of samples in test-time predictions.
influence estimates are fairly accurate for shallow networks.
Hessian regularization is important to get highquality influence estimates.
arXiv Detail & Related papers (2020-06-25T18:25:59Z) - Learning Diverse Representations for Fast Adaptation to Distribution
Shift [78.83747601814669]
We present a method for learning multiple models, incorporating an objective that pressures each to learn a distinct way to solve the task.
We demonstrate our framework's ability to facilitate rapid adaptation to distribution shift.
arXiv Detail & Related papers (2020-06-12T12:23:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.