Influence Functions in Deep Learning Are Fragile
- URL: http://arxiv.org/abs/2006.14651v2
- Date: Wed, 10 Feb 2021 23:45:14 GMT
- Title: Influence Functions in Deep Learning Are Fragile
- Authors: Samyadeep Basu, Philip Pope, Soheil Feizi
- Abstract summary: influence functions approximate the effect of samples in test-time predictions.
influence estimates are fairly accurate for shallow networks.
Hessian regularization is important to get highquality influence estimates.
- Score: 52.31375893260445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Influence functions approximate the effect of training samples in test-time
predictions and have a wide variety of applications in machine learning
interpretability and uncertainty estimation. A commonly-used (first-order)
influence function can be implemented efficiently as a post-hoc method
requiring access only to the gradients and Hessian of the model. For linear
models, influence functions are well-defined due to the convexity of the
underlying loss function and are generally accurate even across difficult
settings where model changes are fairly large such as estimating group
influences. Influence functions, however, are not well-understood in the
context of deep learning with non-convex loss functions. In this paper, we
provide a comprehensive and large-scale empirical study of successes and
failures of influence functions in neural network models trained on datasets
such as Iris, MNIST, CIFAR-10 and ImageNet. Through our extensive experiments,
we show that the network architecture, its depth and width, as well as the
extent of model parameterization and regularization techniques have strong
effects in the accuracy of influence functions. In particular, we find that (i)
influence estimates are fairly accurate for shallow networks, while for deeper
networks the estimates are often erroneous; (ii) for certain network
architectures and datasets, training with weight-decay regularization is
important to get high-quality influence estimates; and (iii) the accuracy of
influence estimates can vary significantly depending on the examined test
points. These results suggest that in general influence functions in deep
learning are fragile and call for developing improved influence estimation
methods to mitigate these issues in non-convex setups.
Related papers
- Enhancing Training Data Attribution for Large Language Models with Fitting Error Consideration [74.09687562334682]
We introduce a novel training data attribution method called Debias and Denoise Attribution (DDA)
Our method significantly outperforms existing approaches, achieving an averaged AUC of 91.64%.
DDA exhibits strong generality and scalability across various sources and different-scale models like LLaMA2, QWEN2, and Mistral.
arXiv Detail & Related papers (2024-10-02T07:14:26Z) - Do Influence Functions Work on Large Language Models? [10.463762448166714]
Influence functions aim to quantify the impact of individual training data points on a model's predictions.
We evaluate influence functions across multiple tasks and find that they consistently perform poorly in most settings.
arXiv Detail & Related papers (2024-09-30T06:50:18Z) - Most Influential Subset Selection: Challenges, Promises, and Beyond [9.479235005673683]
We study the Most Influential Subset Selection (MISS) problem, which aims to identify a subset of training samples with the greatest collective influence.
We conduct a comprehensive analysis of the prevailing approaches in MISS, elucidating their strengths and weaknesses.
We demonstrate that an adaptive version of theses which applies them iteratively, can effectively capture the interactions among samples.
arXiv Detail & Related papers (2024-09-25T20:00:23Z) - Studying Large Language Model Generalization with Influence Functions [29.577692176892135]
Influence functions aim to answer a counterfactual: how would the model's parameters (and hence its outputs) change if a sequence were added to the training set?
We use the Eigenvalue-corrected Kronecker-Factored Approximate Curvature (EK-FAC) approximation to scale influence functions up to large language models (LLMs) with up to 52 billion parameters.
We investigate generalization patterns of LLMs, including the sparsity of the influence patterns, increasing abstraction with scale, math and programming abilities, cross-lingual generalization, and role-playing behavior.
arXiv Detail & Related papers (2023-08-07T04:47:42Z) - If Influence Functions are the Answer, Then What is the Question? [7.873458431535409]
Influence functions efficiently estimate the effect of removing a single training data point on a model's learned parameters.
While influence estimates align well with leave-one-out retraining for linear models, recent works have shown this alignment is often poor in neural networks.
arXiv Detail & Related papers (2022-09-12T16:17:43Z) - Inf-CP: A Reliable Channel Pruning based on Channel Influence [4.692400531340393]
One of the most effective methods of channel pruning is to trim on the basis of the importance of each neuron.
Previous works have proposed to trim by considering the statistics of a single layer or a plurality of successive layers of neurons.
We propose to use ensemble learning to train a model for different batches of data.
arXiv Detail & Related papers (2021-12-05T09:30:43Z) - FastIF: Scalable Influence Functions for Efficient Model Interpretation
and Debugging [112.19994766375231]
Influence functions approximate the 'influences' of training data-points for test predictions.
We present FastIF, a set of simple modifications to influence functions that significantly improves their run-time.
Our experiments demonstrate the potential of influence functions in model interpretation and correcting model errors.
arXiv Detail & Related papers (2020-12-31T18:02:34Z) - Efficient Estimation of Influence of a Training Instance [56.29080605123304]
We propose an efficient method for estimating the influence of a training instance on a neural network model.
Our method is inspired by dropout, which zero-masks a sub-network and prevents the sub-network from learning each training instance.
We demonstrate that the proposed method can capture training influences, enhance the interpretability of error predictions, and cleanse the training dataset for improving generalization.
arXiv Detail & Related papers (2020-12-08T04:31:38Z) - Accurate and Robust Feature Importance Estimation under Distribution
Shifts [49.58991359544005]
PRoFILE is a novel feature importance estimation method.
We show significant improvements over state-of-the-art approaches, both in terms of fidelity and robustness.
arXiv Detail & Related papers (2020-09-30T05:29:01Z) - Explaining Black Box Predictions and Unveiling Data Artifacts through
Influence Functions [55.660255727031725]
Influence functions explain the decisions of a model by identifying influential training examples.
We conduct a comparison between influence functions and common word-saliency methods on representative tasks.
We develop a new measure based on influence functions that can reveal artifacts in training data.
arXiv Detail & Related papers (2020-05-14T00:45:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.