RelatIF: Identifying Explanatory Training Examples via Relative
Influence
- URL: http://arxiv.org/abs/2003.11630v1
- Date: Wed, 25 Mar 2020 20:59:54 GMT
- Title: RelatIF: Identifying Explanatory Training Examples via Relative
Influence
- Authors: Elnaz Barshan, Marc-Etienne Brunet, Gintare Karolina Dziugaite
- Abstract summary: We use influence functions to identify relevant training examples that one might hope "explain" the predictions of a machine learning model.
We introduce RelatIF, a new class of criteria for choosing relevant training examples by way of an optimization objective that places a constraint on global influence.
In empirical evaluations, we find that the examples returned by RelatIF are more intuitive when compared to those found using influence functions.
- Score: 13.87851325824883
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we focus on the use of influence functions to identify relevant
training examples that one might hope "explain" the predictions of a machine
learning model. One shortcoming of influence functions is that the training
examples deemed most "influential" are often outliers or mislabelled, making
them poor choices for explanation. In order to address this shortcoming, we
separate the role of global versus local influence. We introduce RelatIF, a new
class of criteria for choosing relevant training examples by way of an
optimization objective that places a constraint on global influence. RelatIF
considers the local influence that an explanatory example has on a prediction
relative to its global effects on the model. In empirical evaluations, we find
that the examples returned by RelatIF are more intuitive when compared to those
found using influence functions.
Related papers
- Influence Functions for Scalable Data Attribution in Diffusion Models [52.92223039302037]
Diffusion models have led to significant advancements in generative modelling.
Yet their widespread adoption poses challenges regarding data attribution and interpretability.
In this paper, we aim to help address such challenges by developing an textitinfluence functions framework.
arXiv Detail & Related papers (2024-10-17T17:59:02Z) - Do Influence Functions Work on Large Language Models? [10.463762448166714]
Influence functions aim to quantify the impact of individual training data points on a model's predictions.
We evaluate influence functions across multiple tasks and find that they consistently perform poorly in most settings.
arXiv Detail & Related papers (2024-09-30T06:50:18Z) - Most Influential Subset Selection: Challenges, Promises, and Beyond [9.479235005673683]
We study the Most Influential Subset Selection (MISS) problem, which aims to identify a subset of training samples with the greatest collective influence.
We conduct a comprehensive analysis of the prevailing approaches in MISS, elucidating their strengths and weaknesses.
We demonstrate that an adaptive version of theses which applies them iteratively, can effectively capture the interactions among samples.
arXiv Detail & Related papers (2024-09-25T20:00:23Z) - If Influence Functions are the Answer, Then What is the Question? [7.873458431535409]
Influence functions efficiently estimate the effect of removing a single training data point on a model's learned parameters.
While influence estimates align well with leave-one-out retraining for linear models, recent works have shown this alignment is often poor in neural networks.
arXiv Detail & Related papers (2022-09-12T16:17:43Z) - An Additive Instance-Wise Approach to Multi-class Model Interpretation [53.87578024052922]
Interpretable machine learning offers insights into what factors drive a certain prediction of a black-box system.
Existing methods mainly focus on selecting explanatory input features, which follow either locally additive or instance-wise approaches.
This work exploits the strengths of both methods and proposes a global framework for learning local explanations simultaneously for multiple target classes.
arXiv Detail & Related papers (2022-07-07T06:50:27Z) - Understanding Instance-Level Impact of Fairness Constraints [12.866655972682254]
We study the influence of training examples when fairness constraints are imposed.
We find that training on a subset of weighty data examples leads to lower fairness violations with a trade-off of accuracy.
arXiv Detail & Related papers (2022-06-30T17:31:33Z) - Revisiting Methods for Finding Influential Examples [2.094022863940315]
Methods for finding influential training examples for test-time decisions have been proposed.
In this paper, we show that all of the above methods are unstable.
We propose to evaluate such explanations by their ability to detect poisoning attacks.
arXiv Detail & Related papers (2021-11-08T18:00:06Z) - Efficient Estimation of Influence of a Training Instance [56.29080605123304]
We propose an efficient method for estimating the influence of a training instance on a neural network model.
Our method is inspired by dropout, which zero-masks a sub-network and prevents the sub-network from learning each training instance.
We demonstrate that the proposed method can capture training influences, enhance the interpretability of error predictions, and cleanse the training dataset for improving generalization.
arXiv Detail & Related papers (2020-12-08T04:31:38Z) - Understanding Adversarial Examples from the Mutual Influence of Images
and Perturbations [83.60161052867534]
We analyze adversarial examples by disentangling the clean images and adversarial perturbations, and analyze their influence on each other.
Our results suggest a new perspective towards the relationship between images and universal perturbations.
We are the first to achieve the challenging task of a targeted universal attack without utilizing original training data.
arXiv Detail & Related papers (2020-07-13T05:00:09Z) - Influence Functions in Deep Learning Are Fragile [52.31375893260445]
influence functions approximate the effect of samples in test-time predictions.
influence estimates are fairly accurate for shallow networks.
Hessian regularization is important to get highquality influence estimates.
arXiv Detail & Related papers (2020-06-25T18:25:59Z) - Explaining Black Box Predictions and Unveiling Data Artifacts through
Influence Functions [55.660255727031725]
Influence functions explain the decisions of a model by identifying influential training examples.
We conduct a comparison between influence functions and common word-saliency methods on representative tasks.
We develop a new measure based on influence functions that can reveal artifacts in training data.
arXiv Detail & Related papers (2020-05-14T00:45:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.