SAM: The Sensitivity of Attribution Methods to Hyperparameters
- URL: http://arxiv.org/abs/2003.08754v2
- Date: Mon, 13 Apr 2020 03:32:08 GMT
- Title: SAM: The Sensitivity of Attribution Methods to Hyperparameters
- Authors: Naman Bansal, Chirag Agarwal, Anh Nguyen
- Abstract summary: We argue that a key desideratum of an explanation method is its robustness to input hyperparameters which are often randomly set or empirically tuned.
In this paper, we provide a thorough empirical study on the sensitivity of existing attribution methods.
- Score: 13.145335512841557
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Attribution methods can provide powerful insights into the reasons for a
classifier's decision. We argue that a key desideratum of an explanation method
is its robustness to input hyperparameters which are often randomly set or
empirically tuned. High sensitivity to arbitrary hyperparameter choices does
not only impede reproducibility but also questions the correctness of an
explanation and impairs the trust of end-users. In this paper, we provide a
thorough empirical study on the sensitivity of existing attribution methods. We
found an alarming trend that many methods are highly sensitive to changes in
their common hyperparameters e.g. even changing a random seed can yield a
different explanation! Interestingly, such sensitivity is not reflected in the
average explanation accuracy scores over the dataset as commonly reported in
the literature. In addition, explanations generated for robust classifiers
(i.e. which are trained to be invariant to pixel-wise perturbations) are
surprisingly more robust than those generated for regular classifiers.
Related papers
- Classification under Nuisance Parameters and Generalized Label Shift in Likelihood-Free Inference [3.507509142413452]
We propose a new method for robust uncertainty quantification that casts classification as a hypothesis testing problem under nuisance parameters.
Our method effectively endows a pre-trained classifier with domain adaptation capabilities and returns valid prediction sets while maintaining high power.
We demonstrate its performance on two challenging scientific problems in biology and astroparticle physics with data from realistic mechanistic models.
arXiv Detail & Related papers (2024-02-08T00:12:18Z) - Winning Prize Comes from Losing Tickets: Improve Invariant Learning by
Exploring Variant Parameters for Out-of-Distribution Generalization [76.27711056914168]
Out-of-Distribution (OOD) Generalization aims to learn robust models that generalize well to various environments without fitting to distribution-specific features.
Recent studies based on Lottery Ticket Hypothesis (LTH) address this problem by minimizing the learning target to find some of the parameters that are critical to the task.
We propose Exploring Variant parameters for Invariant Learning (EVIL) which also leverages the distribution knowledge to find the parameters that are sensitive to distribution shift.
arXiv Detail & Related papers (2023-10-25T06:10:57Z) - The Lipschitz-Variance-Margin Tradeoff for Enhanced Randomized Smoothing [85.85160896547698]
Real-life applications of deep neural networks are hindered by their unsteady predictions when faced with noisy inputs and adversarial attacks.
We show how to design an efficient classifier with a certified radius by relying on noise injection into the inputs.
Our novel certification procedure allows us to use pre-trained models with randomized smoothing, effectively improving the current certification radius in a zero-shot manner.
arXiv Detail & Related papers (2023-09-28T22:41:47Z) - How adversarial attacks can disrupt seemingly stable accurate classifiers [76.95145661711514]
Adversarial attacks dramatically change the output of an otherwise accurate learning system using a seemingly inconsequential modification to a piece of input data.
Here, we show that this may be seen as a fundamental feature of classifiers working with high dimensional input data.
We introduce a simple generic and generalisable framework for which key behaviours observed in practical systems arise with high probability.
arXiv Detail & Related papers (2023-09-07T12:02:00Z) - Confidence-aware Training of Smoothed Classifiers for Certified
Robustness [75.95332266383417]
We use "accuracy under Gaussian noise" as an easy-to-compute proxy of adversarial robustness for an input.
Our experiments show that the proposed method consistently exhibits improved certified robustness upon state-of-the-art training methods.
arXiv Detail & Related papers (2022-12-18T03:57:12Z) - Shortcomings of Top-Down Randomization-Based Sanity Checks for
Evaluations of Deep Neural Network Explanations [67.40641255908443]
We identify limitations of model-randomization-based sanity checks for the purpose of evaluating explanations.
Top-down model randomization preserves scales of forward pass activations with high probability.
arXiv Detail & Related papers (2022-11-22T18:52:38Z) - Parameter Averaging for Robust Explainability [5.032906900691182]
We introduce a novel method based on parameter averaging for robust explainability in tabular data setting, referred as XTab.
We conduct extensive experiments on a variety of real and synthetic datasets, demonstrating that the proposed method can be used for feature selection as well as to obtain the global feature importance that are not sensitive to sub-optimal model initialisation.
arXiv Detail & Related papers (2022-08-05T16:00:07Z) - More Than Words: Towards Better Quality Interpretations of Text
Classifiers [16.66535643383862]
We show that token-based interpretability, while being a convenient first choice given the input interfaces of the ML models, is not the most effective one in all situations.
We show that higher-level feature attributions offer several advantages: 1) they are more robust as measured by the randomization tests, 2) they lead to lower variability when using approximation-based methods like SHAP, and 3) they are more intelligible to humans in situations where the linguistic coherence resides at a higher level.
arXiv Detail & Related papers (2021-12-23T10:18:50Z) - Deterministic Certification to Adversarial Attacks via Bernstein
Polynomial Approximation [5.392822954974537]
Randomized smoothing has established state-of-the-art provable robustness against $ell$ norm adversarial attacks with high probability.
We come up with a question, "Is it possible to construct a smoothed classifier without randomization while maintaining natural accuracy?"
Our method provides a deterministic algorithm for decision boundary smoothing.
We also introduce a distinctive approach of norm-independent certified robustness via numerical solutions of nonlinear systems of equations.
arXiv Detail & Related papers (2020-11-28T08:27:42Z) - Consistency Regularization for Certified Robustness of Smoothed
Classifiers [89.72878906950208]
A recent technique of randomized smoothing has shown that the worst-case $ell$-robustness can be transformed into the average-case robustness.
We found that the trade-off between accuracy and certified robustness of smoothed classifiers can be greatly controlled by simply regularizing the prediction consistency over noise.
arXiv Detail & Related papers (2020-06-07T06:57:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.