Consistent Explanations in the Face of Model Indeterminacy via
Ensembling
- URL: http://arxiv.org/abs/2306.06193v2
- Date: Tue, 13 Jun 2023 02:04:21 GMT
- Title: Consistent Explanations in the Face of Model Indeterminacy via
Ensembling
- Authors: Dan Ley, Leonard Tang, Matthew Nazari, Hongjin Lin, Suraj Srinivas,
Himabindu Lakkaraju
- Abstract summary: This work addresses the challenge of providing consistent explanations for predictive models in the presence of model indeterminacy.
We introduce ensemble methods to enhance the consistency of the explanations provided in these scenarios.
Our findings highlight the importance of considering model indeterminacy when interpreting explanations.
- Score: 12.661530681518899
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This work addresses the challenge of providing consistent explanations for
predictive models in the presence of model indeterminacy, which arises due to
the existence of multiple (nearly) equally well-performing models for a given
dataset and task. Despite their similar performance, such models often exhibit
inconsistent or even contradictory explanations for their predictions, posing
challenges to end users who rely on these models to make critical decisions.
Recognizing this issue, we introduce ensemble methods as an approach to enhance
the consistency of the explanations provided in these scenarios. Leveraging
insights from recent work on neural network loss landscapes and mode
connectivity, we devise ensemble strategies to efficiently explore the
underspecification set -- the set of models with performance variations
resulting solely from changes in the random seed during training. Experiments
on five benchmark financial datasets reveal that ensembling can yield
significant improvements when it comes to explanation similarity, and
demonstrate the potential of existing ensemble methods to explore the
underspecification set efficiently. Our findings highlight the importance of
considering model indeterminacy when interpreting explanations and showcase the
effectiveness of ensembles in enhancing the reliability of explanations in
machine learning.
Related papers
- Pattern based learning and optimisation through pricing for bin packing problem [50.83768979636913]
We argue that when problem conditions such as the distributions of random variables change, the patterns that performed well in previous circumstances may become less effective.
We propose a novel scheme to efficiently identify patterns and dynamically quantify their values for each specific condition.
Our method quantifies the value of patterns based on their ability to satisfy constraints and their effects on the objective value.
arXiv Detail & Related papers (2024-08-27T17:03:48Z) - Explanatory Model Monitoring to Understand the Effects of Feature Shifts on Performance [61.06245197347139]
We propose a novel approach to explain the behavior of a black-box model under feature shifts.
We refer to our method that combines concepts from Optimal Transport and Shapley Values as Explanatory Performance Estimation.
arXiv Detail & Related papers (2024-08-24T18:28:19Z) - Towards Robust and Adaptive Motion Forecasting: A Causal Representation
Perspective [72.55093886515824]
We introduce a causal formalism of motion forecasting, which casts the problem as a dynamic process with three groups of latent variables.
We devise a modular architecture that factorizes the representations of invariant mechanisms and style confounders to approximate a causal graph.
Experiment results on synthetic and real datasets show that our three proposed components significantly improve the robustness and reusability of the learned motion representations.
arXiv Detail & Related papers (2021-11-29T18:59:09Z) - Discriminative Multimodal Learning via Conditional Priors in Generative
Models [21.166519800652047]
This research studies the realistic scenario in which all modalities and class labels are available for model training.
We show, in this scenario, that the variational lower bound limits mutual information between joint representations and missing modalities.
arXiv Detail & Related papers (2021-10-09T17:22:24Z) - Beyond Trivial Counterfactual Explanations with Diverse Valuable
Explanations [64.85696493596821]
In computer vision applications, generative counterfactual methods indicate how to perturb a model's input to change its prediction.
We propose a counterfactual method that learns a perturbation in a disentangled latent space that is constrained using a diversity-enforcing loss.
Our model improves the success rate of producing high-quality valuable explanations when compared to previous state-of-the-art methods.
arXiv Detail & Related papers (2021-03-18T12:57:34Z) - Characterizing Fairness Over the Set of Good Models Under Selective
Labels [69.64662540443162]
We develop a framework for characterizing predictive fairness properties over the set of models that deliver similar overall performance.
We provide tractable algorithms to compute the range of attainable group-level predictive disparities.
We extend our framework to address the empirically relevant challenge of selectively labelled data.
arXiv Detail & Related papers (2021-01-02T02:11:37Z) - Accounting for Unobserved Confounding in Domain Generalization [107.0464488046289]
This paper investigates the problem of learning robust, generalizable prediction models from a combination of datasets.
Part of the challenge of learning robust models lies in the influence of unobserved confounders.
We demonstrate the empirical performance of our approach on healthcare data from different modalities.
arXiv Detail & Related papers (2020-07-21T08:18:06Z) - Capsule Networks -- A Probabilistic Perspective [42.187785678596384]
'Capsule' models try to explicitly represent the poses of objects, enforcing a linear relationship between an object's pose and that of its constituent parts.
We describe a probabilistic generative model which encodes such capsule assumptions, clearly separating the generative parts of the model from the inference mechanisms.
We experimentally demonstrate the applicability of our unified objective, and demonstrate the use of test time optimisation to solve problems inherent to amortised inference in our model.
arXiv Detail & Related papers (2020-04-07T17:26:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.