Balancing Explainability-Accuracy of Complex Models
- URL: http://arxiv.org/abs/2305.14098v1
- Date: Tue, 23 May 2023 14:20:38 GMT
- Title: Balancing Explainability-Accuracy of Complex Models
- Authors: Poushali Sengupta, Yan Zhang, Sabita Maharjan, Frank Eliassen
- Abstract summary: We introduce a new approach for complex models based on the co-relation impact.
We propose approaches for both scenarios of independent features and dependent features.
We provide an upper bound of the complexity of our proposed approach for the dependent features.
- Score: 8.402048778245165
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Explainability of AI models is an important topic that can have a significant
impact in all domains and applications from autonomous driving to healthcare.
The existing approaches to explainable AI (XAI) are mainly limited to simple
machine learning algorithms, and the research regarding the
explainability-accuracy tradeoff is still in its infancy especially when we are
concerned about complex machine learning techniques like neural networks and
deep learning (DL). In this work, we introduce a new approach for complex
models based on the co-relation impact which enhances the explainability
considerably while also ensuring the accuracy at a high level. We propose
approaches for both scenarios of independent features and dependent features.
In addition, we study the uncertainty associated with features and output.
Furthermore, we provide an upper bound of the computation complexity of our
proposed approach for the dependent features. The complexity bound depends on
the order of logarithmic of the number of observations which provides a
reliable result considering the higher dimension of dependent feature space
with a smaller number of observations.
Related papers
- Explainable Neural Networks with Guarantees: A Sparse Estimation Approach [11.142723510517778]
This paper introduces a novel approach to constructing an explainable neural network that harmonizes predictiveness and explainability.
Our model, termed SparXnet, is designed as a linear combination of a sparse set of jointly learned features.
Our research paves the way for further research on sparse and explainable neural networks with guarantee.
arXiv Detail & Related papers (2025-01-02T12:10:17Z) - Unified Explanations in Machine Learning Models: A Perturbation Approach [0.0]
Inconsistencies between XAI and modeling techniques can have the undesirable effect of casting doubt upon the efficacy of these explainability approaches.
We propose a systematic, perturbation-based analysis against a popular, model-agnostic method in XAI, SHapley Additive exPlanations (Shap)
We devise algorithms to generate relative feature importance in settings of dynamic inference amongst a suite of popular machine learning and deep learning methods, and metrics that allow us to quantify how well explanations generated under the static case hold.
arXiv Detail & Related papers (2024-05-30T16:04:35Z) - On the Benefits of Leveraging Structural Information in Planning Over
the Learned Model [3.3512508970931236]
We investigate the benefits of leveraging structural information about the system in terms of reducing sample complexity.
Our analysis shows that there can be a significant saving in sample complexity by leveraging structural information about the model.
arXiv Detail & Related papers (2023-03-15T18:18:01Z) - Interpretability with full complexity by constraining feature
information [1.52292571922932]
Interpretability is a pressing issue for machine learning.
We approach interpretability from a new angle: constrain the information about the features without restricting the complexity of the model.
We develop a framework for extracting insight from the spectrum of approximate models.
arXiv Detail & Related papers (2022-11-30T18:59:01Z) - Offline Reinforcement Learning with Differentiable Function
Approximation is Provably Efficient [65.08966446962845]
offline reinforcement learning, which aims at optimizing decision-making strategies with historical data, has been extensively applied in real-life applications.
We take a step by considering offline reinforcement learning with differentiable function class approximation (DFA)
Most importantly, we show offline differentiable function approximation is provably efficient by analyzing the pessimistic fitted Q-learning algorithm.
arXiv Detail & Related papers (2022-10-03T07:59:42Z) - Sample-Efficient Reinforcement Learning in the Presence of Exogenous
Information [77.19830787312743]
In real-world reinforcement learning applications the learner's observation space is ubiquitously high-dimensional with both relevant and irrelevant information about the task at hand.
We introduce a new problem setting for reinforcement learning, the Exogenous Decision Process (ExoMDP), in which the state space admits an (unknown) factorization into a small controllable component and a large irrelevant component.
We provide a new algorithm, ExoRL, which learns a near-optimal policy with sample complexity in the size of the endogenous component.
arXiv Detail & Related papers (2022-06-09T05:19:32Z) - Stabilizing Q-learning with Linear Architectures for Provably Efficient
Learning [53.17258888552998]
This work proposes an exploration variant of the basic $Q$-learning protocol with linear function approximation.
We show that the performance of the algorithm degrades very gracefully under a novel and more permissive notion of approximation error.
arXiv Detail & Related papers (2022-06-01T23:26:51Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Relational Neural Markov Random Fields [29.43155380361715]
We introduce Markov Random Fields (RN-MRFs) which allow handling of complex hybrid domains.
We propose a maximum pseudolikelihood estimation-based learning algorithm with importance for training the potential parameters.
arXiv Detail & Related papers (2021-10-18T22:52:54Z) - Counterfactual Explanations as Interventions in Latent Space [62.997667081978825]
Counterfactual explanations aim to provide to end users a set of features that need to be changed in order to achieve a desired outcome.
Current approaches rarely take into account the feasibility of actions needed to achieve the proposed explanations.
We present Counterfactual Explanations as Interventions in Latent Space (CEILS), a methodology to generate counterfactual explanations.
arXiv Detail & Related papers (2021-06-14T20:48:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.