GLocalX -- From Local to Global Explanations of Black Box AI Models
- URL: http://arxiv.org/abs/2101.07685v2
- Date: Tue, 26 Jan 2021 11:26:16 GMT
- Title: GLocalX -- From Local to Global Explanations of Black Box AI Models
- Authors: Mattia Setzu, Riccardo Guidotti, Anna Monreale, Franco Turini, Dino
Pedreschi, Fosca Giannotti
- Abstract summary: We present GLocalX, a "local-first" model agnostic explanation method.
Our goal is to learn accurate yet simple interpretable models to emulate the given black box, and, if possible, replace it entirely.
- Score: 12.065358125757847
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Artificial Intelligence (AI) has come to prominence as one of the major
components of our society, with applications in most aspects of our lives. In
this field, complex and highly nonlinear machine learning models such as
ensemble models, deep neural networks, and Support Vector Machines have
consistently shown remarkable accuracy in solving complex tasks. Although
accurate, AI models often are "black boxes" which we are not able to
understand. Relying on these models has a multifaceted impact and raises
significant concerns about their transparency. Applications in sensitive and
critical domains are a strong motivational factor in trying to understand the
behavior of black boxes. We propose to address this issue by providing an
interpretable layer on top of black box models by aggregating "local"
explanations. We present GLocalX, a "local-first" model agnostic explanation
method. Starting from local explanations expressed in form of local decision
rules, GLocalX iteratively generalizes them into global explanations by
hierarchically aggregating them. Our goal is to learn accurate yet simple
interpretable models to emulate the given black box, and, if possible, replace
it entirely. We validate GLocalX in a set of experiments in standard and
constrained settings with limited or no access to either data or local
explanations. Experiments show that GLocalX is able to accurately emulate
several models with simple and small models, reaching state-of-the-art
performance against natively global solutions. Our findings show how it is
often possible to achieve a high level of both accuracy and comprehensibility
of classification models, even in complex domains with high-dimensional data,
without necessarily trading one property for the other. This is a key
requirement for a trustworthy AI, necessary for adoption in high-stakes
decision making applications.
Related papers
- Explainable AI for Comparative Analysis of Intrusion Detection Models [20.683181384051395]
This research analyzes various machine learning models to the tasks of binary and multi-class classification for intrusion detection from network traffic.
We trained all models to the accuracy of 90% on the UNSW-NB15 dataset.
We also discover that Random Forest provides the best performance in terms of accuracy, time efficiency and robustness.
arXiv Detail & Related papers (2024-06-14T03:11:01Z) - Towards a Generalist and Blind RGB-X Tracker [91.36268768952755]
We develop a single model tracker that can remain blind to any modality X during inference time.
Our training process is extremely simple, integrating multi-label classification loss with a routing function.
Our generalist and blind tracker can achieve competitive performance compared to well-established modal-specific models.
arXiv Detail & Related papers (2024-05-28T03:00:58Z) - Interpretability at Scale: Identifying Causal Mechanisms in Alpaca [62.65877150123775]
We use Boundless DAS to efficiently search for interpretable causal structure in large language models while they follow instructions.
Our findings mark a first step toward faithfully understanding the inner-workings of our ever-growing and most widely deployed language models.
arXiv Detail & Related papers (2023-05-15T17:15:40Z) - When Neural Networks Fail to Generalize? A Model Sensitivity Perspective [82.36758565781153]
Domain generalization (DG) aims to train a model to perform well in unseen domains under different distributions.
This paper considers a more realistic yet more challenging scenario, namely Single Domain Generalization (Single-DG)
We empirically ascertain a property of a model that correlates strongly with its generalization that we coin as "model sensitivity"
We propose a novel strategy of Spectral Adversarial Data Augmentation (SADA) to generate augmented images targeted at the highly sensitive frequencies.
arXiv Detail & Related papers (2022-12-01T20:15:15Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - Greybox XAI: a Neural-Symbolic learning framework to produce
interpretable predictions for image classification [6.940242990198]
Greybox XAI is a framework that composes a DNN and a transparent model thanks to the use of a symbolic Knowledge Base (KB)
We address the problem of the lack of universal criteria for XAI by formalizing what an explanation is.
We show how this new architecture is accurate and explainable in several datasets.
arXiv Detail & Related papers (2022-09-26T08:55:31Z) - Learning outside the Black-Box: The pursuit of interpretable models [78.32475359554395]
This paper proposes an algorithm that produces a continuous global interpretation of any given continuous black-box function.
Our interpretation represents a leap forward from the previous state of the art.
arXiv Detail & Related papers (2020-11-17T12:39:44Z) - Explainable Deep Classification Models for Domain Generalization [94.43131722655617]
Explanations are defined as regions of visual evidence upon which a deep classification network makes a decision.
Our training strategy enforces a periodic saliency-based feedback to encourage the model to focus on the image regions that directly correspond to the ground-truth object.
arXiv Detail & Related papers (2020-03-13T22:22:15Z) - Learning Global Transparent Models Consistent with Local Contrastive
Explanations [34.86847988157447]
We create custom features from sparse local contrastive explanations of the black-box model and then train a globally transparent model on just these.
Based on a key insight we propose a novel method where we create custom features from sparse local contrastive explanations of the black-box model and then train a globally transparent model on just these.
arXiv Detail & Related papers (2020-02-19T15:45:42Z) - MapLUR: Exploring a new Paradigm for Estimating Air Pollution using Deep
Learning on Map Images [4.7791671364702575]
Land-use regression models are important for the assessment of air pollution concentrations in areas without measurement stations.
We propose the Data-driven, Open, Global (DOG) paradigm that entails models based on purely data-driven approaches using only openly and globally available data.
arXiv Detail & Related papers (2020-02-18T11:21:55Z) - Auditing and Debugging Deep Learning Models via Decision Boundaries:
Individual-level and Group-level Analysis [0.0]
We use flip points to explain, audit, and debug deep learning models.
A flip point is any point that lies on the boundary between two output classes.
We demonstrate our methods by investigating several models trained on standard datasets used in social applications of machine learning.
arXiv Detail & Related papers (2020-01-03T01:45:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.