SpecXAI -- Spectral interpretability of Deep Learning Models
- URL: http://arxiv.org/abs/2302.09949v1
- Date: Mon, 20 Feb 2023 12:36:54 GMT
- Title: SpecXAI -- Spectral interpretability of Deep Learning Models
- Authors: Stefan Druc, Peter Wooldridge, Adarsh Krishnamurthy, Soumik Sarkar,
Aditya Balu
- Abstract summary: XAI attempts to develop techniques that temper the impenetrable nature of the models and promote a level of understanding of their behavior.
Here we present our contribution to XAI methods in the form of a framework that we term SpecXAI.
We show how this framework can be used to not only understand the network but also manipulate it into a linear interpretable symbolic representation.
- Score: 11.325580593182414
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning is becoming increasingly adopted in business and industry due
to its ability to transform large quantities of data into high-performing
models. These models, however, are generally regarded as black boxes, which, in
spite of their performance, could prevent their use. In this context, the field
of eXplainable AI attempts to develop techniques that temper the impenetrable
nature of the models and promote a level of understanding of their behavior.
Here we present our contribution to XAI methods in the form of a framework that
we term SpecXAI, which is based on the spectral characterization of the entire
network. We show how this framework can be used to not only understand the
network but also manipulate it into a linear interpretable symbolic
representation.
Related papers
- Explainable Deep Learning Framework for Human Activity Recognition [3.9146761527401424]
We propose a model-agnostic framework that enhances interpretability and efficacy of HAR models.
By implementing competitive data augmentation, our framework provides intuitive and accessible explanations of model decisions.
arXiv Detail & Related papers (2024-08-21T11:59:55Z) - Towards Compositional Interpretability for XAI [3.3768167170511587]
We present an approach to defining AI models and their interpretability based on category theory.
We compare a wide range of AI models as compositional models.
We find that what makes the standard 'intrinsically interpretable' models so transparent is brought out most clearly diagrammatically.
arXiv Detail & Related papers (2024-06-25T14:27:03Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Greybox XAI: a Neural-Symbolic learning framework to produce
interpretable predictions for image classification [6.940242990198]
Greybox XAI is a framework that composes a DNN and a transparent model thanks to the use of a symbolic Knowledge Base (KB)
We address the problem of the lack of universal criteria for XAI by formalizing what an explanation is.
We show how this new architecture is accurate and explainable in several datasets.
arXiv Detail & Related papers (2022-09-26T08:55:31Z) - Beyond Explaining: Opportunities and Challenges of XAI-Based Model
Improvement [75.00655434905417]
Explainable Artificial Intelligence (XAI) is an emerging research field bringing transparency to highly complex machine learning (ML) models.
This paper offers a comprehensive overview over techniques that apply XAI practically for improving various properties of ML models.
We show empirically through experiments on toy and realistic settings how explanations can help improve properties such as model generalization ability or reasoning.
arXiv Detail & Related papers (2022-03-15T15:44:28Z) - Recurrence-Aware Long-Term Cognitive Network for Explainable Pattern
Classification [0.0]
We propose an LTCN-based model for interpretable pattern classification of structured data.
Our method brings its own mechanism for providing explanations by quantifying the relevance of each feature in the decision process.
Our interpretable model obtains competitive performance when compared to the state-of-the-art white and black boxes.
arXiv Detail & Related papers (2021-07-07T18:14:50Z) - EXplainable Neural-Symbolic Learning (X-NeSyL) methodology to fuse deep
learning representations with expert knowledge graphs: the MonuMAI cultural
heritage use case [13.833923272291853]
We present the eXplainable Neural-symbolic learning (X-NeSyL) methodology, designed to learn both symbolic and deep representations.
X-NeSyL methodology involves the concrete use of two notions of explanation at inference and training time respectively.
We showcase X-NeSyL methodology using MonuMAI dataset for monument facade image classification, and demonstrate that our approach improves explainability and performance.
arXiv Detail & Related papers (2021-04-24T09:06:08Z) - Generative Counterfactuals for Neural Networks via Attribute-Informed
Perturbation [51.29486247405601]
We design a framework to generate counterfactuals for raw data instances with the proposed Attribute-Informed Perturbation (AIP)
By utilizing generative models conditioned with different attributes, counterfactuals with desired labels can be obtained effectively and efficiently.
Experimental results on real-world texts and images demonstrate the effectiveness, sample quality as well as efficiency of our designed framework.
arXiv Detail & Related papers (2021-01-18T08:37:13Z) - Explainability in Deep Reinforcement Learning [68.8204255655161]
We review recent works in the direction to attain Explainable Reinforcement Learning (XRL)
In critical situations where it is essential to justify and explain the agent's behaviour, better explainability and interpretability of RL models could help gain scientific insight on the inner workings of what is still considered a black box.
arXiv Detail & Related papers (2020-08-15T10:11:42Z) - Explainable Matrix -- Visualization for Global and Local
Interpretability of Random Forest Classification Ensembles [78.6363825307044]
We propose Explainable Matrix (ExMatrix), a novel visualization method for Random Forest (RF) interpretability.
It employs a simple yet powerful matrix-like visual metaphor, where rows are rules, columns are features, and cells are rules predicates.
ExMatrix applicability is confirmed via different examples, showing how it can be used in practice to promote RF models interpretability.
arXiv Detail & Related papers (2020-05-08T21:03:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.