An Information-Theoretic Approach to Personalized Explainable Machine
Learning
- URL: http://arxiv.org/abs/2003.00484v2
- Date: Sun, 15 Mar 2020 14:38:49 GMT
- Title: An Information-Theoretic Approach to Personalized Explainable Machine
Learning
- Authors: Alexander Jung and Pedro H. J. Nardelli
- Abstract summary: We propose a simple probabilistic model for the predictions and user knowledge.
We quantify the effect of an explanation by the conditional mutual information between the explanation and prediction.
- Score: 92.53970625312665
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Automated decision making is used routinely throughout our everyday life.
Recommender systems decide which jobs, movies, or other user profiles might be
interesting to us. Spell checkers help us to make good use of language. Fraud
detection systems decide if a credit card transactions should be verified more
closely. Many of these decision making systems use machine learning methods
that fit complex models to massive datasets. The successful deployment of
machine learning (ML) methods to many (critical) application domains crucially
depends on its explainability. Indeed, humans have a strong desire to get
explanations that resolve the uncertainty about experienced phenomena like the
predictions and decisions obtained from ML methods. Explainable ML is
challenging since explanations must be tailored (personalized) to individual
users with varying backgrounds. Some users might have received university-level
education in ML, while other users might have no formal training in linear
algebra. Linear regression with few features might be perfectly interpretable
for the first group but might be considered a black-box by the latter. We
propose a simple probabilistic model for the predictions and user knowledge.
This model allows to study explainable ML using information theory. Explaining
is here considered as the task of reducing the "surprise" incurred by a
prediction. We quantify the effect of an explanation by the conditional mutual
information between the explanation and prediction, given the user background.
Related papers
- Pyreal: A Framework for Interpretable ML Explanations [51.14710806705126]
Pyreal is a system for generating a variety of interpretable machine learning explanations.
Pyreal converts data and explanations between the feature spaces expected by the model, relevant explanation algorithms, and human users.
Our studies demonstrate that Pyreal generates more useful explanations than existing systems.
arXiv Detail & Related papers (2023-12-20T15:04:52Z) - Democratizing Reasoning Ability: Tailored Learning from Large Language
Model [97.4921006089966]
We propose a tailored learning approach to distill such reasoning ability to smaller LMs.
We exploit the potential of LLM as a reasoning teacher by building an interactive multi-round learning paradigm.
To exploit the reasoning potential of the smaller LM, we propose self-reflection learning to motivate the student to learn from self-made mistakes.
arXiv Detail & Related papers (2023-10-20T07:50:10Z) - Ticketed Learning-Unlearning Schemes [57.89421552780526]
We propose a new ticketed model for learning--unlearning.
We provide space-efficient ticketed learning--unlearning schemes for a broad family of concept classes.
arXiv Detail & Related papers (2023-06-27T18:54:40Z) - Reason to explain: Interactive contrastive explanations (REASONX) [5.156484100374058]
We present REASONX, an explanation tool based on Constraint Logic Programming (CLP)
REASONX provides interactive contrastive explanations that can be augmented by background knowledge.
It computes factual and constrative decision rules, as well as closest constrative examples.
arXiv Detail & Related papers (2023-05-29T15:13:46Z) - Learning to Scaffold: Optimizing Model Explanations for Teaching [74.25464914078826]
We train models on three natural language processing and computer vision tasks.
We find that students trained with explanations extracted with our framework are able to simulate the teacher significantly more effectively than ones produced with previous methods.
arXiv Detail & Related papers (2022-04-22T16:43:39Z) - Supervised Machine Learning with Plausible Deniability [1.685485565763117]
We study the question of how well machine learning (ML) models trained on a certain data set provide privacy for the training data.
We show that one can take a set of purely random training data, and from this define a suitable learning rule'' that will produce a ML model that is exactly $f$.
arXiv Detail & Related papers (2021-06-08T11:54:51Z) - On Interpretability and Similarity in Concept-Based Machine Learning [2.3986080077861787]
We discuss how notions from cooperative game theory can be used to assess the contribution of individual attributes in classification and clustering processes in concept-based machine learning.
To address the 3rd question, we present some ideas on how to reduce the number of attributes using similarities in large contexts.
arXiv Detail & Related papers (2021-02-25T07:57:28Z) - Teaching the Machine to Explain Itself using Domain Knowledge [4.462334751640166]
Non-technical humans-in-the-loop struggle to comprehend the rationale behind model predictions.
We present JOEL, a neural network-based framework to jointly learn a decision-making task and associated explanations.
We collect the domain feedback from a pool of certified experts and use it to ameliorate the model (human teaching)
arXiv Detail & Related papers (2020-11-27T18:46:34Z) - Explainable Empirical Risk Minimization [0.6299766708197883]
Successful application of machine learning (ML) methods becomes increasingly dependent on their interpretability or explainability.
This paper applies information-theoretic concepts to develop a novel measure for the subjective explainability of predictions delivered by a ML method.
Our main contribution is the explainable empirical risk minimization (EERM) principle of learning a hypothesis that optimally balances between the subjective explainability and risk.
arXiv Detail & Related papers (2020-09-03T07:16:34Z) - The Information Bottleneck Problem and Its Applications in Machine
Learning [53.57797720793437]
Inference capabilities of machine learning systems skyrocketed in recent years, now playing a pivotal role in various aspect of society.
The information bottleneck (IB) theory emerged as a bold information-theoretic paradigm for analyzing deep learning (DL) systems.
In this tutorial we survey the information-theoretic origins of this abstract principle, and its recent impact on DL.
arXiv Detail & Related papers (2020-04-30T16:48:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.