Model Learning with Personalized Interpretability Estimation (ML-PIE)
- URL: http://arxiv.org/abs/2104.06060v2
- Date: Wed, 14 Apr 2021 10:43:12 GMT
- Title: Model Learning with Personalized Interpretability Estimation (ML-PIE)
- Authors: Marco Virgolin, Andrea De Lorenzo, Francesca Randone, Eric Medvet,
Mattias Wahde
- Abstract summary: High-stakes applications require AI-generated models to be interpretable.
Current algorithms for the synthesis of potentially interpretable models rely on objectives or regularization terms.
We propose an approach for the synthesis of models that are tailored to the user.
- Score: 2.862606936691229
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-stakes applications require AI-generated models to be interpretable.
Current algorithms for the synthesis of potentially interpretable models rely
on objectives or regularization terms that represent interpretability only
coarsely (e.g., model size) and are not designed for a specific user. Yet,
interpretability is intrinsically subjective. In this paper, we propose an
approach for the synthesis of models that are tailored to the user by enabling
the user to steer the model synthesis process according to her or his
preferences. We use a bi-objective evolutionary algorithm to synthesize models
with trade-offs between accuracy and a user-specific notion of
interpretability. The latter is estimated by a neural network that is trained
concurrently to the evolution using the feedback of the user, which is
collected using uncertainty-based active learning. To maximize usability, the
user is only asked to tell, given two models at the time, which one is less
complex. With experiments on two real-world datasets involving 61 participants,
we find that our approach is capable of learning estimations of
interpretability that can be very different for different users. Moreover, the
users tend to prefer models found using the proposed approach over models found
using non-personalized interpretability indices.
Related papers
- HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Characterizing and overcoming the greedy nature of learning in
multi-modal deep neural networks [62.48782506095565]
We show that due to the greedy nature of learning in deep neural networks, models tend to rely on just one modality while under-fitting the other modalities.
We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning.
arXiv Detail & Related papers (2022-02-10T20:11:21Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - The Definitions of Interpretability and Learning of Interpretable Models [42.22982369082474]
We propose a mathematical definition for the human-interpretable model.
If a prediction model is interpretable by a human recognition system, the prediction model is defined as a completely human-interpretable model.
arXiv Detail & Related papers (2021-05-29T01:44:12Z) - Intuitively Assessing ML Model Reliability through Example-Based
Explanations and Editing Model Inputs [19.09848738521126]
Interpretability methods aim to help users build trust in and understand the capabilities of machine learning models.
We present two interface modules to facilitate a more intuitive assessment of model reliability.
arXiv Detail & Related papers (2021-02-17T02:41:32Z) - Evaluating the Interpretability of Generative Models by Interactive
Reconstruction [30.441247705313575]
We introduce a task to quantify the human-interpretability of generative model representations.
We find performance on this task much more reliably differentiates entangled and disentangled models than baseline approaches.
arXiv Detail & Related papers (2021-02-02T02:38:14Z) - Distilling Interpretable Models into Human-Readable Code [71.11328360614479]
Human-readability is an important and desirable standard for machine-learned model interpretability.
We propose to train interpretable models using conventional methods, and then distill them into concise, human-readable code.
We describe a piecewise-linear curve-fitting algorithm that produces high-quality results efficiently and reliably across a broad range of use cases.
arXiv Detail & Related papers (2021-01-21T01:46:36Z) - Human-interpretable model explainability on high-dimensional data [8.574682463936007]
We introduce a framework for human-interpretable explainability on high-dimensional data, consisting of two modules.
First, we apply a semantically meaningful latent representation, both to reduce the raw dimensionality of the data, and to ensure its human interpretability.
Second, we adapt the Shapley paradigm for model-agnostic explainability to operate on these latent features. This leads to interpretable model explanations that are both theoretically controlled and computationally tractable.
arXiv Detail & Related papers (2020-10-14T20:06:28Z) - Towards Open-World Recommendation: An Inductive Model-based
Collaborative Filtering Approach [115.76667128325361]
Recommendation models can effectively estimate underlying user interests and predict one's future behaviors.
We propose an inductive collaborative filtering framework that contains two representation models.
Our model achieves promising results for recommendation on few-shot users with limited training ratings and new unseen users.
arXiv Detail & Related papers (2020-07-09T14:31:25Z) - Learning a Formula of Interpretability to Learn Interpretable Formulas [1.7616042687330642]
We show that an ML model of non-objective Proxies of Human Interpretability can be learned from human feedback.
We show this for evolutionary symbolic regression.
Our approach represents an important stepping stone for the design of next-generation interpretable (evolutionary) ML algorithms.
arXiv Detail & Related papers (2020-04-23T13:59:49Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.