Learning a Formula of Interpretability to Learn Interpretable Formulas
- URL: http://arxiv.org/abs/2004.11170v2
- Date: Thu, 28 May 2020 15:08:37 GMT
- Title: Learning a Formula of Interpretability to Learn Interpretable Formulas
- Authors: Marco Virgolin, Andrea De Lorenzo, Eric Medvet, and Francesca Randone
- Abstract summary: We show that an ML model of non-objective Proxies of Human Interpretability can be learned from human feedback.
We show this for evolutionary symbolic regression.
Our approach represents an important stepping stone for the design of next-generation interpretable (evolutionary) ML algorithms.
- Score: 1.7616042687330642
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many risk-sensitive applications require Machine Learning (ML) models to be
interpretable. Attempts to obtain interpretable models typically rely on
tuning, by trial-and-error, hyper-parameters of model complexity that are only
loosely related to interpretability. We show that it is instead possible to
take a meta-learning approach: an ML model of non-trivial Proxies of Human
Interpretability (PHIs) can be learned from human feedback, then this model can
be incorporated within an ML training process to directly optimize for
interpretability. We show this for evolutionary symbolic regression. We first
design and distribute a survey finalized at finding a link between features of
mathematical formulas and two established PHIs, simulatability and
decomposability. Next, we use the resulting dataset to learn an ML model of
interpretability. Lastly, we query this model to estimate the interpretability
of evolving solutions within bi-objective genetic programming. We perform
experiments on five synthetic and eight real-world symbolic regression
problems, comparing to the traditional use of solution size minimization. The
results show that the use of our model leads to formulas that are, for a same
level of accuracy-interpretability trade-off, either significantly more or
equally accurate. Moreover, the formulas are also arguably more interpretable.
Given the very positive results, we believe that our approach represents an
important stepping stone for the design of next-generation interpretable
(evolutionary) ML algorithms.
Related papers
- On Discriminative Probabilistic Modeling for Self-Supervised Representation Learning [85.75164588939185]
We study the discriminative probabilistic modeling problem on a continuous domain for (multimodal) self-supervised representation learning.
We conduct generalization error analysis to reveal the limitation of current InfoNCE-based contrastive loss for self-supervised representation learning.
arXiv Detail & Related papers (2024-10-11T18:02:46Z) - Are Linear Regression Models White Box and Interpretable? [0.0]
Explainable artificial intelligence (XAI) is a set of tools and algorithms that applied or embedded to machine learning models to understand and interpret the models.
Simple models including linear regression are easy to implement, has less computational complexity and easy to visualize the output.
arXiv Detail & Related papers (2024-07-16T21:05:51Z) - Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models [115.501751261878]
Fine-tuning language models(LMs) on human-generated data remains a prevalent practice.
We investigate whether we can go beyond human data on tasks where we have access to scalar feedback.
We find that ReST$EM$ scales favorably with model size and significantly surpasses fine-tuning only on human data.
arXiv Detail & Related papers (2023-12-11T18:17:43Z) - The Languini Kitchen: Enabling Language Modelling Research at Different
Scales of Compute [66.84421705029624]
We introduce an experimental protocol that enables model comparisons based on equivalent compute, measured in accelerator hours.
We pre-process an existing large, diverse, and high-quality dataset of books that surpasses existing academic benchmarks in quality, diversity, and document length.
This work also provides two baseline models: a feed-forward model derived from the GPT-2 architecture and a recurrent model in the form of a novel LSTM with ten-fold throughput.
arXiv Detail & Related papers (2023-09-20T10:31:17Z) - Discovering interpretable elastoplasticity models via the neural
polynomial method enabled symbolic regressions [0.0]
Conventional neural network elastoplasticity models are often perceived as lacking interpretability.
This paper introduces a two-step machine learning approach that returns mathematical models interpretable by human experts.
arXiv Detail & Related papers (2023-07-24T22:22:32Z) - Theoretical Characterization of the Generalization Performance of
Overfitted Meta-Learning [70.52689048213398]
This paper studies the performance of overfitted meta-learning under a linear regression model with Gaussian features.
We find new and interesting properties that do not exist in single-task linear regression.
Our analysis suggests that benign overfitting is more significant and easier to observe when the noise and the diversity/fluctuation of the ground truth of each training task are large.
arXiv Detail & Related papers (2023-04-09T20:36:13Z) - Automated Learning of Interpretable Models with Quantified Uncertainty [0.0]
We introduce a new framework for genetic-programming-based symbolic regression (GPSR)
GPSR uses model evidence to formulate replacement probability during the selection phase of evolution.
It is shown to increase interpretability, improve robustness to noise, and reduce overfitting when compared to a conventional GPSR implementation.
arXiv Detail & Related papers (2022-04-12T19:56:42Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - The Definitions of Interpretability and Learning of Interpretable Models [42.22982369082474]
We propose a mathematical definition for the human-interpretable model.
If a prediction model is interpretable by a human recognition system, the prediction model is defined as a completely human-interpretable model.
arXiv Detail & Related papers (2021-05-29T01:44:12Z) - Model Learning with Personalized Interpretability Estimation (ML-PIE) [2.862606936691229]
High-stakes applications require AI-generated models to be interpretable.
Current algorithms for the synthesis of potentially interpretable models rely on objectives or regularization terms.
We propose an approach for the synthesis of models that are tailored to the user.
arXiv Detail & Related papers (2021-04-13T09:47:48Z) - Distilling Interpretable Models into Human-Readable Code [71.11328360614479]
Human-readability is an important and desirable standard for machine-learned model interpretability.
We propose to train interpretable models using conventional methods, and then distill them into concise, human-readable code.
We describe a piecewise-linear curve-fitting algorithm that produces high-quality results efficiently and reliably across a broad range of use cases.
arXiv Detail & Related papers (2021-01-21T01:46:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.