Breiman's two cultures: You don't have to choose sides
- URL: http://arxiv.org/abs/2104.12219v1
- Date: Sun, 25 Apr 2021 17:58:46 GMT
- Title: Breiman's two cultures: You don't have to choose sides
- Authors: Andrew C. Miller, Nicholas J. Foti, Emily B. Fox
- Abstract summary: Breiman's classic paper casts data analysis as a choice between two cultures.
Data modelers use simple, interpretable models with well-understood theoretical properties to analyze data.
Algorithm modelers prioritize predictive accuracy and use more flexible function approximations to analyze data.
- Score: 10.695407438192527
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Breiman's classic paper casts data analysis as a choice between two cultures:
data modelers and algorithmic modelers. Stated broadly, data modelers use
simple, interpretable models with well-understood theoretical properties to
analyze data. Algorithmic modelers prioritize predictive accuracy and use more
flexible function approximations to analyze data. This dichotomy overlooks a
third set of models $-$ mechanistic models derived from scientific theories
(e.g., ODE/SDE simulators). Mechanistic models encode application-specific
scientific knowledge about the data. And while these categories represent
extreme points in model space, modern computational and algorithmic tools
enable us to interpolate between these points, producing flexible,
interpretable, and scientifically-informed hybrids that can enjoy accurate and
robust predictions, and resolve issues with data analysis that Breiman
describes, such as the Rashomon effect and Occam's dilemma. Challenges still
remain in finding an appropriate point in model space, with many choices on how
to compose model components and the degree to which each component informs
inferences.
Related papers
- Interpretable and Scalable Graphical Models for Complex Spatio-temporal
Processes [3.469001874498102]
thesis focuses on data that has complex-temporal structure and on probabilistic graphical models that learn the structure in an interpretable and interpretable manner.
practical applications of the methodology are considered using real datasets.
This includes brain-connectivity analysis using data, space weather forecasting using solar imaging data, longitudinal analysis of public opinions using Twitter data, and mining of mental health related issues using TalkLife data.
arXiv Detail & Related papers (2023-01-15T05:39:30Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - Design of Dynamic Experiments for Black-Box Model Discrimination [72.2414939419588]
Consider a dynamic model discrimination setting where we wish to chose: (i) what is the best mechanistic, time-varying model and (ii) what are the best model parameter estimates.
For rival mechanistic models where we have access to gradient information, we extend existing methods to incorporate a wider range of problem uncertainty.
We replace these black-box models with Gaussian process surrogate models and thereby extend the model discrimination setting to additionally incorporate rival black-box model.
arXiv Detail & Related papers (2021-02-07T11:34:39Z) - Understanding Neural Abstractive Summarization Models via Uncertainty [54.37665950633147]
seq2seq abstractive summarization models generate text in a free-form manner.
We study the entropy, or uncertainty, of the model's token-level predictions.
We show that uncertainty is a useful perspective for analyzing summarization and text generation models more broadly.
arXiv Detail & Related papers (2020-10-15T16:57:27Z) - Deducing neighborhoods of classes from a fitted model [68.8204255655161]
In this article a new kind of interpretable machine learning method is presented.
It can help to understand the partitioning of the feature space into predicted classes in a classification model using quantile shifts.
Basically, real data points (or specific points of interest) are used and the changes of the prediction after slightly raising or decreasing specific features are observed.
arXiv Detail & Related papers (2020-09-11T16:35:53Z) - A Bayesian machine scientist to aid in the solution of challenging
scientific problems [0.0]
We introduce a Bayesian machine scientist, which establishes the plausibility of models using explicit approximations to the exact marginal posterior over models.
It explores the space of models using Markov chain Monte Carlo.
We show that this approach uncovers accurate models for synthetic and real data and provides out-of-sample predictions that are more accurate than those of existing approaches.
arXiv Detail & Related papers (2020-04-25T14:42:13Z) - Amortized Bayesian model comparison with evidential deep learning [0.12314765641075436]
We propose a novel method for performing Bayesian model comparison using specialized deep learning architectures.
Our method is purely simulation-based and circumvents the step of explicitly fitting all alternative models under consideration to each observed dataset.
We show that our method achieves excellent results in terms of accuracy, calibration, and efficiency across the examples considered in this work.
arXiv Detail & Related papers (2020-04-22T15:15:46Z) - Predicting Multidimensional Data via Tensor Learning [0.0]
We develop a model that retains the intrinsic multidimensional structure of the dataset.
To estimate the model parameters, an Alternating Least Squares algorithm is developed.
The proposed model is able to outperform benchmark models present in the forecasting literature.
arXiv Detail & Related papers (2020-02-11T11:57:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.