Flexible Bayesian Nonlinear Model Configuration
- URL: http://arxiv.org/abs/2003.02929v2
- Date: Tue, 23 Nov 2021 11:15:56 GMT
- Title: Flexible Bayesian Nonlinear Model Configuration
- Authors: Aliaksandr Hubin, Geir Storvik, Florian Frommlet
- Abstract summary: Linear, or simple parametric, models are often not sufficient to describe complex relationships between input variables and a response.
We introduce a flexible approach for the construction and selection of highly flexible nonlinear parametric regression models.
A genetically modified mode jumping chain Monte Carlo algorithm is adopted to perform Bayesian inference.
- Score: 10.865434331546126
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Regression models are used in a wide range of applications providing a
powerful scientific tool for researchers from different fields. Linear, or
simple parametric, models are often not sufficient to describe complex
relationships between input variables and a response. Such relationships can be
better described through flexible approaches such as neural networks, but this
results in less interpretable models and potential overfitting. Alternatively,
specific parametric nonlinear functions can be used, but the specification of
such functions is in general complicated. In this paper, we introduce a
flexible approach for the construction and selection of highly flexible
nonlinear parametric regression models. Nonlinear features are generated
hierarchically, similarly to deep learning, but have additional flexibility on
the possible types of features to be considered. This flexibility, combined
with variable selection, allows us to find a small set of important features
and thereby more interpretable models. Within the space of possible functions,
a Bayesian approach, introducing priors for functions based on their
complexity, is considered. A genetically modified mode jumping Markov chain
Monte Carlo algorithm is adopted to perform Bayesian inference and estimate
posterior probabilities for model averaging. In various applications, we
illustrate how our approach is used to obtain meaningful nonlinear models.
Additionally, we compare its predictive performance with several machine
learning algorithms.
Related papers
- Shape Arithmetic Expressions: Advancing Scientific Discovery Beyond Closed-Form Equations [56.78271181959529]
Generalized Additive Models (GAMs) can capture non-linear relationships between variables and targets, but they cannot capture intricate feature interactions.
We propose Shape Expressions Arithmetic ( SHAREs) that fuses GAM's flexible shape functions with the complex feature interactions found in mathematical expressions.
We also design a set of rules for constructing SHAREs that guarantee transparency of the found expressions beyond the standard constraints.
arXiv Detail & Related papers (2024-04-15T13:44:01Z) - RUMBoost: Gradient Boosted Random Utility Models [0.0]
The RUMBoost model combines the interpretability and behavioural robustness of Random Utility Models (RUMs) with the generalisation and predictive ability of deep learning methods.
We demonstrate the potential of the RUMBoost model compared to various ML and Random Utility benchmark models for revealed preference mode choice data from London.
arXiv Detail & Related papers (2024-01-22T13:54:26Z) - The Contextual Lasso: Sparse Linear Models via Deep Neural Networks [5.607237982617641]
We develop a new statistical estimator that fits a sparse linear model to the explanatory features such that the sparsity pattern and coefficients vary as a function of the contextual features.
An extensive suite of experiments on real and synthetic data suggests that the learned models, which remain highly transparent, can be sparser than the regular lasso.
arXiv Detail & Related papers (2023-02-02T05:00:29Z) - Constructing Effective Machine Learning Models for the Sciences: A
Multidisciplinary Perspective [77.53142165205281]
We show how flexible non-linear solutions will not always improve upon manually adding transforms and interactions between variables to linear regression models.
We discuss how to recognize this before constructing a data-driven model and how such analysis can help us move to intrinsically interpretable regression models.
arXiv Detail & Related papers (2022-11-21T17:48:44Z) - Variational Hierarchical Mixtures for Probabilistic Learning of Inverse
Dynamics [20.953728061894044]
Well-calibrated probabilistic regression models are a crucial learning component in robotics applications as datasets grow rapidly and tasks become more complex.
We consider a probabilistic hierarchical modeling paradigm that combines the benefits of both worlds to deliver computationally efficient representations with inherent complexity regularization.
We derive two efficient variational inference techniques to learn these representations and highlight the advantages of hierarchical infinite local regression models.
arXiv Detail & Related papers (2022-11-02T13:54:07Z) - Decoupling multivariate functions using a nonparametric filtered tensor
decomposition [0.29360071145551075]
Decoupling techniques aim at providing an alternative representation of the nonlinearity.
The so-called decoupled form is often a more efficient parameterisation of the relationship while being highly structured, favouring interpretability.
In this work two new algorithms, based on filtered tensor decompositions of first order derivative information are introduced.
arXiv Detail & Related papers (2022-05-23T09:34:17Z) - Scalable Bayesian Network Structure Learning with Splines [2.741266294612776]
A Bayesian Network (BN) is a probabilistic graphical model consisting of a directed acyclic graph (DAG)
We present a novel approach capable of learning the global DAG structure of a BN and modelling linear and non-linear local relationships between variables.
arXiv Detail & Related papers (2021-10-27T17:54:53Z) - Model-agnostic multi-objective approach for the evolutionary discovery
of mathematical models [55.41644538483948]
In modern data science, it is more interesting to understand the properties of the model, which parts could be replaced to obtain better results.
We use multi-objective evolutionary optimization for composite data-driven model learning to obtain the algorithm's desired properties.
arXiv Detail & Related papers (2021-07-07T11:17:09Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Bilinear Classes: A Structural Framework for Provable Generalization in
RL [119.42509700822484]
Bilinear Classes is a new structural framework which permits generalization in reinforcement learning.
The framework incorporates nearly all existing models in which a sample complexity is achievable.
Our main result provides an RL algorithm which has sample complexity for Bilinear Classes.
arXiv Detail & Related papers (2021-03-19T16:34:20Z) - Hessian Eigenspectra of More Realistic Nonlinear Models [73.31363313577941]
We make a emphprecise characterization of the Hessian eigenspectra for a broad family of nonlinear models.
Our analysis takes a step forward to identify the origin of many striking features observed in more complex machine learning models.
arXiv Detail & Related papers (2021-03-02T06:59:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.