RUMBoost: Gradient Boosted Random Utility Models
- URL: http://arxiv.org/abs/2401.11954v1
- Date: Mon, 22 Jan 2024 13:54:26 GMT
- Title: RUMBoost: Gradient Boosted Random Utility Models
- Authors: Nicolas Salvad\'e, Tim Hillel
- Abstract summary: The RUMBoost model combines the interpretability and behavioural robustness of Random Utility Models (RUMs) with the generalisation and predictive ability of deep learning methods.
We demonstrate the potential of the RUMBoost model compared to various ML and Random Utility benchmark models for revealed preference mode choice data from London.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper introduces the RUMBoost model, a novel discrete choice modelling
approach that combines the interpretability and behavioural robustness of
Random Utility Models (RUMs) with the generalisation and predictive ability of
deep learning methods. We obtain the full functional form of non-linear utility
specifications by replacing each linear parameter in the utility functions of a
RUM with an ensemble of gradient boosted regression trees. This enables
piece-wise constant utility values to be imputed for all alternatives directly
from the data for any possible combination of input variables. We introduce
additional constraints on the ensembles to ensure three crucial features of the
utility specifications: (i) dependency of the utilities of each alternative on
only the attributes of that alternative, (ii) monotonicity of marginal
utilities, and (iii) an intrinsically interpretable functional form, where the
exact response of the model is known throughout the entire input space.
Furthermore, we introduce an optimisation-based smoothing technique that
replaces the piece-wise constant utility values of alternative attributes with
monotonic piece-wise cubic splines to identify non-linear parameters with
defined gradient. We demonstrate the potential of the RUMBoost model compared
to various ML and Random Utility benchmark models for revealed preference mode
choice data from London. The results highlight the great predictive performance
and the direct interpretability of our proposed approach. Furthermore, the
smoothed attribute utility functions allow for the calculation of various
behavioural indicators and marginal utilities. Finally, we demonstrate the
flexibility of our methodology by showing how the RUMBoost model can be
extended to complex model specifications, including attribute interactions,
correlation within alternative error terms and heterogeneity within the
population.
Related papers
- Generalized Sparse Additive Model with Unknown Link Function [19.807823040041896]
We propose a new sparse additive model, named generalized sparse additive model with unknown link function (GSAMUL)
The component functions are estimated by B-spline basis and the unknown link function is estimated by a multi-layer perceptron (MLP) network.
In applications, experimental evaluations on both synthetic and real world data sets consistently validate the effectiveness of the proposed approach.
arXiv Detail & Related papers (2024-10-08T13:13:58Z) - Scaling and renormalization in high-dimensional regression [72.59731158970894]
This paper presents a succinct derivation of the training and generalization performance of a variety of high-dimensional ridge regression models.
We provide an introduction and review of recent results on these topics, aimed at readers with backgrounds in physics and deep learning.
arXiv Detail & Related papers (2024-05-01T15:59:00Z) - Break The Spell Of Total Correlation In betaTCVAE [4.38301148531795]
This paper proposes a new iterative decomposition path of total correlation and explains the disentangled representation ability of VAE.
The novel model enables VAE to adjust the parameter capacity to divide dependent and independent data features flexibly.
arXiv Detail & Related papers (2022-10-17T07:16:53Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - MACE: An Efficient Model-Agnostic Framework for Counterfactual
Explanation [132.77005365032468]
We propose a novel framework of Model-Agnostic Counterfactual Explanation (MACE)
In our MACE approach, we propose a novel RL-based method for finding good counterfactual examples and a gradient-less descent method for improving proximity.
Experiments on public datasets validate the effectiveness with better validity, sparsity and proximity.
arXiv Detail & Related papers (2022-05-31T04:57:06Z) - Spike-and-Slab Generalized Additive Models and Scalable Algorithms for
High-Dimensional Data [0.0]
We propose hierarchical generalized additive models (GAMs) to accommodate high-dimensional data.
We consider the smoothing penalty for proper shrinkage of curve and separation of smoothing function linear and nonlinear spaces.
Two and deterministic algorithms, EM-Coordinate Descent and EM-Iterative Weighted Least Squares, are developed for different utilities.
arXiv Detail & Related papers (2021-10-27T14:11:13Z) - Information-theoretic Feature Selection via Tensor Decomposition and
Submodularity [38.05393186002834]
We introduce a low-rank tensor model of the joint PMF of all variables and indirect targeting as a way of mitigating complexity and maximizing the classification performance for a given number of features.
By indirectly aiming to predict the latent variable of the naive Bayes model instead of the original target variable, it is possible to formulate the feature selection problem as of a monotone submodular function subject to a cardinality constraint.
arXiv Detail & Related papers (2020-10-30T10:36:46Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z) - Identification of Probability weighted ARX models with arbitrary domains [75.91002178647165]
PieceWise Affine models guarantees universal approximation, local linearity and equivalence to other classes of hybrid system.
In this work, we focus on the identification of PieceWise Auto Regressive with eXogenous input models with arbitrary regions (NPWARX)
The architecture is conceived following the Mixture of Expert concept, developed within the machine learning field.
arXiv Detail & Related papers (2020-09-29T12:50:33Z) - Slice Sampling for General Completely Random Measures [74.24975039689893]
We present a novel Markov chain Monte Carlo algorithm for posterior inference that adaptively sets the truncation level using auxiliary slice variables.
The efficacy of the proposed algorithm is evaluated on several popular nonparametric models.
arXiv Detail & Related papers (2020-06-24T17:53:53Z) - Misspecification-robust likelihood-free inference in high dimensions [13.934999364767918]
We introduce an extension of the popular Bayesian optimisation based approach to approximate discrepancy functions in a probabilistic manner.
Our approach achieves computational scalability for higher dimensional parameter spaces by using separate acquisition functions and discrepancies for each parameter.
The method successfully performs computationally efficient inference in a 100-dimensional space on canonical examples and compares favourably to existing modularised ABC methods.
arXiv Detail & Related papers (2020-02-21T16:06:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.