Towards a Unified Framework for Uncertainty-aware Nonlinear Variable
Selection with Theoretical Guarantees
- URL: http://arxiv.org/abs/2204.07293v1
- Date: Fri, 15 Apr 2022 02:12:00 GMT
- Title: Towards a Unified Framework for Uncertainty-aware Nonlinear Variable
Selection with Theoretical Guarantees
- Authors: Wenying Deng, Beau Coker, Jeremiah Zhe Liu, Brent A. Coull
- Abstract summary: We develop a simple and unified framework for nonlinear variable selection that incorporates model uncertainty.
We show that the approach is generalizable even to non-differentiable models such as tree ensembles.
- Score: 2.1506382989223782
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We develop a simple and unified framework for nonlinear variable selection
that incorporates model uncertainty and is compatible with a wide range of
machine learning models (e.g., tree ensembles, kernel methods and neural
network). In particular, for a learned nonlinear model $f(\mathbf{x})$, we
consider quantifying the importance of an input variable $\mathbf{x}^j$ using
the integrated gradient measure $\psi_j = \Vert \frac{\partial}{\partial
\mathbf{x}^j} f(\mathbf{x})\Vert^2_2$. We then (1) provide a principled
approach for quantifying variable selection uncertainty by deriving its
posterior distribution, and (2) show that the approach is generalizable even to
non-differentiable models such as tree ensembles. Rigorous Bayesian
nonparametric theorems are derived to guarantee the posterior consistency and
asymptotic uncertainty of the proposed approach. Extensive simulation confirms
that the proposed algorithm outperforms existing classic and recent variable
selection methods.
Related papers
- A Unified Theory of Stochastic Proximal Point Methods without Smoothness [52.30944052987393]
Proximal point methods have attracted considerable interest owing to their numerical stability and robustness against imperfect tuning.
This paper presents a comprehensive analysis of a broad range of variations of the proximal point method (SPPM)
arXiv Detail & Related papers (2024-05-24T21:09:19Z) - Nonlinear Distributionally Robust Optimization [5.0490573482829335]
This article focuses on a class of distributionally robust optimization (DRO) problems where the objective function is potentially nonlinear in the distribution.
We propose an alternative notion for the derivative and corresponding smoothness based on Gateaux (G)-derivative for generic risk measures.
We use the set-up of the FW algorithm to devise a methodology to compute a saddle point of the nonlinear DRO problem.
arXiv Detail & Related papers (2023-06-05T19:22:02Z) - Gradient-Free Methods for Deterministic and Stochastic Nonsmooth
Nonconvex Optimization [94.19177623349947]
Non-smooth non optimization problems emerge in machine learning and business making.
Two core challenges impede the development of efficient methods with finitetime convergence guarantee.
Two-phase versions of GFM and SGFM are also proposed and proven to achieve improved large-deviation results.
arXiv Detail & Related papers (2022-09-12T06:53:24Z) - A Variational Inference Approach to Inverse Problems with Gamma
Hyperpriors [60.489902135153415]
This paper introduces a variational iterative alternating scheme for hierarchical inverse problems with gamma hyperpriors.
The proposed variational inference approach yields accurate reconstruction, provides meaningful uncertainty quantification, and is easy to implement.
arXiv Detail & Related papers (2021-11-26T06:33:29Z) - Universal and data-adaptive algorithms for model selection in linear
contextual bandits [52.47796554359261]
We consider the simplest non-trivial instance of model-selection: distinguishing a simple multi-armed bandit problem from a linear contextual bandit problem.
We introduce new algorithms that explore in a data-adaptive manner and provide guarantees of the form $mathcalO(dalpha T1- alpha)$.
Our approach extends to model selection among nested linear contextual bandits under some additional assumptions.
arXiv Detail & Related papers (2021-11-08T18:05:35Z) - Combining Gaussian processes and polynomial chaos expansions for
stochastic nonlinear model predictive control [0.0]
We introduce a new algorithm to explicitly consider time-invariant uncertainties in optimal control problems.
The main novelty in this paper is to use this combination in an efficient fashion to obtain mean and variance estimates of nonlinear transformations.
It is shown how to formulate both chance-constraints and a probabilistic objective for the optimal control problem.
arXiv Detail & Related papers (2021-03-09T14:25:08Z) - Probabilistic Circuits for Variational Inference in Discrete Graphical
Models [101.28528515775842]
Inference in discrete graphical models with variational methods is difficult.
Many sampling-based methods have been proposed for estimating Evidence Lower Bound (ELBO)
We propose a new approach that leverages the tractability of probabilistic circuit models, such as Sum Product Networks (SPN)
We show that selective-SPNs are suitable as an expressive variational distribution, and prove that when the log-density of the target model is aweighted the corresponding ELBO can be computed analytically.
arXiv Detail & Related papers (2020-10-22T05:04:38Z) - Identification of Probability weighted ARX models with arbitrary domains [75.91002178647165]
PieceWise Affine models guarantees universal approximation, local linearity and equivalence to other classes of hybrid system.
In this work, we focus on the identification of PieceWise Auto Regressive with eXogenous input models with arbitrary regions (NPWARX)
The architecture is conceived following the Mixture of Expert concept, developed within the machine learning field.
arXiv Detail & Related papers (2020-09-29T12:50:33Z) - Mean-Field Approximation to Gaussian-Softmax Integral with Application
to Uncertainty Estimation [23.38076756988258]
We propose a new single-model based approach to quantify uncertainty in deep neural networks.
We use a mean-field approximation formula to compute an analytically intractable integral.
Empirically, the proposed approach performs competitively when compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-06-13T07:32:38Z) - A Precise High-Dimensional Asymptotic Theory for Boosting and
Minimum-$\ell_1$-Norm Interpolated Classifiers [3.167685495996986]
This paper establishes a precise high-dimensional theory for boosting on separable data.
Under a class of statistical models, we provide an exact analysis of the universality error of boosting.
We also explicitly pin down the relation between the boosting test error and the optimal Bayes error.
arXiv Detail & Related papers (2020-02-05T00:24:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.