Quantile Extreme Gradient Boosting for Uncertainty Quantification
- URL: http://arxiv.org/abs/2304.11732v1
- Date: Sun, 23 Apr 2023 19:46:19 GMT
- Title: Quantile Extreme Gradient Boosting for Uncertainty Quantification
- Authors: Xiaozhe Yin, Masoud Fallah-Shorshani, Rob McConnell, Scott Fruin,
Yao-Yi Chiang, Meredith Franklin
- Abstract summary: Extreme Gradient Boosting (XGBoost) is one of the most popular machine learning (ML) methods.
We propose enhancements to XGBoost whereby a modified quantile regression is used as the objective function to estimate uncertainty (QXGBoost)
Our proposed method had comparable or better performance than the uncertainty estimates generated for regular and quantile light gradient boosting.
- Score: 1.7685947618629572
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As the availability, size and complexity of data have increased in recent
years, machine learning (ML) techniques have become popular for modeling.
Predictions resulting from applying ML models are often used for inference,
decision-making, and downstream applications. A crucial yet often overlooked
aspect of ML is uncertainty quantification, which can significantly impact how
predictions from models are used and interpreted.
Extreme Gradient Boosting (XGBoost) is one of the most popular ML methods
given its simple implementation, fast computation, and sequential learning,
which make its predictions highly accurate compared to other methods. However,
techniques for uncertainty determination in ML models such as XGBoost have not
yet been universally agreed among its varying applications. We propose
enhancements to XGBoost whereby a modified quantile regression is used as the
objective function to estimate uncertainty (QXGBoost). Specifically, we
included the Huber norm in the quantile regression model to construct a
differentiable approximation to the quantile regression error function. This
key step allows XGBoost, which uses a gradient-based optimization algorithm, to
make probabilistic predictions efficiently.
QXGBoost was applied to create 90\% prediction intervals for one simulated
dataset and one real-world environmental dataset of measured traffic noise. Our
proposed method had comparable or better performance than the uncertainty
estimates generated for regular and quantile light gradient boosting. For both
the simulated and traffic noise datasets, the overall performance of the
prediction intervals from QXGBoost were better than other models based on
coverage width-based criterion.
Related papers
- Gradient-free variational learning with conditional mixture networks [39.827869318925494]
Conditional mixture networks (CMNs) are suitable for fast, gradient-free inference and can solve complex classification tasks.
We validate this approach by training two-layer CMNs on standard benchmarks from the UCI repository.
Our method, CAVI-CMN, achieves competitive and often superior predictive accuracy compared to maximum likelihood estimation (MLE) with backpropagation.
arXiv Detail & Related papers (2024-08-29T10:43:55Z) - Bayesian Deep Learning for Remaining Useful Life Estimation via Stein
Variational Gradient Descent [14.784809634505903]
We show that Bayesian deep learning models trained via Stein variational gradient descent consistently outperform with respect to convergence speed and predictive performance.
We propose a method to enhance performance based on the uncertainty information provided by the Bayesian models.
arXiv Detail & Related papers (2024-02-02T02:21:06Z) - Beyond mirkwood: Enhancing SED Modeling with Conformal Predictions [0.0]
We propose an advanced machine learning-based approach that enhances flexibility and uncertainty in SED fitting.
We incorporate conformalized quantile regression to convert point predictions into error bars, enhancing interpretability and reliability.
arXiv Detail & Related papers (2023-12-21T11:27:20Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Efficient and Differentiable Conformal Prediction with General Function
Classes [96.74055810115456]
We propose a generalization of conformal prediction to multiple learnable parameters.
We show that it achieves approximate valid population coverage and near-optimal efficiency within class.
Experiments show that our algorithm is able to learn valid prediction sets and improve the efficiency significantly.
arXiv Detail & Related papers (2022-02-22T18:37:23Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Flexible Model Aggregation for Quantile Regression [92.63075261170302]
Quantile regression is a fundamental problem in statistical learning motivated by a need to quantify uncertainty in predictions.
We investigate methods for aggregating any number of conditional quantile models.
All of the models we consider in this paper can be fit using modern deep learning toolkits.
arXiv Detail & Related papers (2021-02-26T23:21:16Z) - Causal Gradient Boosting: Boosted Instrumental Variable Regression [2.831053006774813]
We propose an alternative algorithm called boostIV that builds on the traditional gradient boosting algorithm and corrects for the endogeneity bias.
Our approach is data driven, meaning that the researcher does not have to make a stance on neither the form of the target function approximation nor the choice of instruments.
We show that boostIV is at worst on par with the existing methods and on average significantly outperforms them.
arXiv Detail & Related papers (2021-01-15T11:54:25Z) - Evaluating Prediction-Time Batch Normalization for Robustness under
Covariate Shift [81.74795324629712]
We call prediction-time batch normalization, which significantly improves model accuracy and calibration under covariate shift.
We show that prediction-time batch normalization provides complementary benefits to existing state-of-the-art approaches for improving robustness.
The method has mixed results when used alongside pre-training, and does not seem to perform as well under more natural types of dataset shift.
arXiv Detail & Related papers (2020-06-19T05:08:43Z) - Gaussian Process Boosting [13.162429430481982]
We introduce a novel way to combine boosting with Gaussian process and mixed effects models.
We obtain increased prediction accuracy compared to existing approaches on simulated and real-world data sets.
arXiv Detail & Related papers (2020-04-06T13:19:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.