Stable Update of Regression Trees
- URL: http://arxiv.org/abs/2402.13655v1
- Date: Wed, 21 Feb 2024 09:41:56 GMT
- Title: Stable Update of Regression Trees
- Authors: Morten Bl{\o}rstad, Berent {\AA}. S. Lunde, Nello Blaser
- Abstract summary: We focus on the stability of an inherently explainable machine learning method, namely regression trees.
We propose a regularization method, where data points are weighted based on the uncertainty in the initial model.
Results show that the proposed update method improves stability while achieving similar or better predictive performance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Updating machine learning models with new information usually improves their
predictive performance, yet, in many applications, it is also desirable to
avoid changing the model predictions too much. This property is called
stability. In most cases when stability matters, so does explainability. We
therefore focus on the stability of an inherently explainable machine learning
method, namely regression trees. We aim to use the notion of empirical
stability and design algorithms for updating regression trees that provide a
way to balance between predictability and empirical stability. To achieve this,
we propose a regularization method, where data points are weighted based on the
uncertainty in the initial model. The balance between predictability and
empirical stability can be adjusted through hyperparameters. This
regularization method is evaluated in terms of loss and stability and assessed
on a broad range of data characteristics. The results show that the proposed
update method improves stability while achieving similar or better predictive
performance. This shows that it is possible to achieve both predictive and
stable results when updating regression trees.
Related papers
- Using dynamic loss weighting to boost improvements in forecast stability [0.9332308328407303]
Rolling origin forecast instability refers to variability in forecasts for a specific period induced by updating the forecast.
In this paper, we empirically investigate whether further improvements in stability can be obtained without compromising accuracy by applying dynamic loss weighting algorithms.
arXiv Detail & Related papers (2024-09-26T20:21:46Z) - Predictability Analysis of Regression Problems via Conditional Entropy Estimations [1.8913544072080544]
Conditional entropy estimators are developed to assess predictability in regression problems.
Experiments on synthesized and real-world datasets demonstrate the robustness and utility of these estimators.
arXiv Detail & Related papers (2024-06-06T07:59:19Z) - On Forecast Stability [3.5789787644375495]
We explore two types of forecast stability that we call vertical stability and horizontal stability.
We propose a simple linear-interpolation-based approach that is applicable to stabilise the forecasts provided by any base model vertically and horizontally.
arXiv Detail & Related papers (2023-10-26T11:55:30Z) - Bagging Provides Assumption-free Stability [11.456416081243654]
Bagging is an important technique for stabilizing machine learning models.
In this paper, we derive a finite-sample guarantee on the stability of bagging for any model.
arXiv Detail & Related papers (2023-01-30T01:18:05Z) - Reliability-Aware Prediction via Uncertainty Learning for Person Image
Retrieval [51.83967175585896]
UAL aims at providing reliability-aware predictions by considering data uncertainty and model uncertainty simultaneously.
Data uncertainty captures the noise" inherent in the sample, while model uncertainty depicts the model's confidence in the sample's prediction.
arXiv Detail & Related papers (2022-10-24T17:53:20Z) - Numerically Stable Sparse Gaussian Processes via Minimum Separation
using Cover Trees [57.67528738886731]
We study the numerical stability of scalable sparse approximations based on inducing points.
For low-dimensional tasks such as geospatial modeling, we propose an automated method for computing inducing points satisfying these conditions.
arXiv Detail & Related papers (2022-10-14T15:20:17Z) - Probabilistic robust linear quadratic regulators with Gaussian processes [73.0364959221845]
Probabilistic models such as Gaussian processes (GPs) are powerful tools to learn unknown dynamical systems from data for subsequent use in control design.
We present a novel controller synthesis for linearized GP dynamics that yields robust controllers with respect to a probabilistic stability margin.
arXiv Detail & Related papers (2021-05-17T08:36:18Z) - Learning Probabilistic Ordinal Embeddings for Uncertainty-Aware
Regression [91.3373131262391]
Uncertainty is the only certainty there is.
Traditionally, the direct regression formulation is considered and the uncertainty is modeled by modifying the output space to a certain family of probabilistic distributions.
How to model the uncertainty within the present-day technologies for regression remains an open issue.
arXiv Detail & Related papers (2021-03-25T06:56:09Z) - Versatile and Robust Transient Stability Assessment via Instance
Transfer Learning [6.760999627905228]
This paper introduces a new data collection method in a data-driven algorithm incorporating the knowledge of power system dynamics.
We introduce a new concept called Fault-Affected Area, which provides crucial information regarding the unstable region of operation.
The test results on the IEEE 39-bus system verify that this model can accurately predict the stability of previously unseen operational scenarios.
arXiv Detail & Related papers (2021-02-20T09:10:29Z) - Fine-Grained Analysis of Stability and Generalization for Stochastic
Gradient Descent [55.85456985750134]
We introduce a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iterates.
This yields generalization bounds depending on the behavior of the best model, and leads to the first-ever-known fast bounds in the low-noise setting.
To our best knowledge, this gives the firstever-known stability and generalization for SGD with even non-differentiable loss functions.
arXiv Detail & Related papers (2020-06-15T06:30:19Z) - Balance-Subsampled Stable Prediction [55.13512328954456]
We propose a novel balance-subsampled stable prediction (BSSP) algorithm based on the theory of fractional factorial design.
A design-theoretic analysis shows that the proposed method can reduce the confounding effects among predictors induced by the distribution shift.
Numerical experiments on both synthetic and real-world data sets demonstrate that our BSSP algorithm significantly outperforms the baseline methods for stable prediction across unknown test data.
arXiv Detail & Related papers (2020-06-08T07:01:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.