SETAR-Tree: A Novel and Accurate Tree Algorithm for Global Time Series
Forecasting
- URL: http://arxiv.org/abs/2211.08661v1
- Date: Wed, 16 Nov 2022 04:30:42 GMT
- Title: SETAR-Tree: A Novel and Accurate Tree Algorithm for Global Time Series
Forecasting
- Authors: Rakshitha Godahewa, Geoffrey I. Webb, Daniel Schmidt, Christoph
Bergmeir
- Abstract summary: In this paper, we explore the close connections between TAR models and regression trees.
We introduce a new forecasting-specific tree algorithm that trains global Pooled Regression (PR) models in the leaves.
In our evaluation, the proposed tree and forest models are able to achieve significantly higher accuracy than a set of state-of-the-art tree-based algorithms.
- Score: 7.206754802573034
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Threshold Autoregressive (TAR) models have been widely used by statisticians
for non-linear time series forecasting during the past few decades, due to
their simplicity and mathematical properties. On the other hand, in the
forecasting community, general-purpose tree-based regression algorithms
(forests, gradient-boosting) have become popular recently due to their ease of
use and accuracy. In this paper, we explore the close connections between TAR
models and regression trees. These enable us to use the rich methodology from
the literature on TAR models to define a hierarchical TAR model as a regression
tree that trains globally across series, which we call SETAR-Tree. In contrast
to the general-purpose tree-based models that do not primarily focus on
forecasting, and calculate averages at the leaf nodes, we introduce a new
forecasting-specific tree algorithm that trains global Pooled Regression (PR)
models in the leaves allowing the models to learn cross-series information and
also uses some time-series-specific splitting and stopping procedures. The
depth of the tree is controlled by conducting a statistical linearity test
commonly employed in TAR models, as well as measuring the error reduction
percentage at each node split. Thus, the proposed tree model requires minimal
external hyperparameter tuning and provides competitive results under its
default configuration. We also use this tree algorithm to develop a forest
where the forecasts provided by a collection of diverse SETAR-Trees are
combined during the forecasting process. In our evaluation on eight publicly
available datasets, the proposed tree and forest models are able to achieve
significantly higher accuracy than a set of state-of-the-art tree-based
algorithms and forecasting benchmarks across four evaluation metrics.
Related papers
- Modern Neighborhood Components Analysis: A Deep Tabular Baseline Two Decades Later [59.88557193062348]
We revisit the classic Neighborhood Component Analysis (NCA), designed to learn a linear projection that captures semantic similarities between instances.
We find that minor modifications, such as adjustments to the learning objectives and the integration of deep learning architectures, significantly enhance NCA's performance.
We also introduce a neighbor sampling strategy that improves both the efficiency and predictive accuracy of our proposed ModernNCA.
arXiv Detail & Related papers (2024-07-03T16:38:57Z) - Forecasting with Hyper-Trees [50.72190208487953]
Hyper-Trees are designed to learn the parameters of time series models.
By relating the parameters of a target time series model to features, Hyper-Trees also address the issue of parameter non-stationarity.
In this novel approach, the trees first generate informative representations from the input features, which a shallow network then maps to the target model parameters.
arXiv Detail & Related papers (2024-05-13T15:22:15Z) - Hierarchical clustering with dot products recovers hidden tree structure [53.68551192799585]
In this paper we offer a new perspective on the well established agglomerative clustering algorithm, focusing on recovery of hierarchical structure.
We recommend a simple variant of the standard algorithm, in which clusters are merged by maximum average dot product and not, for example, by minimum distance or within-cluster variance.
We demonstrate that the tree output by this algorithm provides a bona fide estimate of generative hierarchical structure in data, under a generic probabilistic graphical model.
arXiv Detail & Related papers (2023-05-24T11:05:12Z) - Hierarchical Shrinkage: improving the accuracy and interpretability of
tree-based methods [10.289846887751079]
We introduce Hierarchical Shrinkage (HS), a post-hoc algorithm that does not modify the tree structure.
HS substantially increases the predictive performance of decision trees, even when used in conjunction with other regularization techniques.
All code and models are released in a full-fledged package available on Github.
arXiv Detail & Related papers (2022-02-02T02:43:23Z) - A cautionary tale on fitting decision trees to data from additive
models: generalization lower bounds [9.546094657606178]
We study the generalization performance of decision trees with respect to different generative regression models.
This allows us to elicit their inductive bias, that is, the assumptions the algorithms make (or do not make) to generalize to new data.
We prove a sharp squared error generalization lower bound for a large class of decision tree algorithms fitted to sparse additive models.
arXiv Detail & Related papers (2021-10-18T21:22:40Z) - Complex Event Forecasting with Prediction Suffix Trees: Extended
Technical Report [70.7321040534471]
Complex Event Recognition (CER) systems have become popular in the past two decades due to their ability to "instantly" detect patterns on real-time streams of events.
There is a lack of methods for forecasting when a pattern might occur before such an occurrence is actually detected by a CER engine.
We present a formal framework that attempts to address the issue of Complex Event Forecasting.
arXiv Detail & Related papers (2021-09-01T09:52:31Z) - Data-driven advice for interpreting local and global model predictions
in bioinformatics problems [17.685881417954782]
Conditional feature contributions (CFCs) provide textitlocal, case-by-case explanations of a prediction.
We compare the explanations computed by both methods on a set of 164 publicly available classification problems.
For random forests, we find extremely high similarities and correlations of both local and global SHAP values and CFC scores.
arXiv Detail & Related papers (2021-08-13T12:41:39Z) - Spectral Top-Down Recovery of Latent Tree Models [13.681975313065477]
Spectral Top-Down Recovery (STDR) is a divide-and-conquer approach for inference of large latent tree models.
STDR's partitioning step is non-random. Instead, it is based on the Fiedler vector of a suitable Laplacian matrix related to the observed nodes.
We prove that STDR is statistically consistent, and bound the number of samples required to accurately recover the tree with high probability.
arXiv Detail & Related papers (2021-02-26T02:47:42Z) - JSRT: James-Stein Regression Tree [55.2059664267247]
Regression tree (RT) has been widely used in machine learning and data mining community.
In practice, the performance of RT relies heavily on the local mean of samples from an individual node during the tree construction/prediction stage.
We propose a novel regression tree, named James-Stein Regression Tree (JSRT) by considering global information from different nodes.
arXiv Detail & Related papers (2020-10-18T16:28:49Z) - MurTree: Optimal Classification Trees via Dynamic Programming and Search [61.817059565926336]
We present a novel algorithm for learning optimal classification trees based on dynamic programming and search.
Our approach uses only a fraction of the time required by the state-of-the-art and can handle datasets with tens of thousands of instances.
arXiv Detail & Related papers (2020-07-24T17:06:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.