Multi-Objective Model Selection for Time Series Forecasting
- URL: http://arxiv.org/abs/2202.08485v1
- Date: Thu, 17 Feb 2022 07:40:15 GMT
- Title: Multi-Objective Model Selection for Time Series Forecasting
- Authors: Oliver Borchert, David Salinas, Valentin Flunkert, Tim Januschowski,
Stephan G\"unnemann
- Abstract summary: We present a benchmark, evaluating 7 classical and 6 deep learning forecasting methods on 44 datasets.
We leverage the benchmark evaluations to learn good defaults that consider multiple objectives such as accuracy and latency.
By learning a mapping from forecasting models to performance metrics, we show that our method PARETOSELECT is able to accurately select models.
- Score: 9.473440847947492
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Research on time series forecasting has predominantly focused on developing
methods that improve accuracy. However, other criteria such as training time or
latency are critical in many real-world applications. We therefore address the
question of how to choose an appropriate forecasting model for a given dataset
among the plethora of available forecasting methods when accuracy is only one
of many criteria. For this, our contributions are two-fold. First, we present a
comprehensive benchmark, evaluating 7 classical and 6 deep learning forecasting
methods on 44 heterogeneous, publicly available datasets. The benchmark code is
open-sourced along with evaluations and forecasts for all methods. These
evaluations enable us to answer open questions such as the amount of data
required for deep learning models to outperform classical ones. Second, we
leverage the benchmark evaluations to learn good defaults that consider
multiple objectives such as accuracy and latency. By learning a mapping from
forecasting models to performance metrics, we show that our method PARETOSELECT
is able to accurately select models from the Pareto front -- alleviating the
need to train or evaluate many forecasting models for model selection. To the
best of our knowledge, PARETOSELECT constitutes the first method to learn
default models in a multi-objective setting.
Related papers
- Context is Key: A Benchmark for Forecasting with Essential Textual Information [87.3175915185287]
"Context is Key" (CiK) is a time series forecasting benchmark that pairs numerical data with diverse types of carefully crafted textual context.
We evaluate a range of approaches, including statistical models, time series foundation models, and LLM-based forecasters.
Our experiments highlight the importance of incorporating contextual information, demonstrate surprising performance when using LLM-based forecasting models, and also reveal some of their critical shortcomings.
arXiv Detail & Related papers (2024-10-24T17:56:08Z) - GIFT-Eval: A Benchmark For General Time Series Forecasting Model Evaluation [90.53485251837235]
GIFT-Eval is a pioneering benchmark aimed at promoting evaluation across diverse datasets.
GIFT-Eval encompasses 28 datasets over 144,000 time series and 177 million data points.
We also provide a non-leaking pretraining dataset containing approximately 230 billion data points.
arXiv Detail & Related papers (2024-10-14T11:29:38Z) - A Two-Phase Recall-and-Select Framework for Fast Model Selection [13.385915962994806]
We propose a two-phase (coarse-recall and fine-selection) model selection framework.
It aims to enhance the efficiency of selecting a robust model by leveraging the models' training performances on benchmark datasets.
It has been demonstrated that the proposed methodology facilitates the selection of a high-performing model at a rate about 3x times faster than conventional baseline methods.
arXiv Detail & Related papers (2024-03-28T14:44:44Z) - DsDm: Model-Aware Dataset Selection with Datamodels [81.01744199870043]
Standard practice is to filter for examples that match human notions of data quality.
We find that selecting according to similarity with "high quality" data sources may not increase (and can even hurt) performance compared to randomly selecting data.
Our framework avoids handpicked notions of data quality, and instead models explicitly how the learning process uses train datapoints to predict on the target tasks.
arXiv Detail & Related papers (2024-01-23T17:22:00Z) - Anchor Points: Benchmarking Models with Much Fewer Examples [88.02417913161356]
In six popular language classification benchmarks, model confidence in the correct class on many pairs of points is strongly correlated across models.
We propose Anchor Point Selection, a technique to select small subsets of datasets that capture model behavior across the entire dataset.
Just several anchor points can be used to estimate model per-class predictions on all other points in a dataset with low mean absolute error.
arXiv Detail & Related papers (2023-09-14T17:45:51Z) - Post-Selection Confidence Bounds for Prediction Performance [2.28438857884398]
In machine learning, the selection of a promising model from a potentially large number of competing models and the assessment of its generalization performance are critical tasks.
We propose an algorithm how to compute valid lower confidence bounds for multiple models that have been selected based on their prediction performances in the evaluation set.
arXiv Detail & Related papers (2022-10-24T13:28:43Z) - fETSmcs: Feature-based ETS model component selection [8.99236558175168]
We propose an efficient approach for ETS model selection by training classifiers on simulated data to predict appropriate model component forms for a given time series.
We evaluate our approach on the widely used forecasting competition data set M4 in terms of both point forecasts and prediction intervals.
arXiv Detail & Related papers (2022-06-26T13:52:43Z) - Model Selection for Time Series Forecasting: Empirical Analysis of
Different Estimators [1.6328866317851185]
We compare a set of estimation methods for model selection in time series forecasting tasks.
We empirically found that the accuracy of the estimators for selecting the best solution is low.
Some factors, such as the sample size, are important in the relative performance of the estimators.
arXiv Detail & Related papers (2021-04-01T16:08:25Z) - Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual
Model-Based Reinforcement Learning [109.74041512359476]
We study a number of design decisions for the predictive model in visual MBRL algorithms.
We find that a range of design decisions that are often considered crucial, such as the use of latent spaces, have little effect on task performance.
We show how this phenomenon is related to exploration and how some of the lower-scoring models on standard benchmarks will perform the same as the best-performing models when trained on the same training data.
arXiv Detail & Related papers (2020-12-08T18:03:21Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z) - For2For: Learning to forecast from forecasts [1.6752182911522522]
This paper presents a time series forecasting framework which combines standard forecasting methods and a machine learning model.
Tested on the M4 competition dataset, this approach outperforms all submissions for quarterly series, and is more accurate than all but the winning algorithm for monthly series.
arXiv Detail & Related papers (2020-01-14T03:06:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.