Lag-Llama: Towards Foundation Models for Probabilistic Time Series
Forecasting
- URL: http://arxiv.org/abs/2310.08278v3
- Date: Thu, 8 Feb 2024 05:49:04 GMT
- Title: Lag-Llama: Towards Foundation Models for Probabilistic Time Series
Forecasting
- Authors: Kashif Rasul, Arjun Ashok, Andrew Robert Williams, Hena Ghonia,
Rishika Bhagwatkar, Arian Khorasani, Mohammad Javad Darvishi Bayazi, George
Adamopoulos, Roland Riachi, Nadhir Hassen, Marin Bilo\v{s}, Sahil Garg,
Anderson Schneider, Nicolas Chapados, Alexandre Drouin, Valentina
Zantedeschi, Yuriy Nevmyvaka, Irina Rish
- Abstract summary: We present Lag-Llama, a general-purpose foundation model for time series forecasting based on a decoder-only transformer architecture.
Lag-Llama is pretrained on a large corpus of diverse time series data from several domains, and demonstrates strong zero-shot generalization capabilities.
When fine-tuned on relatively small fractions of such previously unseen datasets, Lag-Llama achieves state-of-the-art performance.
- Score: 54.04430089029033
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Over the past years, foundation models have caused a paradigm shift in
machine learning due to their unprecedented capabilities for zero-shot and
few-shot generalization. However, despite the success of foundation models in
modalities such as natural language processing and computer vision, the
development of foundation models for time series forecasting has lagged behind.
We present Lag-Llama, a general-purpose foundation model for univariate
probabilistic time series forecasting based on a decoder-only transformer
architecture that uses lags as covariates. Lag-Llama is pretrained on a large
corpus of diverse time series data from several domains, and demonstrates
strong zero-shot generalization capabilities compared to a wide range of
forecasting models on downstream datasets across domains. Moreover, when
fine-tuned on relatively small fractions of such previously unseen datasets,
Lag-Llama achieves state-of-the-art performance, outperforming prior deep
learning approaches, emerging as the best general-purpose model on average.
Lag-Llama serves as a strong contender to the current state-of-art in time
series forecasting and paves the way for future advancements in foundation
models tailored to time series data.
Related papers
- GIFT-Eval: A Benchmark For General Time Series Forecasting Model Evaluation [90.53485251837235]
Time series foundation models excel in zero-shot forecasting, handling diverse tasks without explicit training.
GIFT-Eval is a pioneering benchmark aimed at promoting evaluation across diverse datasets.
GIFT-Eval encompasses 23 datasets over 144,000 time series and 177 million data points.
arXiv Detail & Related papers (2024-10-14T11:29:38Z) - Scaling Law for Time Series Forecasting [8.967263259533036]
Scaling law that rewards large datasets, complex models and enhanced data granularity has been observed in various fields of deep learning.
Yet, studies on time series forecasting have cast doubt on scaling behaviors of deep learning methods for time series forecasting.
We propose a theory for scaling law for time series forecasting that can explain these seemingly abnormal behaviors.
arXiv Detail & Related papers (2024-05-24T00:46:27Z) - Unified Training of Universal Time Series Forecasting Transformers [104.56318980466742]
We present a Masked-based Universal Time Series Forecasting Transformer (Moirai)
Moirai is trained on our newly introduced Large-scale Open Time Series Archive (LOTSA) featuring over 27B observations across nine domains.
Moirai achieves competitive or superior performance as a zero-shot forecaster when compared to full-shot models.
arXiv Detail & Related papers (2024-02-04T20:00:45Z) - Timer: Generative Pre-trained Transformers Are Large Time Series Models [83.03091523806668]
This paper aims at the early development of large time series models (LTSM)
During pre-training, we curate large-scale datasets with up to 1 billion time points.
To meet diverse application needs, we convert forecasting, imputation, and anomaly detection of time series into a unified generative task.
arXiv Detail & Related papers (2024-02-04T06:55:55Z) - A decoder-only foundation model for time-series forecasting [23.824504640087753]
Our model is based on pretraining a patched-decoder style attention model on a large time-series corpus.
It can work well across different forecasting history lengths, prediction lengths and temporal granularities.
arXiv Detail & Related papers (2023-10-14T17:01:37Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - Time-LLM: Time Series Forecasting by Reprogramming Large Language Models [110.20279343734548]
Time series forecasting holds significant importance in many real-world dynamic systems.
We present Time-LLM, a reprogramming framework to repurpose large language models for time series forecasting.
Time-LLM is a powerful time series learner that outperforms state-of-the-art, specialized forecasting models.
arXiv Detail & Related papers (2023-10-03T01:31:25Z) - Improving the Accuracy of Global Forecasting Models using Time Series
Data Augmentation [7.38079566297881]
Forecasting models that are trained across sets of many time series, known as Global Forecasting Models (GFM), have shown promising results in forecasting competitions and real-world applications.
We propose a novel, data augmentation based forecasting framework that is capable of improving the baseline accuracy of GFM models in less data-abundant settings.
arXiv Detail & Related papers (2020-08-06T13:52:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.