Correlated Time Series Self-Supervised Representation Learning via
Spatiotemporal Bootstrapping
- URL: http://arxiv.org/abs/2306.06994v2
- Date: Tue, 20 Jun 2023 15:29:51 GMT
- Title: Correlated Time Series Self-Supervised Representation Learning via
Spatiotemporal Bootstrapping
- Authors: Luxuan Wang, Lei Bai, Ziyue Li, Rui Zhao, Fugee Tsung
- Abstract summary: Time series analysis plays an important role in many real-world industries.
In this paper, we propose a time-step-level representation learning framework for individual instances.
A linear regression model trained on top of the learned representations demonstrates our model performs best in most cases.
- Score: 13.988624652592259
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Correlated time series analysis plays an important role in many real-world
industries. Learning an efficient representation of this large-scale data for
further downstream tasks is necessary but challenging. In this paper, we
propose a time-step-level representation learning framework for individual
instances via bootstrapped spatiotemporal representation prediction. We
evaluated the effectiveness and flexibility of our representation learning
framework on correlated time series forecasting and cold-start transferring the
forecasting model to new instances with limited data. A linear regression model
trained on top of the learned representations demonstrates our model performs
best in most cases. Especially compared to representation learning models, we
reduce the RMSE, MAE, and MAPE by 37%, 49%, and 48% on the PeMS-BAY dataset,
respectively. Furthermore, in real-world metro passenger flow data, our
framework demonstrates the ability to transfer to infer future information of
new cold-start instances, with gains of 15%, 19%, and 18%. The source code will
be released under the GitHub
https://github.com/bonaldli/Spatiotemporal-TS-Representation-Learning
Related papers
- A Mamba Foundation Model for Time Series Forecasting [13.593170999506889]
We introduce TSMamba, a linear-complexity foundation model for time series forecasting built on the Mamba architecture.
The model captures temporal dependencies through both forward and backward Mamba encoders, achieving high prediction accuracy.
It also achieves competitive or superior full-shot performance compared to task-specific prediction models.
arXiv Detail & Related papers (2024-11-05T09:34:05Z) - Time Series Representation Models [2.724184832774005]
Time series analysis remains a major challenge due to its sparse characteristics, high dimensionality, and inconsistent data quality.
Recent advancements in transformer-based techniques have enhanced capabilities in forecasting and imputation.
We propose a new architectural concept for time series analysis based on introspection.
arXiv Detail & Related papers (2024-05-28T13:25:31Z) - No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency Determines Multimodal Model Performance [68.18779562801762]
multimodal models require exponentially more data to achieve linear improvements in downstream "zero-shot" performance.
Our study reveals an exponential need for training data which implies that the key to "zero-shot" generalization capabilities under large-scale training paradigms remains to be found.
arXiv Detail & Related papers (2024-04-04T17:58:02Z) - Pushing the Limits of Pre-training for Time Series Forecasting in the
CloudOps Domain [54.67888148566323]
We introduce three large-scale time series forecasting datasets from the cloud operations domain.
We show it is a strong zero-shot baseline and benefits from further scaling, both in model and dataset size.
Accompanying these datasets and results is a suite of comprehensive benchmark results comparing classical and deep learning baselines to our pre-trained method.
arXiv Detail & Related papers (2023-10-08T08:09:51Z) - Instance-Conditional Timescales of Decay for Non-Stationary Learning [11.90763787610444]
Slow concept drift is a ubiquitous, yet under-studied problem in machine learning systems.
We propose an optimization-driven approach towards balancing instance importance over large training windows.
Experiments on a large real-world dataset of 39M photos over a 9 year period show upto 15% relative gains in accuracy.
arXiv Detail & Related papers (2022-12-12T14:16:26Z) - The future is different: Large pre-trained language models fail in
prediction tasks [2.9005223064604078]
We introduce four new REDDIT datasets, namely the WALLSTREETBETS, ASKSCIENCE, THE DONALD, and POLITICS sub-reddits.
First, we empirically demonstrate that LPLM can display average performance drops of about 88% when predicting the popularity of future posts from sub-reddits whose topic distribution changes with time.
We then introduce a simple methodology that leverages neural variational dynamic topic models and attention mechanisms to infer temporal language model representations for regression tasks.
arXiv Detail & Related papers (2022-11-01T11:01:36Z) - Beyond Transfer Learning: Co-finetuning for Action Localisation [64.07196901012153]
We propose co-finetuning -- simultaneously training a single model on multiple upstream'' and downstream'' tasks.
We demonstrate that co-finetuning outperforms traditional transfer learning when using the same total amount of data.
We also show how we can easily extend our approach to multiple upstream'' datasets to further improve performance.
arXiv Detail & Related papers (2022-07-08T10:25:47Z) - Self-Supervised Pre-Training for Transformer-Based Person
Re-Identification [54.55281692768765]
Transformer-based supervised pre-training achieves great performance in person re-identification (ReID)
Due to the domain gap between ImageNet and ReID datasets, it usually needs a larger pre-training dataset to boost the performance.
This work aims to mitigate the gap between the pre-training and ReID datasets from the perspective of data and model structure.
arXiv Detail & Related papers (2021-11-23T18:59:08Z) - Relation-Guided Representation Learning [53.60351496449232]
We propose a new representation learning method that explicitly models and leverages sample relations.
Our framework well preserves the relations between samples.
By seeking to embed samples into subspace, we show that our method can address the large-scale and out-of-sample problem.
arXiv Detail & Related papers (2020-07-11T10:57:45Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.