Self-Supervised Contrastive Pre-Training For Time Series via
Time-Frequency Consistency
- URL: http://arxiv.org/abs/2206.08496v1
- Date: Fri, 17 Jun 2022 00:45:04 GMT
- Title: Self-Supervised Contrastive Pre-Training For Time Series via
Time-Frequency Consistency
- Authors: Xiang Zhang, Ziyuan Zhao, Theodoros Tsiligkaridis, Marinka Zitnik
- Abstract summary: Methods need to accommodate target domains with different temporal dynamics.
Time-frequency consistency (TF-C) is desirable for pre-training.
Experiments show TF-C outperforms baselines by 15.4% (F1 score) on average in one-to-one settings.
- Score: 19.1862172442857
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Pre-training on time series poses a unique challenge due to the potential
mismatch between pre-training and target domains, such as shifts in temporal
dynamics, fast-evolving trends, and long-range and short cyclic effects, which
can lead to poor downstream performance. While domain adaptation methods can
mitigate these shifts, most methods need examples directly from the target
domain, making them suboptimal for pre-training. To address this challenge,
methods need to accommodate target domains with different temporal dynamics and
be capable of doing so without seeing any target examples during pre-training.
Relative to other modalities, in time series, we expect that time-based and
frequency-based representations of the same example are located close together
in the time-frequency space. To this end, we posit that time-frequency
consistency (TF-C) -- embedding a time-based neighborhood of a particular
example close to its frequency-based neighborhood and back -- is desirable for
pre-training. Motivated by TF-C, we define a decomposable pre-training model,
where the self-supervised signal is provided by the distance between time and
frequency components, each individually trained by contrastive estimation. We
evaluate the new method on eight datasets, including electrodiagnostic testing,
human activity recognition, mechanical fault detection, and physical status
monitoring. Experiments against eight state-of-the-art methods show that TF-C
outperforms baselines by 15.4% (F1 score) on average in one-to-one settings
(e.g., fine-tuning an EEG-pretrained model on EMG data) and by up to 8.4% (F1
score) in challenging one-to-many settings, reflecting the breadth of scenarios
that arise in real-world applications. The source code and datasets are
available at https: //anonymous.4open.science/r/TFC-pretraining-6B07.
Related papers
- Towards Generalisable Time Series Understanding Across Domains [10.350643783811174]
We introduce OTiS, an open model for general time series analysis.
We propose a novel pre-training paradigm including a tokeniser with learnable domain-specific signatures.
Our model is pre-trained on a large corpus of 640,187 samples and 11 billion time points spanning 8 distinct domains.
arXiv Detail & Related papers (2024-10-09T17:09:30Z) - Fast Information Streaming Handler (FisH): A Unified Seismic Neural Network for Single Station Real-Time Earthquake Early Warning [56.45067876391473]
Existing EEW approaches treat phase picking, location estimation, and magnitude estimation as separate tasks, lacking a unified framework.
We propose a novel unified seismic neural network called Fast Information Streaming Handler (FisH)
FisH is designed to process real-time streaming seismic data and generate simultaneous results for phase picking, location estimation, and magnitude estimation in an end-to-end fashion.
arXiv Detail & Related papers (2024-08-13T04:33:23Z) - PeFAD: A Parameter-Efficient Federated Framework for Time Series Anomaly Detection [51.20479454379662]
We propose a.
Federated Anomaly Detection framework named PeFAD with the increasing privacy concerns.
We conduct extensive evaluations on four real datasets, where PeFAD outperforms existing state-of-the-art baselines by up to 28.74%.
arXiv Detail & Related papers (2024-06-04T13:51:08Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Time-series Generation by Contrastive Imitation [87.51882102248395]
We study a generative framework that seeks to combine the strengths of both: Motivated by a moment-matching objective to mitigate compounding error, we optimize a local (but forward-looking) transition policy.
At inference, the learned policy serves as the generator for iterative sampling, and the learned energy serves as a trajectory-level measure for evaluating sample quality.
arXiv Detail & Related papers (2023-11-02T16:45:25Z) - Toward a Foundation Model for Time Series Data [34.1973242428317]
A foundation model is a machine learning model trained on a large and diverse set of data.
We develop an effective time series foundation model by leveraging unlabeled samples from multiple domains.
arXiv Detail & Related papers (2023-10-05T21:44:50Z) - Cross-Frequency Time Series Meta-Forecasting [7.809667883159047]
We introduce the Continuous Frequency Adapter (CFA), specifically designed to learn frequency-invariant representations.
CFA greatly improves performance when generalizing to unseen frequencies, providing a first step towards forecasting over larger multi-frequency datasets.
arXiv Detail & Related papers (2023-02-04T03:22:16Z) - Continual Test-Time Domain Adaptation [94.51284735268597]
Test-time domain adaptation aims to adapt a source pre-trained model to a target domain without using any source data.
CoTTA is easy to implement and can be readily incorporated in off-the-shelf pre-trained models.
arXiv Detail & Related papers (2022-03-25T11:42:02Z) - Adaptive Risk Minimization: Learning to Adapt to Domain Shift [109.87561509436016]
A fundamental assumption of most machine learning algorithms is that the training and test data are drawn from the same underlying distribution.
In this work, we consider the problem setting of domain generalization, where the training data are structured into domains and there may be multiple test time shifts.
We introduce the framework of adaptive risk minimization (ARM), in which models are directly optimized for effective adaptation to shift by learning to adapt on the training domains.
arXiv Detail & Related papers (2020-07-06T17:59:30Z) - Bridging the Gap Between Training and Inference for Spatio-Temporal
Forecasting [16.06369357595426]
We propose a novel curriculum learning based strategy named Temporal Progressive Growing Sampling to bridge the gap between training and inference for S-temporal sequence forecasting.
Experimental results demonstrate that our proposed method better models long term dependencies and outperforms baseline approaches on two competitive datasets.
arXiv Detail & Related papers (2020-05-19T10:14:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.