Error-bounded Approximate Time Series Joins Using Compact Dictionary
Representations of Time Series
- URL: http://arxiv.org/abs/2112.12965v2
- Date: Sun, 5 Nov 2023 04:34:23 GMT
- Title: Error-bounded Approximate Time Series Joins Using Compact Dictionary
Representations of Time Series
- Authors: Chin-Chia Michael Yeh, Yan Zheng, Junpeng Wang, Huiyuan Chen,
Zhongfang Zhuang, Wei Zhang, Eamonn Keogh
- Abstract summary: We show that it is possible to efficiently perform inter-time series similarity joins with error bounded guarantees by creating a compact "dictionary" representation of time series.
We demonstrate the utility of our dictionary-based inter-time series similarity joins on domains as diverse as medicine and transportation.
- Score: 29.83535690719436
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The matrix profile is an effective data mining tool that provides similarity
join functionality for time series data. Users of the matrix profile can either
join a time series with itself using intra-similarity join (i.e., self-join) or
join a time series with another time series using inter-similarity join. By
invoking either or both types of joins, the matrix profile can help users
discover both conserved and anomalous structures in the data. Since the
introduction of the matrix profile five years ago, multiple efforts have been
made to speed up the computation with approximate joins; however, the majority
of these efforts only focus on self-joins. In this work, we show that it is
possible to efficiently perform approximate inter-time series similarity joins
with error bounded guarantees by creating a compact "dictionary" representation
of time series. Using the dictionary representation instead of the original
time series, we are able to improve the throughput of an anomaly mining system
by at least 20X, with essentially no decrease in accuracy. As a side effect,
the dictionaries also summarize the time series in a semantically meaningful
way and can provide intuitive and actionable insights. We demonstrate the
utility of our dictionary-based inter-time series similarity joins on domains
as diverse as medicine and transportation.
Related papers
- TiVy: Time Series Visual Summary for Scalable Visualization [32.33793043326047]
We propose TiVy, a new algorithm that summarizes time series using sequential patterns.<n>We also present an interactive time series visualization that renders large-scale time series in real-time.
arXiv Detail & Related papers (2025-07-25T05:50:01Z) - Inferring the Most Similar Variable-length Subsequences between Multidimensional Time Series [0.0]
We propose an algorithm that provides the exact solution of finding the most similar multidimensional subsequences between time series.<n>The algorithm is built based on theoretical guarantee of correctness and efficiency.<n>In real-world datasets, it extracted the most similar subsequences even faster.
arXiv Detail & Related papers (2025-05-16T10:39:46Z) - Language in the Flow of Time: Time-Series-Paired Texts Weaved into a Unified Temporal Narrative [65.84249211767921]
Texts as Time Series (TaTS) can be plugged into any existing numerical-only time series models.<n>We show that TaTS can enhance predictive performance without modifying model architectures.
arXiv Detail & Related papers (2025-02-13T03:43:27Z) - TimeSiam: A Pre-Training Framework for Siamese Time-Series Modeling [67.02157180089573]
Time series pre-training has recently garnered wide attention for its potential to reduce labeling expenses and benefit various downstream tasks.
This paper proposes TimeSiam as a simple but effective self-supervised pre-training framework for Time series based on Siamese networks.
arXiv Detail & Related papers (2024-02-04T13:10:51Z) - Series2Vec: Similarity-based Self-supervised Representation Learning for
Time Series Classification [13.775977945756415]
We introduce a novel approach called textitSeries2Vec for self-supervised representation learning.
Series2Vec is trained to predict the similarity between two series in both temporal and spectral domains.
We show that Series2Vec performs comparably with fully supervised training and offers high efficiency in datasets with limited-labeled data.
arXiv Detail & Related papers (2023-12-07T02:30:40Z) - Time Series Synthesis Using the Matrix Profile for Anonymization [32.22243483781984]
Many researchers cannot release their data due to privacy regulations or fear of leaking confidential business information.
We propose the Time Series Synthesis Using the Matrix Profile (TSSUMP) method, where synthesized time series can be released in lieu of the original data.
We test our method on a case study of ECG and gender masking prediction.
arXiv Detail & Related papers (2023-11-05T04:27:24Z) - Robust Detection of Lead-Lag Relationships in Lagged Multi-Factor Models [61.10851158749843]
Key insights can be obtained by discovering lead-lag relationships inherent in the data.
We develop a clustering-driven methodology for robust detection of lead-lag relationships in lagged multi-factor models.
arXiv Detail & Related papers (2023-05-11T10:30:35Z) - TimeMAE: Self-Supervised Representations of Time Series with Decoupled
Masked Autoencoders [55.00904795497786]
We propose TimeMAE, a novel self-supervised paradigm for learning transferrable time series representations based on transformer networks.
The TimeMAE learns enriched contextual representations of time series with a bidirectional encoding scheme.
To solve the discrepancy issue incurred by newly injected masked embeddings, we design a decoupled autoencoder architecture.
arXiv Detail & Related papers (2023-03-01T08:33:16Z) - Matrix Profile XXVII: A Novel Distance Measure for Comparing Long Time
Series [18.205595410817327]
We introduce PRCIS, which stands for Pattern Representation Comparison in Series.
PRCIS is a distance measure for long time series, which exploits recent progress in our ability to summarize time series with dictionaries.
arXiv Detail & Related papers (2022-12-09T23:02:23Z) - HyperTime: Implicit Neural Representation for Time Series [131.57172578210256]
Implicit neural representations (INRs) have recently emerged as a powerful tool that provides an accurate and resolution-independent encoding of data.
In this paper, we analyze the representation of time series using INRs, comparing different activation functions in terms of reconstruction accuracy and training convergence speed.
We propose a hypernetwork architecture that leverages INRs to learn a compressed latent representation of an entire time series dataset.
arXiv Detail & Related papers (2022-08-11T14:05:51Z) - Elastic Product Quantization for Time Series [19.839572576189187]
We propose the use of product quantization for efficient similarity-based comparison of time series under time warping.
The proposed solution emerges as a highly efficient (both in terms of memory usage and time) replacement for elastic measures in time series applications.
arXiv Detail & Related papers (2022-01-04T09:23:06Z) - Cluster-and-Conquer: A Framework For Time-Series Forecasting [94.63501563413725]
We propose a three-stage framework for forecasting high-dimensional time-series data.
Our framework is highly general, allowing for any time-series forecasting and clustering method to be used in each step.
When instantiated with simple linear autoregressive models, we are able to achieve state-of-the-art results on several benchmark datasets.
arXiv Detail & Related papers (2021-10-26T20:41:19Z) - Novel Features for Time Series Analysis: A Complex Networks Approach [62.997667081978825]
Time series data are ubiquitous in several domains as climate, economics and health care.
Recent conceptual approach relies on time series mapping to complex networks.
Network analysis can be used to characterize different types of time series.
arXiv Detail & Related papers (2021-10-11T13:46:28Z) - Self-Supervised Time Series Representation Learning by Inter-Intra
Relational Reasoning [18.72937677485634]
We present SelfTime: a general self-supervised time series representation learning framework.
We explore the inter-sample relation and intra-temporal relation of time series to learn the underlying structure feature on the unlabeled time series.
The useful representations of time series are extracted from the backbone under the supervision of relation reasoning heads.
arXiv Detail & Related papers (2020-11-27T04:04:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.