MrSQM: Fast Time Series Classification with Symbolic Representations
- URL: http://arxiv.org/abs/2109.01036v1
- Date: Thu, 2 Sep 2021 15:54:46 GMT
- Title: MrSQM: Fast Time Series Classification with Symbolic Representations
- Authors: Thach Le Nguyen and Georgiana Ifrim
- Abstract summary: MrSQM uses multiple symbolic representations and efficient sequence mining to extract important time series features.
We study four feature selection approaches on symbolic sequences, ranging from fully supervised, to unsupervised and hybrids.
Our experiments on 112 datasets of the UEA/UCR benchmark demonstrate that MrSQM can quickly extract useful features.
- Score: 11.853438514668207
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Symbolic representations of time series have proven to be effective for time
series classification, with many recent approaches including SAX-VSM, BOSS,
WEASEL, and MrSEQL. The key idea is to transform numerical time series to
symbolic representations in the time or frequency domain, i.e., sequences of
symbols, and then extract features from these sequences. While achieving high
accuracy, existing symbolic classifiers are computationally expensive. In this
paper we present MrSQM, a new time series classifier which uses multiple
symbolic representations and efficient sequence mining, to extract important
time series features. We study four feature selection approaches on symbolic
sequences, ranging from fully supervised, to unsupervised and hybrids. We
propose a new approach for optimal supervised symbolic feature selection in
all-subsequence space, by adapting a Chi-squared bound developed for
discriminative pattern mining, to time series. Our extensive experiments on 112
datasets of the UEA/UCR benchmark demonstrate that MrSQM can quickly extract
useful features and learn accurate classifiers with the classic logistic
regression algorithm. Interestingly, we find that a very simple and fast
feature selection strategy can be highly effective as compared with more
sophisticated and expensive methods. MrSQM advances the state-of-the-art for
symbolic time series classifiers and it is an effective method to achieve high
accuracy, with fast runtime.
Related papers
- TimeSiam: A Pre-Training Framework for Siamese Time-Series Modeling [67.02157180089573]
Time series pre-training has recently garnered wide attention for its potential to reduce labeling expenses and benefit various downstream tasks.
This paper proposes TimeSiam as a simple but effective self-supervised pre-training framework for Time series based on Siamese networks.
arXiv Detail & Related papers (2024-02-04T13:10:51Z) - RED CoMETS: An ensemble classifier for symbolically represented
multivariate time series [1.0878040851638]
This paper introduces a novel ensemble classifier called RED CoMETS.
Red CoMETS builds upon the success of Co-eye, an ensemble classifier specifically designed for symbolically represented univariate time series.
It achieves the highest reported accuracy in the literature for the 'HandMovementDirection' dataset.
arXiv Detail & Related papers (2023-07-25T17:36:34Z) - Quick Adaptive Ternary Segmentation: An Efficient Decoding Procedure For
Hidden Markov Models [70.26374282390401]
Decoding the original signal (i.e., hidden chain) from the noisy observations is one of the main goals in nearly all HMM based data analyses.
We present Quick Adaptive Ternary (QATS), a divide-and-conquer procedure which decodes the hidden sequence in polylogarithmic computational complexity.
arXiv Detail & Related papers (2023-05-29T19:37:48Z) - TimeMAE: Self-Supervised Representations of Time Series with Decoupled
Masked Autoencoders [55.00904795497786]
We propose TimeMAE, a novel self-supervised paradigm for learning transferrable time series representations based on transformer networks.
The TimeMAE learns enriched contextual representations of time series with a bidirectional encoding scheme.
To solve the discrepancy issue incurred by newly injected masked embeddings, we design a decoupled autoencoder architecture.
arXiv Detail & Related papers (2023-03-01T08:33:16Z) - ASTRIDE: Adaptive Symbolization for Time Series Databases [6.8820425565516095]
We introduce ASTRIDE, a novel symbolic representation of time series, along with its accelerated variant FASTRIDE (Fast ASTRIDE)
Unlike most symbolization procedures, ASTRIDE is adaptive during both the segmentation step by performing change-point detection and the quantization step by using quantiles.
We demonstrate the performance of the ASTRIDE and FASTRIDE representations compared to SAX, 1d-SAX, SFA, and ABBA on reconstruction and, when applicable, on classification tasks.
arXiv Detail & Related papers (2023-02-08T14:46:24Z) - HyperTime: Implicit Neural Representation for Time Series [131.57172578210256]
Implicit neural representations (INRs) have recently emerged as a powerful tool that provides an accurate and resolution-independent encoding of data.
In this paper, we analyze the representation of time series using INRs, comparing different activation functions in terms of reconstruction accuracy and training convergence speed.
We propose a hypernetwork architecture that leverages INRs to learn a compressed latent representation of an entire time series dataset.
arXiv Detail & Related papers (2022-08-11T14:05:51Z) - The FreshPRINCE: A Simple Transformation Based Pipeline Time Series
Classifier [0.0]
We look at whether the complexity of the algorithms considered state of the art is really necessary.
Many times the first approach suggested is a simple pipeline of summary statistics or other time series feature extraction approaches.
We test these approaches on the UCR time series dataset archive, looking to see if TSC literature has overlooked the effectiveness of these approaches.
arXiv Detail & Related papers (2022-01-28T11:23:58Z) - An efficient aggregation method for the symbolic representation of
temporal data [0.0]
We present a new variant of the adaptive Brownian bridge-based aggregation (ABBA) method, called fABBA.
This variant utilizes a new aggregation approach tailored to the piecewise representation of time series.
In contrast to the original method, the new approach does not require the number of time series symbols to be specified in advance.
arXiv Detail & Related papers (2022-01-14T22:51:24Z) - Novel Features for Time Series Analysis: A Complex Networks Approach [62.997667081978825]
Time series data are ubiquitous in several domains as climate, economics and health care.
Recent conceptual approach relies on time series mapping to complex networks.
Network analysis can be used to characterize different types of time series.
arXiv Detail & Related papers (2021-10-11T13:46:28Z) - Interpretable Time Series Classification using Linear Models and
Multi-resolution Multi-domain Symbolic Representations [6.6147550436077776]
We propose new time series classification algorithms to address gaps in current approaches.
Our approach is based on symbolic representations of time series, efficient sequence mining algorithms and linear classification models.
Our models are as accurate as deep learning models but are more efficient regarding running time and memory, can work with variable-length time series and can be interpreted by highlighting the discriminative symbolic features on the original time series.
arXiv Detail & Related papers (2020-05-31T15:32:08Z) - Supervised Feature Subset Selection and Feature Ranking for Multivariate
Time Series without Feature Extraction [78.84356269545157]
We introduce supervised feature ranking and feature subset selection algorithms for MTS classification.
Unlike most existing supervised/unsupervised feature selection algorithms for MTS our techniques do not require a feature extraction step to generate a one-dimensional feature vector from the time series.
arXiv Detail & Related papers (2020-05-01T07:46:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.