Enhancing Multivariate Time Series Classifiers through Self-Attention
and Relative Positioning Infusion
- URL: http://arxiv.org/abs/2302.06683v1
- Date: Mon, 13 Feb 2023 20:50:34 GMT
- Title: Enhancing Multivariate Time Series Classifiers through Self-Attention
and Relative Positioning Infusion
- Authors: Mehryar Abbasi, Parvaneh Saeedi
- Abstract summary: Time Series Classification (TSC) is an important and challenging task for many visual computing applications.
We propose two novel attention blocks that can enhance deep learning-based TSC approaches.
We show that adding the proposed attention blocks improves base models' average accuracy by up to 3.6%.
- Score: 4.18804572788063
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Time Series Classification (TSC) is an important and challenging task for
many visual computing applications. Despite the extensive range of methods
developed for TSC, relatively few utilized Deep Neural Networks (DNNs). In this
paper, we propose two novel attention blocks (Global Temporal Attention and
Temporal Pseudo-Gaussian augmented Self-Attention) that can enhance deep
learning-based TSC approaches, even when such approaches are designed and
optimized for a specific dataset or task. We validate this claim by evaluating
multiple state-of-the-art deep learning-based TSC models on the University of
East Anglia (UEA) benchmark, a standardized collection of 30 Multivariate Time
Series Classification (MTSC) datasets. We show that adding the proposed
attention blocks improves base models' average accuracy by up to 3.6%.
Additionally, the proposed TPS block uses a new injection module to include the
relative positional information in transformers. As a standalone unit with less
computational complexity, it enables TPS to perform better than most of the
state-of-the-art DNN-based TSC methods. The source codes for our experimental
setups and proposed attention blocks are made publicly available.
Related papers
- DiTMoS: Delving into Diverse Tiny-Model Selection on Microcontrollers [34.282971510732736]
We introduce DiTMoS, a novel DNN training and inference framework with a selector-classifiers architecture.
A composition of weak models can exhibit high diversity and the union of them can significantly boost the accuracy upper bound.
We deploy DiTMoS on the Neucleo STM32F767ZI board and evaluate it based on three time-series datasets for human activity recognition, keywords spotting, and emotion recognition.
arXiv Detail & Related papers (2024-03-14T02:11:38Z) - Rethinking Few-shot 3D Point Cloud Semantic Segmentation [62.80639841429669]
This paper revisits few-shot 3D point cloud semantic segmentation (FS-PCS)
We focus on two significant issues in the state-of-the-art: foreground leakage and sparse point distribution.
To address these issues, we introduce a standardized FS-PCS setting, upon which a new benchmark is built.
arXiv Detail & Related papers (2024-03-01T15:14:47Z) - Finding Foundation Models for Time Series Classification with a PreText
Task [7.197233473373693]
This paper introduces pre-trained domain foundation models for Time Series Classification.
A key aspect of our methodology is a novel pretext task that spans multiple datasets.
Our experiments on the UCR archive demonstrate that this pre-training strategy significantly outperforms the conventional training approach without pre-training.
arXiv Detail & Related papers (2023-11-24T15:03:55Z) - FormerTime: Hierarchical Multi-Scale Representations for Multivariate
Time Series Classification [53.55504611255664]
FormerTime is a hierarchical representation model for improving the classification capacity for the multivariate time series classification task.
It exhibits three aspects of merits: (1) learning hierarchical multi-scale representations from time series data, (2) inheriting the strength of both transformers and convolutional networks, and (3) tacking the efficiency challenges incurred by the self-attention mechanism.
arXiv Detail & Related papers (2023-02-20T07:46:14Z) - Dynamic Sparse Network for Time Series Classification: Learning What to
"see'' [17.705879794117912]
We propose a dynamic sparse network (DSN) with sparse connections for time series classification (TSC)
The kernels in each sparse layer are sparse and can be explored under the constraint regions by dynamic sparse training, which makes it possible to reduce the resource cost.
arXiv Detail & Related papers (2022-12-19T20:32:27Z) - Enhancing Transformer Efficiency for Multivariate Time Series
Classification [12.128991867050487]
We propose a methodology to investigate the relationship between model efficiency and accuracy, as well as its complexity.
Comprehensive experiments on benchmark MTS datasets illustrate the effectiveness of our method.
arXiv Detail & Related papers (2022-03-28T03:25:19Z) - Coarse-to-Fine Sparse Transformer for Hyperspectral Image Reconstruction [138.04956118993934]
We propose a novel Transformer-based method, coarse-to-fine sparse Transformer (CST)
CST embedding HSI sparsity into deep learning for HSI reconstruction.
In particular, CST uses our proposed spectra-aware screening mechanism (SASM) for coarse patch selecting. Then the selected patches are fed into our customized spectra-aggregation hashing multi-head self-attention (SAH-MSA) for fine pixel clustering and self-similarity capturing.
arXiv Detail & Related papers (2022-03-09T16:17:47Z) - Towards Similarity-Aware Time-Series Classification [51.2400839966489]
We study time-series classification (TSC), a fundamental task of time-series data mining.
We propose Similarity-Aware Time-Series Classification (SimTSC), a framework that models similarity information with graph neural networks (GNNs)
arXiv Detail & Related papers (2022-01-05T02:14:57Z) - Adaptive Anomaly Detection for Internet of Things in Hierarchical Edge
Computing: A Contextual-Bandit Approach [81.5261621619557]
We propose an adaptive anomaly detection scheme with hierarchical edge computing (HEC)
We first construct multiple anomaly detection DNN models with increasing complexity, and associate each of them to a corresponding HEC layer.
Then, we design an adaptive model selection scheme that is formulated as a contextual-bandit problem and solved by using a reinforcement learning policy network.
arXiv Detail & Related papers (2021-08-09T08:45:47Z) - Classification of multivariate weakly-labelled time-series with
attention [0.0]
Weakly labelled time-series are time-series containing noise and significant redundancies.
This paper proposes an approach of exploiting context relevance of subsequences to improve classification accuracy.
arXiv Detail & Related papers (2021-02-16T16:05:38Z) - Benchmarking Multivariate Time Series Classification Algorithms [69.12151492736524]
Time Series Classification (TSC) involved building predictive models for a discrete target variable from ordered, real valued, attributes.
Over recent years, a new set of TSC algorithms have been developed which have made significant improvement over the previous state of the art.
We review recently proposed bespoke MTSC algorithms based on deep learning, shapelets and bag of words approaches.
arXiv Detail & Related papers (2020-07-26T15:56:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.