Taking ROCKET on an Efficiency Mission: Multivariate Time Series
Classification with LightWaveS
- URL: http://arxiv.org/abs/2204.01379v2
- Date: Tue, 5 Apr 2022 08:27:56 GMT
- Title: Taking ROCKET on an Efficiency Mission: Multivariate Time Series
Classification with LightWaveS
- Authors: Leonardos Pantiskas, Kees Verstoep, Mark Hoogendoorn, Henri Bal
- Abstract summary: We present LightWaveS, a framework for accurate multivariate time series classification.
It employs just 2.5% of the ROCKET features, while achieving accuracy comparable to recent deep learning models.
We show that we achieve speedup ranging from 9x to 65x compared to ROCKET during inference on an edge device.
- Score: 3.5786621294068373
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Nowadays, with the rising number of sensors in sectors such as healthcare and
industry, the problem of multivariate time series classification (MTSC) is
getting increasingly relevant and is a prime target for machine and deep
learning approaches. Their expanding adoption in real-world environments is
causing a shift in focus from the pursuit of ever higher prediction accuracy
with complex models towards practical, deployable solutions that balance
accuracy and parameters such as prediction speed. An MTSC model that has
attracted attention recently is ROCKET, based on random convolutional kernels,
both because of its very fast training process and its state-of-the-art
accuracy. However, the large number of features it utilizes may be detrimental
to inference time. Examining its theoretical background and limitations enables
us to address potential drawbacks and present LightWaveS: a framework for
accurate MTSC, which is fast both during training and inference. Specifically,
utilizing wavelet scattering transformation and distributed feature selection,
we manage to create a solution which employs just 2.5% of the ROCKET features,
while achieving accuracy comparable to recent deep learning models. LightWaveS
also scales well across multiple compute nodes and with the number of input
channels during training. In addition, it can significantly reduce the input
size and provide insight to an MTSC problem by keeping only the most useful
channels. We present three versions of our algorithm and their results on
distributed training time and scalability, accuracy and inference speedup. We
show that we achieve speedup ranging from 9x to 65x compared to ROCKET during
inference on an edge device, on datasets with comparable accuracy.
Related papers
- Nearly Lossless Adaptive Bit Switching [8.485009775430411]
Experimental results on the ImageNet-1K classification demonstrate that our methods have enough advantages to state-of-the-art one-shot joint QAT in both multi-precision and mixed-precision.
arXiv Detail & Related papers (2025-02-03T09:46:26Z) - Classification of Raw MEG/EEG Data with Detach-Rocket Ensemble: An Improved ROCKET Algorithm for Multivariate Time Series Analysis [0.0]
We present a novel ROCKET-based algorithm, named Detach-Rocket Ensemble, specifically designed to deal with high-dimensional data such as EEG and MEG.
Our algorithm leverages pruning to provide an integrated estimation of channel importance, and ensembles to achieve better accuracy and provide a label probability.
We show that Detach-Rocket Ensemble is able to provide both interpretable channel relevance and competitive classification accuracy, even when applied directly to the raw brain data.
arXiv Detail & Related papers (2024-08-05T18:24:09Z) - DOCTOR: Dynamic On-Chip Temporal Variation Remediation Toward Self-Corrected Photonic Tensor Accelerators [5.873308516576125]
Photonic tensor accelerators offer unparalleled speed and energy efficiency.
Off-chip noise-aware training and on-chip training have been proposed to enhance the variation tolerance of optical neural accelerators.
We propose a lightweight dynamic on-chip framework, dubbed DOCTOR, providing adaptive, in-situ accuracy recovery against temporally drifting noise.
arXiv Detail & Related papers (2024-03-05T06:17:13Z) - Efficient Asynchronous Federated Learning with Sparsification and
Quantization [55.6801207905772]
Federated Learning (FL) is attracting more and more attention to collaboratively train a machine learning model without transferring raw data.
FL generally exploits a parameter server and a large number of edge devices during the whole process of the model training.
We propose TEASQ-Fed to exploit edge devices to asynchronously participate in the training process by actively applying for tasks.
arXiv Detail & Related papers (2023-12-23T07:47:07Z) - Detach-ROCKET: Sequential feature selection for time series classification with random convolutional kernels [0.7499722271664144]
We introduce Sequential Feature Detachment (SFD) to identify and prune non-essential features in ROCKET-based models.
SFD can produce models with better test accuracy using only 10% of the original features.
We also present an end-to-end procedure for determining an optimal balance between the number of features and model accuracy.
arXiv Detail & Related papers (2023-09-25T20:24:36Z) - Efficient Graph Neural Network Inference at Large Scale [54.89457550773165]
Graph neural networks (GNNs) have demonstrated excellent performance in a wide range of applications.
Existing scalable GNNs leverage linear propagation to preprocess the features and accelerate the training and inference procedure.
We propose a novel adaptive propagation order approach that generates the personalized propagation order for each node based on its topological information.
arXiv Detail & Related papers (2022-11-01T14:38:18Z) - Correlating sparse sensing for large-scale traffic speed estimation: A
Laplacian-enhanced low-rank tensor kriging approach [76.45949280328838]
We propose a Laplacian enhanced low-rank tensor (LETC) framework featuring both lowrankness and multi-temporal correlations for large-scale traffic speed kriging.
We then design an efficient solution algorithm via several effective numeric techniques to scale up the proposed model to network-wide kriging.
arXiv Detail & Related papers (2022-10-21T07:25:57Z) - Adaptive Latent Space Tuning for Non-Stationary Distributions [62.997667081978825]
We present a method for adaptive tuning of the low-dimensional latent space of deep encoder-decoder style CNNs.
We demonstrate our approach for predicting the properties of a time-varying charged particle beam in a particle accelerator.
arXiv Detail & Related papers (2021-05-08T03:50:45Z) - FastFlowNet: A Lightweight Network for Fast Optical Flow Estimation [81.76975488010213]
Dense optical flow estimation plays a key role in many robotic vision tasks.
Current networks often occupy large number of parameters and require heavy computation costs.
Our proposed FastFlowNet works in the well-known coarse-to-fine manner with following innovations.
arXiv Detail & Related papers (2021-03-08T03:09:37Z) - Fast Modeling and Understanding Fluid Dynamics Systems with
Encoder-Decoder Networks [0.0]
We show that an accurate deep-learning-based proxy model can be taught efficiently by a finite-volume-based simulator.
Compared to traditional simulation, the proposed deep learning approach enables much faster forward computation.
We quantify the sensitivity of the deep learning model to key physical parameters and hence demonstrate that the inversion problems can be solved with great acceleration.
arXiv Detail & Related papers (2020-06-09T17:14:08Z) - FMA-ETA: Estimating Travel Time Entirely Based on FFN With Attention [88.33372574562824]
We propose a novel framework based on feed-forward network (FFN) for ETA, FFN with Multi-factor self-Attention (FMA-ETA)
The novel Multi-factor self-attention mechanism is proposed to deal with different category features and aggregate the information purposefully.
Experiments show FMA-ETA is competitive with state-of-the-art methods in terms of the prediction accuracy with significantly better inference speed.
arXiv Detail & Related papers (2020-06-07T08:10:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.