TSFEDL: A Python Library for Time Series Spatio-Temporal Feature
Extraction and Prediction using Deep Learning (with Appendices on Detailed
Network Architectures and Experimental Cases of Study)
- URL: http://arxiv.org/abs/2206.03179v2
- Date: Wed, 8 Jun 2022 09:49:38 GMT
- Title: TSFEDL: A Python Library for Time Series Spatio-Temporal Feature
Extraction and Prediction using Deep Learning (with Appendices on Detailed
Network Architectures and Experimental Cases of Study)
- Authors: Ignacio Aguilera-Martos, \'Angel M. Garc\'ia-Vico, Juli\'an Luengo,
Sergio Damas, Francisco J. Melero, Jos\'e Javier Valle-Alonso, Francisco
Herrera
- Abstract summary: The TSFE library is built upon a set offlow+Keras and PyTorch modules under the AGPLv3 license.
The performance validation of the architectures included in this proposal confirms the usefulness of this Python package.
- Score: 9.445070013080601
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The combination of convolutional and recurrent neural networks is a promising
framework that allows the extraction of high-quality spatio-temporal features
together with its temporal dependencies, which is key for time series
prediction problems such as forecasting, classification or anomaly detection,
amongst others. In this paper, the TSFEDL library is introduced. It compiles 20
state-of-the-art methods for both time series feature extraction and
prediction, employing convolutional and recurrent deep neural networks for its
use in several data mining tasks. The library is built upon a set of
Tensorflow+Keras and PyTorch modules under the AGPLv3 license. The performance
validation of the architectures included in this proposal confirms the
usefulness of this Python package.
Related papers
- FocusLearn: Fully-Interpretable, High-Performance Modular Neural Networks for Time Series [0.3277163122167434]
This paper proposes a novel modular neural network model for time series prediction that is interpretable by construction.
A recurrent neural network learns the temporal dependencies in the data while an attention-based feature selection component selects the most relevant features.
A modular deep network is trained from the selected features independently to show the users how features influence outcomes, making the model interpretable.
arXiv Detail & Related papers (2023-11-28T14:51:06Z) - PyPOTS: A Python Toolbox for Data Mining on Partially-Observed Time
Series [0.0]
PyPOTS is an open-source Python library dedicated to data mining and analysis on partially-observed time series.
It provides easy access to diverse algorithms categorized into four tasks: imputation, classification, clustering, and forecasting.
arXiv Detail & Related papers (2023-05-30T07:57:05Z) - PARTIME: Scalable and Parallel Processing Over Time with Deep Neural
Networks [68.96484488899901]
We present PARTIME, a library designed to speed up neural networks whenever data is continuously streamed over time.
PARTIME starts processing each data sample at the time in which it becomes available from the stream.
Experiments are performed in order to empirically compare PARTIME with classic non-parallel neural computations in online learning.
arXiv Detail & Related papers (2022-10-17T14:49:14Z) - Feature-Based Time-Series Analysis in R using the theft Package [0.0]
Many open-source software packages for computing sets of time-series features exist across multiple programming languages.
Here we introduce a solution to these issues in an R software package called theft.
arXiv Detail & Related papers (2022-08-12T07:29:29Z) - HyperTime: Implicit Neural Representation for Time Series [131.57172578210256]
Implicit neural representations (INRs) have recently emerged as a powerful tool that provides an accurate and resolution-independent encoding of data.
In this paper, we analyze the representation of time series using INRs, comparing different activation functions in terms of reconstruction accuracy and training convergence speed.
We propose a hypernetwork architecture that leverages INRs to learn a compressed latent representation of an entire time series dataset.
arXiv Detail & Related papers (2022-08-11T14:05:51Z) - TorchNTK: A Library for Calculation of Neural Tangent Kernels of PyTorch
Models [16.30276204466139]
We introduce torchNTK, a python library to calculate the empirical neural tangent kernel (NTK) of neural network models in the PyTorch framework.
A feature of the library is that we expose the user to layerwise NTK components, and show that in some regimes a layerwise calculation is more memory efficient.
arXiv Detail & Related papers (2022-05-24T21:27:58Z) - PyTorch Geometric Temporal: Spatiotemporal Signal Processing with Neural
Machine Learning Models [8.572409162523735]
PyTorch Temporal is a temporal deep learning framework for neural signal processing.
It was created with foundations on existing libraries in the PyTorch eco-system.
Experiments show that the framework can potentially operate on web-scale datasets with rich temporal features and spatial structure.
arXiv Detail & Related papers (2021-04-15T21:45:57Z) - Meta-Learning for Koopman Spectral Analysis with Short Time-series [49.41640137945938]
Existing methods require long time-series for training neural networks.
We propose a meta-learning method for estimating embedding functions from unseen short time-series.
We experimentally demonstrate that the proposed method achieves better performance in terms of eigenvalue estimation and future prediction.
arXiv Detail & Related papers (2021-02-09T07:19:19Z) - Deep Cellular Recurrent Network for Efficient Analysis of Time-Series
Data with Spatial Information [52.635997570873194]
This work proposes a novel deep cellular recurrent neural network (DCRNN) architecture to process complex multi-dimensional time series data with spatial information.
The proposed architecture achieves state-of-the-art performance while utilizing substantially less trainable parameters when compared to comparable methods in the literature.
arXiv Detail & Related papers (2021-01-12T20:08:18Z) - Captum: A unified and generic model interpretability library for PyTorch [49.72749684393332]
We introduce a novel, unified, open-source model interpretability library for PyTorch.
The library contains generic implementations of a number of gradient and perturbation-based attribution algorithms.
It can be used for both classification and non-classification models.
arXiv Detail & Related papers (2020-09-16T18:57:57Z) - Convolutional Tensor-Train LSTM for Spatio-temporal Learning [116.24172387469994]
We propose a higher-order LSTM model that can efficiently learn long-term correlations in the video sequence.
This is accomplished through a novel tensor train module that performs prediction by combining convolutional features across time.
Our results achieve state-of-the-art performance-art in a wide range of applications and datasets.
arXiv Detail & Related papers (2020-02-21T05:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.