State Sequences Prediction via Fourier Transform for Representation
Learning
- URL: http://arxiv.org/abs/2310.15888v1
- Date: Tue, 24 Oct 2023 14:47:02 GMT
- Title: State Sequences Prediction via Fourier Transform for Representation
Learning
- Authors: Mingxuan Ye, Yufei Kuang, Jie Wang, Rui Yang, Wengang Zhou, Houqiang
Li, Feng Wu
- Abstract summary: We propose State Sequences Prediction via Fourier Transform (SPF), a novel method for learning expressive representations efficiently.
We theoretically analyze the existence of structural information in state sequences, which is closely related to policy performance and signal regularity.
Experiments demonstrate that the proposed method outperforms several state-of-the-art algorithms in terms of both sample efficiency and performance.
- Score: 111.82376793413746
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While deep reinforcement learning (RL) has been demonstrated effective in
solving complex control tasks, sample efficiency remains a key challenge due to
the large amounts of data required for remarkable performance. Existing
research explores the application of representation learning for data-efficient
RL, e.g., learning predictive representations by predicting long-term future
states. However, many existing methods do not fully exploit the structural
information inherent in sequential state signals, which can potentially improve
the quality of long-term decision-making but is difficult to discern in the
time domain. To tackle this problem, we propose State Sequences Prediction via
Fourier Transform (SPF), a novel method that exploits the frequency domain of
state sequences to extract the underlying patterns in time series data for
learning expressive representations efficiently. Specifically, we theoretically
analyze the existence of structural information in state sequences, which is
closely related to policy performance and signal regularity, and then propose
to predict the Fourier transform of infinite-step future state sequences to
extract such information. One of the appealing features of SPF is that it is
simple to implement while not requiring storage of infinite-step future states
as prediction targets. Experiments demonstrate that the proposed method
outperforms several state-of-the-art algorithms in terms of both sample
efficiency and performance.
Related papers
- Graph Spatiotemporal Process for Multivariate Time Series Anomaly
Detection with Missing Values [67.76168547245237]
We introduce a novel framework called GST-Pro, which utilizes a graphtemporal process and anomaly scorer to detect anomalies.
Our experimental results show that the GST-Pro method can effectively detect anomalies in time series data and outperforms state-of-the-art methods.
arXiv Detail & Related papers (2024-01-11T10:10:16Z) - Representation Learning of Multivariate Time Series using Attention and
Adversarial Training [2.0577627277681887]
A Transformer-based autoencoder is proposed that is regularized using an adversarial training scheme to generate artificial time series signals.
Our results indicate that the generated signals exhibit higher similarity to an exemplary dataset than using a convolutional network approach.
arXiv Detail & Related papers (2024-01-03T21:32:46Z) - Enhancing Algorithm Performance Understanding through tsMorph: Generating Semi-Synthetic Time Series for Robust Forecasting Evaluation [0.393259574660092]
We present tsMorph, a tool for generating semi-synthetic time series through dataset morphing.
In this paper, we show the benefits of tsMorph by assessing the predictive performance of the Long Short-Term Memory Network and DeepAR forecasting algorithms.
arXiv Detail & Related papers (2023-12-03T10:40:07Z) - MPR-Net:Multi-Scale Pattern Reproduction Guided Universality Time Series
Interpretable Forecasting [13.790498420659636]
Time series forecasting has received wide interest from existing research due to its broad applications inherent challenging.
This paper proposes a forecasting model, MPR-Net. It first adaptively decomposes multi-scale historical series patterns using convolution operation, then constructs a pattern extension forecasting method based on the prior knowledge of pattern reproduction, and finally reconstructs future patterns into future series using deconvolution operation.
By leveraging the temporal dependencies present in the time series, MPR-Net not only achieves linear time complexity, but also makes the forecasting process interpretable.
arXiv Detail & Related papers (2023-07-13T13:16:01Z) - On the Importance of Feature Decorrelation for Unsupervised
Representation Learning in Reinforcement Learning [23.876039876806182]
unsupervised representation learning (URL) has improved the sample efficiency of Reinforcement Learning (RL)
We propose a novel URL framework that causally predicts future states while increasing the dimension of the latent manifold.
Our framework effectively learns predictive representations without collapse, which significantly improves the sample efficiency of state-of-the-art URL methods on the Atari 100k benchmark.
arXiv Detail & Related papers (2023-06-09T02:47:21Z) - Task-Guided IRL in POMDPs that Scales [22.594913269327353]
In inverse linear reinforcement learning (IRL), a learning agent infers a reward function encoding the underlying task using demonstrations from experts.
Most IRL techniques require the computationally forward problem -- computing an optimal policy given a reward function -- in POMDPs.
We develop an algorithm that reduces the information while increasing the data efficiency.
arXiv Detail & Related papers (2022-12-30T21:08:57Z) - Value-Consistent Representation Learning for Data-Efficient
Reinforcement Learning [105.70602423944148]
We propose a novel method, called value-consistent representation learning (VCR), to learn representations that are directly related to decision-making.
Instead of aligning this imagined state with a real state returned by the environment, VCR applies a $Q$-value head on both states and obtains two distributions of action values.
It has been demonstrated that our methods achieve new state-of-the-art performance for search-free RL algorithms.
arXiv Detail & Related papers (2022-06-25T03:02:25Z) - Voice2Series: Reprogramming Acoustic Models for Time Series
Classification [65.94154001167608]
Voice2Series is a novel end-to-end approach that reprograms acoustic models for time series classification.
We show that V2S either outperforms or is tied with state-of-the-art methods on 20 tasks, and improves their average accuracy by 1.84%.
arXiv Detail & Related papers (2021-06-17T07:59:15Z) - Optimizing for the Future in Non-Stationary MDPs [52.373873622008944]
We present a policy gradient algorithm that maximizes a forecast of future performance.
We show that our algorithm, called Prognosticator, is more robust to non-stationarity than two online adaptation techniques.
arXiv Detail & Related papers (2020-05-17T03:41:19Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.