Reinforcement Learning based dynamic weighing of Ensemble Models for
Time Series Forecasting
- URL: http://arxiv.org/abs/2008.08878v1
- Date: Thu, 20 Aug 2020 10:40:42 GMT
- Title: Reinforcement Learning based dynamic weighing of Ensemble Models for
Time Series Forecasting
- Authors: Satheesh K. Perepu, Bala Shyamala Balaji, Hemanth Kumar Tanneru,
Sudhakar Kathari, Vivek Shankar Pinnamaraju
- Abstract summary: It is known that if models selected for data modelling are distinct (linear/non-linear, static/dynamic) and independent (minimally correlated) models, the accuracy of the predictions is improved.
Various approaches suggested in the literature to weigh the ensemble models use a static set of weights.
To address this issue, a Reinforcement Learning (RL) approach to dynamically assign and update weights of each of the models at different time instants.
- Score: 0.8399688944263843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Ensemble models are powerful model building tools that are developed with a
focus to improve the accuracy of model predictions. They find applications in
time series forecasting in varied scenarios including but not limited to
process industries, health care, and economics where a single model might not
provide optimal performance. It is known that if models selected for data
modelling are distinct (linear/non-linear, static/dynamic) and independent
(minimally correlated models), the accuracy of the predictions is improved.
Various approaches suggested in the literature to weigh the ensemble models use
a static set of weights. Due to this limitation, approaches using a static set
of weights for weighing ensemble models cannot capture the dynamic changes or
local features of the data effectively. To address this issue, a Reinforcement
Learning (RL) approach to dynamically assign and update weights of each of the
models at different time instants depending on the nature of data and the
individual model predictions is proposed in this work. The RL method
implemented online, essentially learns to update the weights and reduce the
errors as the time progresses. Simulation studies on time series data showed
that the dynamic weighted approach using RL learns the weight better than
existing approaches. The accuracy of the proposed method is compared with an
existing approach of online Neural Network tuning quantitatively through
normalized mean square error(NMSE) values.
Related papers
- Learning Augmentation Policies from A Model Zoo for Time Series Forecasting [58.66211334969299]
We introduce AutoTSAug, a learnable data augmentation method based on reinforcement learning.
By augmenting the marginal samples with a learnable policy, AutoTSAug substantially improves forecasting performance.
arXiv Detail & Related papers (2024-09-10T07:34:19Z) - Few-Shot Load Forecasting Under Data Scarcity in Smart Grids: A Meta-Learning Approach [0.18641315013048293]
This paper proposes adapting an established model-agnostic meta-learning algorithm for short-term load forecasting.
The proposed method can rapidly adapt and generalize within any unknown load time series of arbitrary length.
The proposed model is evaluated using a dataset of historical load consumption data from real-world consumers.
arXiv Detail & Related papers (2024-06-09T18:59:08Z) - Towards Stable Machine Learning Model Retraining via Slowly Varying Sequences [6.067007470552307]
We propose a methodology for finding sequences of machine learning models that are stable across retraining iterations.
We develop a mixed-integer optimization formulation that is guaranteed to recover optimal models.
Our method shows stronger stability than greedily trained models with a small, controllable sacrifice in predictive power.
arXiv Detail & Related papers (2024-03-28T22:45:38Z) - Deep autoregressive density nets vs neural ensembles for model-based
offline reinforcement learning [2.9158689853305693]
We consider a model-based reinforcement learning algorithm that infers the system dynamics from the available data and performs policy optimization on imaginary model rollouts.
This approach is vulnerable to exploiting model errors which can lead to catastrophic failures on the real system.
We show that better performance can be obtained with a single well-calibrated autoregressive model on the D4RL benchmark.
arXiv Detail & Related papers (2024-02-05T10:18:15Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - Online Dynamics Learning for Predictive Control with an Application to
Aerial Robots [3.673994921516517]
Even though prediction models can be learned and applied to model-based controllers, these models are often learned offline.
In this offline setting, training data is first collected and a prediction model is learned through an elaborated training procedure.
We propose an online dynamics learning framework that continually improves the accuracy of the dynamic model during deployment.
arXiv Detail & Related papers (2022-07-19T15:51:25Z) - Autoregressive Dynamics Models for Offline Policy Evaluation and
Optimization [60.73540999409032]
We show that expressive autoregressive dynamics models generate different dimensions of the next state and reward sequentially conditioned on previous dimensions.
We also show that autoregressive dynamics models are useful for offline policy optimization by serving as a way to enrich the replay buffer.
arXiv Detail & Related papers (2021-04-28T16:48:44Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.