AAMDRL: Augmented Asset Management with Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2010.08497v1
- Date: Wed, 30 Sep 2020 03:55:47 GMT
- Title: AAMDRL: Augmented Asset Management with Deep Reinforcement Learning
- Authors: Eric Benhamou and David Saltiel and Sandrine Ungari and Abhishek
Mukhopadhyay and Jamal Atif
- Abstract summary: We show how Deep Reinforcement Learning can tackle this challenge.
Our contributions are threefold: (i) the use of contextual information also referred to as augmented state in DRL, (ii) the impact of a one period lag between observations and actions, and (iii) the implementation of a new repetitive train test method called walk forward analysis.
Although our experiment is on trading bots, it can easily be translated to other bot environments that operate in sequential environment with regime changes and noisy data.
- Score: 5.801876281373619
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Can an agent learn efficiently in a noisy and self adapting environment with
sequential, non-stationary and non-homogeneous observations? Through trading
bots, we illustrate how Deep Reinforcement Learning (DRL) can tackle this
challenge. Our contributions are threefold: (i) the use of contextual
information also referred to as augmented state in DRL, (ii) the impact of a
one period lag between observations and actions that is more realistic for an
asset management environment, (iii) the implementation of a new repetitive
train test method called walk forward analysis, similar in spirit to cross
validation for time series. Although our experiment is on trading bots, it can
easily be translated to other bot environments that operate in sequential
environment with regime changes and noisy data. Our experiment for an augmented
asset manager interested in finding the best portfolio for hedging strategies
shows that AAMDRL achieves superior returns and lower risk.
Related papers
- Discovering Minimal Reinforcement Learning Environments [24.6408931194983]
Reinforcement learning (RL) agents are commonly trained and evaluated in the same environment.
Humans often train in a specialized environment before being evaluated, such as studying a book before taking an exam.
arXiv Detail & Related papers (2024-06-18T13:19:26Z) - Contrastive Initial State Buffer for Reinforcement Learning [25.849626996870526]
In Reinforcement Learning, the trade-off between exploration and exploitation poses a complex challenge for achieving efficient learning from limited samples.
We introduce the concept of a Contrastive Initial State Buffer, which strategically selects states from past experiences and uses them to initialize the agent in the environment.
We validate our approach on two complex robotic tasks without relying on any prior information about the environment.
arXiv Detail & Related papers (2023-09-18T13:26:40Z) - Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels [112.63440666617494]
Reinforcement learning algorithms can succeed but require large amounts of interactions between the agent and the environment.
We propose a new method to solve it, using unsupervised model-based RL, for pre-training the agent.
We show robust performance on the Real-Word RL benchmark, hinting at resiliency to environment perturbations during adaptation.
arXiv Detail & Related papers (2022-09-24T14:22:29Z) - Retrieval-Augmented Reinforcement Learning [63.32076191982944]
We train a network to map a dataset of past experiences to optimal behavior.
The retrieval process is trained to retrieve information from the dataset that may be useful in the current context.
We show that retrieval-augmented R2D2 learns significantly faster than the baseline R2D2 agent and achieves higher scores.
arXiv Detail & Related papers (2022-02-17T02:44:05Z) - Multitask Adaptation by Retrospective Exploration with Learned World
Models [77.34726150561087]
We propose a meta-learned addressing model called RAMa that provides training samples for the MBRL agent taken from task-agnostic storage.
The model is trained to maximize the expected agent's performance by selecting promising trajectories solving prior tasks from the storage.
arXiv Detail & Related papers (2021-10-25T20:02:57Z) - Simplifying Deep Reinforcement Learning via Self-Supervision [51.2400839966489]
Self-Supervised Reinforcement Learning (SSRL) is a simple algorithm that optimize policies with purely supervised losses.
We show that SSRL is surprisingly competitive to contemporary algorithms with more stable performance and less running time.
arXiv Detail & Related papers (2021-06-10T06:29:59Z) - PEBBLE: Feedback-Efficient Interactive Reinforcement Learning via
Relabeling Experience and Unsupervised Pre-training [94.87393610927812]
We present an off-policy, interactive reinforcement learning algorithm that capitalizes on the strengths of both feedback and off-policy learning.
We demonstrate that our approach is capable of learning tasks of higher complexity than previously considered by human-in-the-loop methods.
arXiv Detail & Related papers (2021-06-09T14:10:50Z) - PsiPhi-Learning: Reinforcement Learning with Demonstrations using
Successor Features and Inverse Temporal Difference Learning [102.36450942613091]
We propose an inverse reinforcement learning algorithm, called emphinverse temporal difference learning (ITD)
We show how to seamlessly integrate ITD with learning from online environment interactions, arriving at a novel algorithm for reinforcement learning with demonstrations, called $Psi Phi$-learning.
arXiv Detail & Related papers (2021-02-24T21:12:09Z) - Deep Reinforcement Learning for Active High Frequency Trading [1.6874375111244329]
We introduce the first end-to-end Deep Reinforcement Learning (DRL) based framework for active high frequency trading in the stock market.
We train DRL agents to trade one unit of Intel Corporation stock by employing the Proximal Policy Optimization algorithm.
arXiv Detail & Related papers (2021-01-18T15:09:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.