Time your hedge with Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2009.14136v2
- Date: Mon, 9 Nov 2020 07:56:27 GMT
- Title: Time your hedge with Deep Reinforcement Learning
- Authors: Eric Benhamou, David Saltiel, Sandrine Ungari, Abhishek Mukhopadhyay
- Abstract summary: Deep Reinforcement Learning (DRL) can tackle this challenge by creating a dynamic dependency between market information and hedging strategies allocation decisions.
We present a realistic and augmented DRL framework that: (i) uses additional contextual information to decide an action, (ii) has a one period lag between observations and actions to account for one day lag turnover of common asset managers to rebalance their hedge, (iii) is fully tested in terms of stability and robustness thanks to a repetitive train test method called anchored walk forward training, similar in spirit to k fold cross validation for time series and (iv) allows managing leverage of our hedging
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Can an asset manager plan the optimal timing for her/his hedging strategies
given market conditions? The standard approach based on Markowitz or other more
or less sophisticated financial rules aims to find the best portfolio
allocation thanks to forecasted expected returns and risk but fails to fully
relate market conditions to hedging strategies decision. In contrast, Deep
Reinforcement Learning (DRL) can tackle this challenge by creating a dynamic
dependency between market information and hedging strategies allocation
decisions. In this paper, we present a realistic and augmented DRL framework
that: (i) uses additional contextual information to decide an action, (ii) has
a one period lag between observations and actions to account for one day lag
turnover of common asset managers to rebalance their hedge, (iii) is fully
tested in terms of stability and robustness thanks to a repetitive train test
method called anchored walk forward training, similar in spirit to k fold cross
validation for time series and (iv) allows managing leverage of our hedging
strategy. Our experiment for an augmented asset manager interested in sizing
and timing his hedges shows that our approach achieves superior returns and
lower risk.
Related papers
- From Bandits Model to Deep Deterministic Policy Gradient, Reinforcement
Learning with Contextual Information [4.42532447134568]
In this study, we use two methods to overcome the issue with contextual information.
In order to investigate strategic trading in quantitative markets, we merged the earlier financial trading strategy known as constant proportion portfolio insurance ( CPPI) into deep deterministic policy gradient (DDPG)
The experimental results show that both methods can accelerate the progress of reinforcement learning to obtain the optimal solution.
arXiv Detail & Related papers (2023-10-01T11:25:20Z) - Robust Risk-Aware Option Hedging [2.405471533561618]
We showcase the potential of robust risk-aware reinforcement learning (RL) in mitigating the risks associated with path-dependent financial derivatives.
We apply this methodology to the hedging of barrier options, and highlight how the optimal hedging strategy undergoes distortions as the agent moves from being risk-averse to risk-seeking.
arXiv Detail & Related papers (2023-03-27T13:57:13Z) - Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels [112.63440666617494]
Reinforcement learning algorithms can succeed but require large amounts of interactions between the agent and the environment.
We propose a new method to solve it, using unsupervised model-based RL, for pre-training the agent.
We show robust performance on the Real-Word RL benchmark, hinting at resiliency to environment perturbations during adaptation.
arXiv Detail & Related papers (2022-09-24T14:22:29Z) - MetaTrader: An Reinforcement Learning Approach Integrating Diverse
Policies for Portfolio Optimization [17.759687104376855]
We propose a novel two-stage-based approach for portfolio management.
In the first stage, incorporates an imitation learning into the reinforcement learning framework.
In the second stage, learns a meta-policy to recognize the market conditions and decide on the most proper learned policy to follow.
arXiv Detail & Related papers (2022-09-01T07:58:06Z) - Reinforcement learning for options on target volatility funds [0.0]
We deal with the funding costs rising from hedging the risky securities underlying a target volatility strategy (TVS)
We derive an analytical solution of the problem in the Black and Scholes (BS) scenario.
Then we use Reinforcement Learning (RL) techniques to determine the fund composition leading to the most conservative price under the local volatility (LV) model.
arXiv Detail & Related papers (2021-12-03T10:55:11Z) - Deep Learning Statistical Arbitrage [0.0]
We propose a unifying conceptual framework for statistical arbitrage and develop a novel deep learning solution.
We construct arbitrage portfolios of similar assets as residual portfolios from conditional latent asset pricing factors.
We extract the time series signals of these residual portfolios with one of the most powerful machine learning time-series solutions.
arXiv Detail & Related papers (2021-06-08T00:48:25Z) - Universal Trading for Order Execution with Oracle Policy Distillation [99.57416828489568]
We propose a novel universal trading policy optimization framework to bridge the gap between the noisy yet imperfect market states and the optimal action sequences for order execution.
We show that our framework can better guide the learning of the common policy towards practically optimal execution by an oracle teacher with perfect information.
arXiv Detail & Related papers (2021-01-28T05:52:18Z) - Robust Reinforcement Learning on State Observations with Learned Optimal
Adversary [86.0846119254031]
We study the robustness of reinforcement learning with adversarially perturbed state observations.
With a fixed agent policy, we demonstrate that an optimal adversary to perturb state observations can be found.
For DRL settings, this leads to a novel empirical adversarial attack to RL agents via a learned adversary that is much stronger than previous ones.
arXiv Detail & Related papers (2021-01-21T05:38:52Z) - Learning Strategies in Decentralized Matching Markets under Uncertain
Preferences [91.3755431537592]
We study the problem of decision-making in the setting of a scarcity of shared resources when the preferences of agents are unknown a priori.
Our approach is based on the representation of preferences in a reproducing kernel Hilbert space.
We derive optimal strategies that maximize agents' expected payoffs.
arXiv Detail & Related papers (2020-10-29T03:08:22Z) - A Deep Reinforcement Learning Framework for Continuous Intraday Market
Bidding [69.37299910149981]
A key component for the successful renewable energy sources integration is the usage of energy storage.
We propose a novel modelling framework for the strategic participation of energy storage in the European continuous intraday market.
An distributed version of the fitted Q algorithm is chosen for solving this problem due to its sample efficiency.
Results indicate that the agent converges to a policy that achieves in average higher total revenues than the benchmark strategy.
arXiv Detail & Related papers (2020-04-13T13:50:13Z) - Reinforcement-Learning based Portfolio Management with Augmented Asset
Movement Prediction States [71.54651874063865]
Portfolio management (PM) aims to achieve investment goals such as maximal profits or minimal risks.
In this paper, we propose SARL, a novel State-Augmented RL framework for PM.
Our framework aims to address two unique challenges in financial PM: (1) data Heterogeneous data -- the collected information for each asset is usually diverse, noisy and imbalanced (e.g., news articles); and (2) environment uncertainty -- the financial market is versatile and non-stationary.
arXiv Detail & Related papers (2020-02-09T08:10:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.