Asynchronous Deep Double Duelling Q-Learning for Trading-Signal
Execution in Limit Order Book Markets
- URL: http://arxiv.org/abs/2301.08688v2
- Date: Mon, 25 Sep 2023 15:57:24 GMT
- Title: Asynchronous Deep Double Duelling Q-Learning for Trading-Signal
Execution in Limit Order Book Markets
- Authors: Peer Nagy, Jan-Peter Calliess and Stefan Zohren
- Abstract summary: We employ deep reinforcement learning to train an agent to translate a high-frequency trading signal into a trading strategy that places individual limit orders.
Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment.
We find that the RL agent learns an effective trading strategy for inventory management and order placing that outperforms a benchmark trading strategy having access to the same signal.
- Score: 5.202524136984542
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We employ deep reinforcement learning (RL) to train an agent to successfully
translate a high-frequency trading signal into a trading strategy that places
individual limit orders. Based on the ABIDES limit order book simulator, we
build a reinforcement learning OpenAI gym environment and utilise it to
simulate a realistic trading environment for NASDAQ equities based on historic
order book messages. To train a trading agent that learns to maximise its
trading return in this environment, we use Deep Duelling Double Q-learning with
the APEX (asynchronous prioritised experience replay) architecture. The agent
observes the current limit order book state, its recent history, and a
short-term directional forecast. To investigate the performance of RL for
adaptive trading independently from a concrete forecasting algorithm, we study
the performance of our approach utilising synthetic alpha signals obtained by
perturbing forward-looking returns with varying levels of noise. Here, we find
that the RL agent learns an effective trading strategy for inventory management
and order placing that outperforms a heuristic benchmark trading strategy
having access to the same signal.
Related papers
- When AI Meets Finance (StockAgent): Large Language Model-based Stock Trading in Simulated Real-world Environments [55.19252983108372]
We have developed a multi-agent AI system called StockAgent, driven by LLMs.
The StockAgent allows users to evaluate the impact of different external factors on investor trading.
It avoids the test set leakage issue present in existing trading simulation systems based on AI Agents.
arXiv Detail & Related papers (2024-07-15T06:49:30Z) - MOT: A Mixture of Actors Reinforcement Learning Method by Optimal Transport for Algorithmic Trading [6.305870529904885]
We propose MOT, which designs multiple actors with disentangled representation learning to model the different patterns of the market.
Experimental results on real futures market data demonstrate that MOT exhibits excellent profit capabilities while balancing risks.
arXiv Detail & Related papers (2024-06-03T01:42:52Z) - Combining Deep Learning on Order Books with Reinforcement Learning for
Profitable Trading [0.0]
This work focuses on forecasting returns across multiple horizons using order flow and training three temporal-difference imbalance learning models for five financial instruments.
The results prove potential but require further minimal modifications for consistently profitable trading to fully handle retail trading costs, slippage, and spread fluctuation.
arXiv Detail & Related papers (2023-10-24T15:58:58Z) - Action-Quantized Offline Reinforcement Learning for Robotic Skill
Learning [68.16998247593209]
offline reinforcement learning (RL) paradigm provides recipe to convert static behavior datasets into policies that can perform better than the policy that collected the data.
In this paper, we propose an adaptive scheme for action quantization.
We show that several state-of-the-art offline RL methods such as IQL, CQL, and BRAC improve in performance on benchmarks when combined with our proposed discretization scheme.
arXiv Detail & Related papers (2023-10-18T06:07:10Z) - A Novel Deep Reinforcement Learning Based Automated Stock Trading System
Using Cascaded LSTM Networks [3.593955557310285]
We propose a DRL based stock trading system using cascaded LSTM, which first uses LSTM to extract the time-series features from stock daily data, and then the features extracted are fed to the agent for training.
Experiments in DJI in the US market and SSE50 in the Chinese stock market show that our model outperforms previous baseline models in terms of cumulative returns and Sharp ratio.
arXiv Detail & Related papers (2022-12-06T03:22:06Z) - Mastering the Unsupervised Reinforcement Learning Benchmark from Pixels [112.63440666617494]
Reinforcement learning algorithms can succeed but require large amounts of interactions between the agent and the environment.
We propose a new method to solve it, using unsupervised model-based RL, for pre-training the agent.
We show robust performance on the Real-Word RL benchmark, hinting at resiliency to environment perturbations during adaptation.
arXiv Detail & Related papers (2022-09-24T14:22:29Z) - Deep Q-Learning Market Makers in a Multi-Agent Simulated Stock Market [58.720142291102135]
This paper focuses precisely on the study of these markets makers strategies from an agent-based perspective.
We propose the application of Reinforcement Learning (RL) for the creation of intelligent market markers in simulated stock markets.
arXiv Detail & Related papers (2021-12-08T14:55:21Z) - Deep Reinforcement Learning for Active High Frequency Trading [1.6874375111244329]
We introduce the first end-to-end Deep Reinforcement Learning (DRL) based framework for active high frequency trading in the stock market.
We train DRL agents to trade one unit of Intel Corporation stock by employing the Proximal Policy Optimization algorithm.
arXiv Detail & Related papers (2021-01-18T15:09:28Z) - AAMDRL: Augmented Asset Management with Deep Reinforcement Learning [5.801876281373619]
We show how Deep Reinforcement Learning can tackle this challenge.
Our contributions are threefold: (i) the use of contextual information also referred to as augmented state in DRL, (ii) the impact of a one period lag between observations and actions, and (iii) the implementation of a new repetitive train test method called walk forward analysis.
Although our experiment is on trading bots, it can easily be translated to other bot environments that operate in sequential environment with regime changes and noisy data.
arXiv Detail & Related papers (2020-09-30T03:55:47Z) - Stock2Vec: A Hybrid Deep Learning Framework for Stock Market Prediction
with Representation Learning and Temporal Convolutional Network [71.25144476293507]
We have proposed to develop a global hybrid deep learning framework to predict the daily prices in the stock market.
With representation learning, we derived an embedding called Stock2Vec, which gives us insight for the relationship among different stocks.
Our hybrid framework integrates both advantages and achieves better performance on the stock price prediction task than several popular benchmarked models.
arXiv Detail & Related papers (2020-09-29T22:54:30Z) - Deep Stock Predictions [58.720142291102135]
We consider the design of a trading strategy that performs portfolio optimization using Long Short Term Memory (LSTM) neural networks.
We then customize the loss function used to train the LSTM to increase the profit earned.
We find the LSTM model with the customized loss function to have an improved performance in the training bot over a regressive baseline such as ARIMA.
arXiv Detail & Related papers (2020-06-08T23:37:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.