Onflow: an online portfolio allocation algorithm
- URL: http://arxiv.org/abs/2312.05169v1
- Date: Fri, 8 Dec 2023 16:49:19 GMT
- Title: Onflow: an online portfolio allocation algorithm
- Authors: Gabriel Turinici and Pierre Brugiere
- Abstract summary: We introduce Onflow, a reinforcement learning technique that enables online optimization of portfolio allocation policies.
For log-normal assets, the strategy learned by Onflow, with transaction costs at zero, mimics Markowitz's optimal portfolio.
Onflow can remain efficient in regimes where other dynamical allocation techniques do not work anymore.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We introduce Onflow, a reinforcement learning technique that enables online
optimization of portfolio allocation policies based on gradient flows. We
devise dynamic allocations of an investment portfolio to maximize its expected
log return while taking into account transaction fees. The portfolio allocation
is parameterized through a softmax function, and at each time step, the
gradient flow method leads to an ordinary differential equation whose solutions
correspond to the updated allocations. This algorithm belongs to the large
class of stochastic optimization procedures; we measure its efficiency by
comparing our results to the mathematical theoretical values in a log-normal
framework and to standard benchmarks from the 'old NYSE' dataset. For
log-normal assets, the strategy learned by Onflow, with transaction costs at
zero, mimics Markowitz's optimal portfolio and thus the best possible asset
allocation strategy. Numerical experiments from the 'old NYSE' dataset show
that Onflow leads to dynamic asset allocation strategies whose performances
are: a) comparable to benchmark strategies such as Cover's Universal Portfolio
or Helmbold et al. "multiplicative updates" approach when transaction costs are
zero, and b) better than previous procedures when transaction costs are high.
Onflow can even remain efficient in regimes where other dynamical allocation
techniques do not work anymore. Therefore, as far as tested, Onflow appears to
be a promising dynamic portfolio management strategy based on observed prices
only and without any assumption on the laws of distributions of the underlying
assets' returns. In particular it could avoid model risk when building a
trading strategy.
Related papers
- $f$-PO: Generalizing Preference Optimization with $f$-divergence Minimization [91.43730624072226]
$f$-PO is a novel framework that generalizes and extends existing approaches.
We conduct experiments on state-of-the-art language models using benchmark datasets.
arXiv Detail & Related papers (2024-10-29T02:11:45Z) - Dynamic Portfolio Rebalancing: A Hybrid new Model Using GNNs and Pathfinding for Cost Efficiency [0.0]
This paper introduces a novel approach to optimizing portfolio rebalancing by integrating Graph Neural Networks (GNNs) for predicting transaction costs and Dijkstra's algorithm for identifying cost-efficient rebalancing paths.
Empirical results show that this hybrid approach significantly reduces transaction costs, offering a powerful tool for portfolio managers.
arXiv Detail & Related papers (2024-10-02T11:00:52Z) - Stochastic Q-learning for Large Discrete Action Spaces [79.1700188160944]
In complex environments with discrete action spaces, effective decision-making is critical in reinforcement learning (RL)
We present value-based RL approaches which, as opposed to optimizing over the entire set of $n$ actions, only consider a variable set of actions, possibly as small as $mathcalO(log(n)$)$.
The presented value-based RL methods include, among others, Q-learning, StochDQN, StochDDQN, all of which integrate this approach for both value-function updates and action selection.
arXiv Detail & Related papers (2024-05-16T17:58:44Z) - Optimizing Portfolio Management and Risk Assessment in Digital Assets
Using Deep Learning for Predictive Analysis [5.015409508372732]
This paper introduces the DQN algorithm into asset management portfolios in a novel and straightforward way.
The performance greatly exceeds the benchmark, which fully proves the effectiveness of the DRL algorithm in portfolio management.
Since different assets are trained separately as environments, there may be a phenomenon of Q value drift among different assets.
arXiv Detail & Related papers (2024-02-25T05:23:57Z) - Cryptocurrency Portfolio Optimization by Neural Networks [81.20955733184398]
This paper proposes an effective algorithm based on neural networks to take advantage of these investment products.
A deep neural network, which outputs the allocation weight of each asset at a time interval, is trained to maximize the Sharpe ratio.
A novel loss term is proposed to regulate the network's bias towards a specific asset, thus enforcing the network to learn an allocation strategy that is close to a minimum variance strategy.
arXiv Detail & Related papers (2023-10-02T12:33:28Z) - Sparse Index Tracking: Simultaneous Asset Selection and Capital Allocation via $\ell_0$-Constrained Portfolio [7.5684339230894135]
A sparse portfolio is preferable to a full portfolio in terms of reducing transaction costs and avoiding illiquid assets.
We propose a new problem formulation of sparse index tracking using an $ell_p$-norm constraint.
Our approach offers a choice between constraints on portfolio and turnover sparsity, further reducing transaction costs by limiting asset updates at each rebalancing interval.
arXiv Detail & Related papers (2023-07-22T04:47:30Z) - Towards Understanding and Improving GFlowNet Training [71.85707593318297]
We introduce an efficient evaluation strategy to compare the learned sampling distribution to the target reward distribution.
We propose prioritized replay training of high-reward $x$, relative edge flow policy parametrization, and a novel guided trajectory balance objective.
arXiv Detail & Related papers (2023-05-11T22:50:41Z) - High-Dimensional Stock Portfolio Trading with Deep Reinforcement
Learning [0.0]
The algorithm is capable of trading high-dimensional portfolios from cross-sectional datasets of any size.
We sequentially set up environments by sampling one asset for each environment while rewarding investments with the resulting asset's return and cash reservation with the average return of the set of assets.
arXiv Detail & Related papers (2021-12-09T08:30:45Z) - Deep Hedging of Derivatives Using Reinforcement Learning [0.3313576045747072]
We show how reinforcement learning can be used to derive optimal hedging strategies for derivatives when there are transaction costs.
We find that a hybrid approach involving the use of an accounting P&L approach that incorporates a relatively simple valuation model works well.
arXiv Detail & Related papers (2021-03-29T07:43:30Z) - Semi-supervised Batch Active Learning via Bilevel Optimization [89.37476066973336]
We formulate our approach as a data summarization problem via bilevel optimization.
We show that our method is highly effective in keyword detection tasks in the regime when only few labeled samples are available.
arXiv Detail & Related papers (2020-10-19T16:53:24Z) - Provably Efficient Exploration in Policy Optimization [117.09887790160406]
This paper proposes an Optimistic variant of the Proximal Policy Optimization algorithm (OPPO)
OPPO achieves $tildeO(sqrtd2 H3 T )$ regret.
To the best of our knowledge, OPPO is the first provably efficient policy optimization algorithm that explores.
arXiv Detail & Related papers (2019-12-12T08:40:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.