Reinforcement Learning with Intrinsic Affinity for Personalized Asset
Management
- URL: http://arxiv.org/abs/2204.09218v1
- Date: Wed, 20 Apr 2022 04:33:32 GMT
- Title: Reinforcement Learning with Intrinsic Affinity for Personalized Asset
Management
- Authors: Charl Maree and Christian W. Omlin
- Abstract summary: We develop a regularization method that ensures that strategies have global intrinsic affinities.
We capitalize on these intrinsic affinities to make our model inherently interpretable.
We demonstrate how RL agents can be trained to orchestrate such individual policies for particular personality profiles and still achieve high returns.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The common purpose of applying reinforcement learning (RL) to asset
management is the maximization of profit. The extrinsic reward function used to
learn an optimal strategy typically does not take into account any other
preferences or constraints. We have developed a regularization method that
ensures that strategies have global intrinsic affinities, i.e., different
personalities may have preferences for certain assets which may change over
time. We capitalize on these intrinsic policy affinities to make our RL model
inherently interpretable. We demonstrate how RL agents can be trained to
orchestrate such individual policies for particular personality profiles and
still achieve high returns.
Related papers
- COPR: Continual Learning Human Preference through Optimal Policy Regularization [32.54658750353585]
We propose a new method called Continual Optimal Policy Regularization (COPR)
COPR involves a single learning phase and doesn't necessitate complex reinforcement learning.
Our experimental results show that COPR outperforms strong Continuous Learning (CL) baselines.
arXiv Detail & Related papers (2023-10-24T10:05:32Z) - Reinforcement Learning from Diverse Human Preferences [68.4294547285359]
This paper develops a method for crowd-sourcing preference labels and learning from diverse human preferences.
The proposed method is tested on a variety of tasks in DMcontrol and Meta-world.
It has shown consistent and significant improvements over existing preference-based RL algorithms when learning from diverse feedback.
arXiv Detail & Related papers (2023-01-27T15:18:54Z) - Emergent Behaviors in Multi-Agent Target Acquisition [0.0]
We simulate a Multi-Agent System (MAS) using Reinforcement Learning (RL) in a pursuit-evasion game.
We create different adversarial scenarios by replacing RL-trained pursuers' policies with two distinct (non-RL) analytical strategies.
The novelty of our approach entails the creation of an influential feature set that reveals underlying data regularities.
arXiv Detail & Related papers (2022-12-15T15:20:58Z) - Flexible Attention-Based Multi-Policy Fusion for Efficient Deep
Reinforcement Learning [78.31888150539258]
Reinforcement learning (RL) agents have long sought to approach the efficiency of human learning.
Prior studies in RL have incorporated external knowledge policies to help agents improve sample efficiency.
We present Knowledge-Grounded RL (KGRL), an RL paradigm fusing multiple knowledge policies and aiming for human-like efficiency and flexibility.
arXiv Detail & Related papers (2022-10-07T17:56:57Z) - Policy Gradient for Reinforcement Learning with General Utilities [50.65940899590487]
In Reinforcement Learning (RL), the goal of agents is to discover an optimal policy that maximizes the expected cumulative rewards.
Many supervised and unsupervised RL problems are not covered in the Linear RL framework.
We derive the policy gradient theorem for RL with general utilities.
arXiv Detail & Related papers (2022-10-03T14:57:46Z) - Learning Meta Representations for Agents in Multi-Agent Reinforcement
Learning [12.170248966278281]
In multi-agent reinforcement learning, behaviors that agents learn in a single Markov Game (MG) are typically confined to the given agent number.
In this work, our focus is on creating agents that can generalize across population-varying MGs.
Instead of learning a unimodal policy, each agent learns a policy set comprising effective strategies across a variety of games.
arXiv Detail & Related papers (2021-08-30T04:30:53Z) - Policy Mirror Descent for Regularized Reinforcement Learning: A
Generalized Framework with Linear Convergence [60.20076757208645]
This paper proposes a general policy mirror descent (GPMD) algorithm for solving regularized RL.
We demonstrate that our algorithm converges linearly over an entire range learning rates, in a dimension-free fashion, to the global solution.
arXiv Detail & Related papers (2021-05-24T02:21:34Z) - DisCo RL: Distribution-Conditioned Reinforcement Learning for
General-Purpose Policies [116.12670064963625]
We develop an off-policy algorithm called distribution-conditioned reinforcement learning (DisCo RL) to efficiently learn contextual policies.
We evaluate DisCo RL on a variety of robot manipulation tasks and find that it significantly outperforms prior methods on tasks that require generalization to new goal distributions.
arXiv Detail & Related papers (2021-04-23T16:51:58Z) - Dirichlet policies for reinforced factor portfolios [1.3706331473063877]
This article aims to combine factor investing and reinforcement learning (RL)
The agent learns through sequential random allocations which rely on firms' characteristics.
Across a large range of parametric choices, our result indicates that RL-based portfolios are very close to the equally-weighted (1/N) allocation.
arXiv Detail & Related papers (2020-11-10T20:25:41Z) - Preference-based Reinforcement Learning with Finite-Time Guarantees [76.88632321436472]
Preference-based Reinforcement Learning (PbRL) replaces reward values in traditional reinforcement learning to better elicit human opinion on the target objective.
Despite promising results in applications, the theoretical understanding of PbRL is still in its infancy.
We present the first finite-time analysis for general PbRL problems.
arXiv Detail & Related papers (2020-06-16T03:52:41Z) - A General Framework on Enhancing Portfolio Management with Reinforcement
Learning [3.6985496077087743]
Portfolio management concerns continuous reallocation of funds and assets across financial instruments to meet the desired returns to risk profile.
Deep reinforcement learning (RL) has gained increasing interest in portfolio management, where RL agents are trained base on financial data to optimize the asset reallocation process.
We propose a general RL framework for asset management that enables continuous asset weights, short selling and making decisions with relevant features.
arXiv Detail & Related papers (2019-11-26T23:41:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.