Cost-Sensitive Portfolio Selection via Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2003.03051v1
- Date: Fri, 6 Mar 2020 06:28:17 GMT
- Title: Cost-Sensitive Portfolio Selection via Deep Reinforcement Learning
- Authors: Yifan Zhang, Peilin Zhao, Qingyao Wu, Bin Li, Junzhou Huang, and
Mingkui Tan
- Abstract summary: We propose a cost-sensitive portfolio selection method with deep reinforcement learning.
Specifically, a novel two-stream portfolio policy network is devised to extract both price series patterns and asset correlations.
A new cost-sensitive reward function is developed to maximize the accumulated return and constrain both costs via reinforcement learning.
- Score: 100.73223416589596
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Portfolio Selection is an important real-world financial task and has
attracted extensive attention in artificial intelligence communities. This
task, however, has two main difficulties: (i) the non-stationary price series
and complex asset correlations make the learning of feature representation very
hard; (ii) the practicality principle in financial markets requires controlling
both transaction and risk costs. Most existing methods adopt handcraft features
and/or consider no constraints for the costs, which may make them perform
unsatisfactorily and fail to control both costs in practice. In this paper, we
propose a cost-sensitive portfolio selection method with deep reinforcement
learning. Specifically, a novel two-stream portfolio policy network is devised
to extract both price series patterns and asset correlations, while a new
cost-sensitive reward function is developed to maximize the accumulated return
and constrain both costs via reinforcement learning. We theoretically analyze
the near-optimality of the proposed reward, which shows that the growth rate of
the policy regarding this reward function can approach the theoretical optimum.
We also empirically evaluate the proposed method on real-world datasets.
Promising results demonstrate the effectiveness and superiority of the proposed
method in terms of profitability, cost-sensitivity and representation
abilities.
Related papers
- Deep Generative Demand Learning for Newsvendor and Pricing [7.594251468240168]
We consider data-driven inventory and pricing decisions in the feature-based newsvendor problem.
We propose a novel approach leveraging conditional deep generative models (cDGMs) to address these challenges.
We provide theoretical guarantees for our approach, including the consistency of profit estimation and convergence of our decisions to the optimal solution.
arXiv Detail & Related papers (2024-11-13T14:17:26Z) - Neural Active Learning Beyond Bandits [69.99592173038903]
We study both stream-based and pool-based active learning with neural network approximations.
We propose two algorithms based on the newly designed exploitation and exploration neural networks for stream-based and pool-based active learning.
arXiv Detail & Related papers (2024-04-18T21:52:14Z) - Utility Fairness in Contextual Dynamic Pricing with Demand Learning [23.26236046836737]
This paper introduces a novel contextual bandit algorithm for personalized pricing under utility fairness constraints.
Our approach, which incorporates dynamic pricing and demand learning, addresses the critical challenge of fairness in pricing strategies.
arXiv Detail & Related papers (2023-11-28T05:19:23Z) - Online Learning under Budget and ROI Constraints via Weak Adaptivity [57.097119428915796]
Existing primal-dual algorithms for constrained online learning problems rely on two fundamental assumptions.
We show how such assumptions can be circumvented by endowing standard primal-dual templates with weakly adaptive regret minimizers.
We prove the first best-of-both-worlds no-regret guarantees which hold in absence of the two aforementioned assumptions.
arXiv Detail & Related papers (2023-02-02T16:30:33Z) - Online Learning with Knapsacks: the Best of Both Worlds [54.28273783164608]
We casting online learning problems in which a decision maker wants to maximize their expected reward without violating a finite set of $m$m resource constraints.
Our framework allows the decision maker to handle its evidence flexibility and costoretic functions.
arXiv Detail & Related papers (2022-02-28T12:10:48Z) - Present-Biased Optimization [8.775878711928591]
The paper extends the framework proposed by Akerlof (1991) for studying various aspects of human behavior related to time-inconsistent planning.
We show that the ratio between the cost of the solutions computed by present-biased agents and the cost of the optimal solutions may differ significantly depending on the problem constraints.
arXiv Detail & Related papers (2020-12-29T12:40:59Z) - Cost-effective Variational Active Entity Resolution [4.238343046459798]
We devise an entity resolution method that builds on the robustness conferred by deep autoencoders to reduce human-involvement costs.
Specifically, we reduce the cost of training deep entity resolution models by performing unsupervised representation learning.
Finally, we reduce the cost of labelling training data through an active learning approach that builds on the properties conferred by the use of deep autoencoders.
arXiv Detail & Related papers (2020-11-20T13:47:11Z) - Reinforcement Learning with Efficient Active Feature Acquisition [59.91808801541007]
In real-life, information acquisition might correspond to performing a medical test on a patient.
We propose a model-based reinforcement learning framework that learns an active feature acquisition policy.
Key to the success is a novel sequential variational auto-encoder that learns high-quality representations from partially observed states.
arXiv Detail & Related papers (2020-11-02T08:46:27Z) - Toward Optimal Probabilistic Active Learning Using a Bayesian Approach [4.380488084997317]
Active learning aims at reducing the labeling costs by an efficient and effective allocation of costly labeling resources.
By reformulating existing selection strategies within our proposed model, we can explain which aspects are not covered in current state-of-the-art.
arXiv Detail & Related papers (2020-06-02T15:59:42Z) - Hierarchical Adaptive Contextual Bandits for Resource Constraint based
Recommendation [49.69139684065241]
Contextual multi-armed bandit (MAB) achieves cutting-edge performance on a variety of problems.
In this paper, we propose a hierarchical adaptive contextual bandit method (HATCH) to conduct the policy learning of contextual bandits with a budget constraint.
arXiv Detail & Related papers (2020-04-02T17:04:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.