Optimizing Credit Limit Adjustments Under Adversarial Goals Using
Reinforcement Learning
- URL: http://arxiv.org/abs/2306.15585v2
- Date: Fri, 16 Feb 2024 16:12:21 GMT
- Title: Optimizing Credit Limit Adjustments Under Adversarial Goals Using
Reinforcement Learning
- Authors: Sherly Alfonso-S\'anchez, Jes\'us Solano, Alejandro Correa-Bahnsen,
Kristina P. Sendova, and Cristi\'an Bravo
- Abstract summary: We seek to find and automatize an optimal credit card limit adjustment policy by employing reinforcement learning techniques.
Our research establishes a conceptual structure for applying reinforcement learning framework to credit limit adjustment.
- Score: 42.303733194571905
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Reinforcement learning has been explored for many problems, from video games
with deterministic environments to portfolio and operations management in which
scenarios are stochastic; however, there have been few attempts to test these
methods in banking problems. In this study, we sought to find and automatize an
optimal credit card limit adjustment policy by employing reinforcement learning
techniques. Because of the historical data available, we considered two
possible actions per customer, namely increasing or maintaining an individual's
current credit limit. To find this policy, we first formulated this
decision-making question as an optimization problem in which the expected
profit was maximized; therefore, we balanced two adversarial goals: maximizing
the portfolio's revenue and minimizing the portfolio's provisions. Second,
given the particularities of our problem, we used an offline learning strategy
to simulate the impact of the action based on historical data from a super-app
in Latin America to train our reinforcement learning agent. Our results, based
on the proposed methodology involving synthetic experimentation, show that a
Double Q-learning agent with optimized hyperparameters can outperform other
strategies and generate a non-trivial optimal policy not only reflecting the
complex nature of this decision but offering an incentive to explore
reinforcement learning in real-world banking scenarios. Our research
establishes a conceptual structure for applying reinforcement learning
framework to credit limit adjustment, presenting an objective technique to make
these decisions primarily based on data-driven methods rather than relying only
on expert-driven systems. We also study the use of alternative data for the
problem of balance prediction, as the latter is a requirement of our proposed
model. We find the use of such data does not always bring prediction gains.
Related papers
- Optimal Execution with Reinforcement Learning [0.4972323953932129]
This study investigates the development of an optimal execution strategy through reinforcement learning.
We present a custom MDP formulation followed by the results of our methodology and benchmark the performance against standard execution strategies.
arXiv Detail & Related papers (2024-11-10T08:21:03Z) - Learning to Cover: Online Learning and Optimization with Irreversible Decisions [50.5775508521174]
We find that regret grows sub-linearly at a rate $Thetaleft(mfrac12cdotfrac11-2-Tright)$, thus converging exponentially fast to $Theta(sqrtm)$.
These findings underscore the benefits of limited online learning and optimization, in that even a few rounds can provide significant benefits as compared to a no-learning baseline.
arXiv Detail & Related papers (2024-06-20T23:00:25Z) - Leaving the Nest: Going Beyond Local Loss Functions for
Predict-Then-Optimize [57.22851616806617]
We show that our method achieves state-of-the-art results in four domains from the literature.
Our approach outperforms the best existing method by nearly 200% when the localness assumption is broken.
arXiv Detail & Related papers (2023-05-26T11:17:45Z) - Re-thinking Data Availablity Attacks Against Deep Neural Networks [53.64624167867274]
In this paper, we re-examine the concept of unlearnable examples and discern that the existing robust error-minimizing noise presents an inaccurate optimization objective.
We introduce a novel optimization paradigm that yields improved protection results with reduced computational time requirements.
arXiv Detail & Related papers (2023-05-18T04:03:51Z) - When Demonstrations Meet Generative World Models: A Maximum Likelihood
Framework for Offline Inverse Reinforcement Learning [62.00672284480755]
This paper aims to recover the structure of rewards and environment dynamics that underlie observed actions in a fixed, finite set of demonstrations from an expert agent.
Accurate models of expertise in executing a task has applications in safety-sensitive applications such as clinical decision making and autonomous driving.
arXiv Detail & Related papers (2023-02-15T04:14:20Z) - Constrained Policy Optimization for Controlled Self-Learning in
Conversational AI Systems [18.546197100318693]
We introduce a scalable framework for supporting fine-grained exploration targets for individual domains via user-defined constraints.
We present a novel meta-gradient learning approach that is scalable and practical to address this problem.
We conduct extensive experiments using data from a real-world conversational AI on a set of realistic constraint benchmarks.
arXiv Detail & Related papers (2022-09-17T23:44:13Z) - Off-Policy Optimization of Portfolio Allocation Policies under
Constraints [0.8848340429852071]
Dynamic portfolio optimization problem in finance frequently requires learning policies that adhere to various constraints, driven by investor preferences and risk.
We motivate this problem of finding an allocation policy within a sequential decision making framework and study the effects of: (a) using data collected under previously employed policies, which may be sub-optimal and constraint-violating, and (b) imposing desired constraints while computing near-optimal policies with this data.
arXiv Detail & Related papers (2020-12-21T22:22:04Z) - Toward Optimal Probabilistic Active Learning Using a Bayesian Approach [4.380488084997317]
Active learning aims at reducing the labeling costs by an efficient and effective allocation of costly labeling resources.
By reformulating existing selection strategies within our proposed model, we can explain which aspects are not covered in current state-of-the-art.
arXiv Detail & Related papers (2020-06-02T15:59:42Z) - Interpretable Personalization via Policy Learning with Linear Decision
Boundaries [14.817218449140338]
effective personalization of goods and services has become a core business for companies to improve revenues and maintain a competitive edge.
This paper studies the personalization problem through the lens of policy learning.
We propose a class of policies with linear decision boundaries and propose learning algorithms using tools from causal inference.
arXiv Detail & Related papers (2020-03-17T05:48:27Z) - Cost-Sensitive Portfolio Selection via Deep Reinforcement Learning [100.73223416589596]
We propose a cost-sensitive portfolio selection method with deep reinforcement learning.
Specifically, a novel two-stream portfolio policy network is devised to extract both price series patterns and asset correlations.
A new cost-sensitive reward function is developed to maximize the accumulated return and constrain both costs via reinforcement learning.
arXiv Detail & Related papers (2020-03-06T06:28:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.