A new perspective on classification: optimally allocating limited
resources to uncertain tasks
- URL: http://arxiv.org/abs/2202.04369v1
- Date: Wed, 9 Feb 2022 10:14:45 GMT
- Title: A new perspective on classification: optimally allocating limited
resources to uncertain tasks
- Authors: Toon Vanderschueren, Bart Baesens, Tim Verdonck, and Wouter Verbeke
- Abstract summary: In credit card fraud detection, for instance, a bank can only assign a small subset of transactions to their fraud investigations team.
We argue that using classification to address task uncertainty is inherently suboptimal as it does not take into account the available capacity.
We present a novel solution using learning to rank by directly optimizing the assignment's expected profit given limited capacity.
- Score: 4.169130102668252
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A central problem in business concerns the optimal allocation of limited
resources to a set of available tasks, where the payoff of these tasks is
inherently uncertain. In credit card fraud detection, for instance, a bank can
only assign a small subset of transactions to their fraud investigations team.
Typically, such problems are solved using a classification framework, where the
focus is on predicting task outcomes given a set of characteristics. Resources
are then allocated to the tasks that are predicted to be the most likely to
succeed. However, we argue that using classification to address task
uncertainty is inherently suboptimal as it does not take into account the
available capacity. Therefore, we first frame the problem as a type of
assignment problem. Then, we present a novel solution using learning to rank by
directly optimizing the assignment's expected profit given limited, stochastic
capacity. This is achieved by optimizing a specific instance of the net
discounted cumulative gain, a commonly used class of metrics in learning to
rank. Empirically, we demonstrate that our new method achieves higher expected
profit and expected precision compared to a classification approach for a wide
variety of application areas and data sets. This illustrates the benefit of an
integrated approach and of explicitly considering the available resources when
learning a predictive model.
Related papers
- Autoregressive Policy Optimization for Constrained Allocation Tasks [4.316765170255551]
We propose a new method for constrained allocation tasks based on an autoregressive process to sequentially sample allocations for each entity.
In addition, we introduce a novel de-biasing mechanism to counter the initial bias caused by sequential sampling.
arXiv Detail & Related papers (2024-09-27T13:27:15Z) - Leaving the Nest: Going Beyond Local Loss Functions for
Predict-Then-Optimize [57.22851616806617]
We show that our method achieves state-of-the-art results in four domains from the literature.
Our approach outperforms the best existing method by nearly 200% when the localness assumption is broken.
arXiv Detail & Related papers (2023-05-26T11:17:45Z) - Minimalistic Predictions to Schedule Jobs with Online Precedence
Constraints [117.8317521974783]
We consider non-clairvoyant scheduling with online precedence constraints.
An algorithm is oblivious to any job dependencies and learns about a job only if all of its predecessors have been completed.
arXiv Detail & Related papers (2023-01-30T13:17:15Z) - Budgeted Classification with Rejection: An Evolutionary Method with
Multiple Objectives [0.0]
Budgeted, sequential classifiers (BSCs) process inputs through a sequence of partial feature acquisition and evaluation steps.
This allows for an efficient evaluation of inputs that prevents unneeded feature acquisition.
We propose a problem-specific genetic algorithm to build budgeted, sequential classifiers with confidence-based reject options.
arXiv Detail & Related papers (2022-05-01T22:05:16Z) - Learning (Local) Surrogate Loss Functions for Predict-Then-Optimize
Problems [58.954414264760956]
Decision-Focused Learning (DFL) is a paradigm for tailoring a predictive model to a downstream optimisation task.
We present an approach to learn faithful task-specific surrogates which (a) only requires access to a black-box oracle that can solve the optimisation problem and is thus generalizable, and (b) can be convex by gradients and so can be easily optimized over.
arXiv Detail & Related papers (2022-03-30T05:46:54Z) - MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven
Reinforcement Learning [65.52675802289775]
We show that an uncertainty aware classifier can solve challenging reinforcement learning problems.
We propose a novel method for computing the normalized maximum likelihood (NML) distribution.
We show that the resulting algorithm has a number of intriguing connections to both count-based exploration methods and prior algorithms for learning reward functions.
arXiv Detail & Related papers (2021-07-15T08:19:57Z) - The Single-Noun Prior for Image Clustering [34.97652735163338]
Self-supervised clustering methods have achieved increasing accuracy in recent years but do not yet perform as well as supervised classification methods.
We introduce the "single-noun" prior - which states that semantic clusters tend to correspond to concepts that humans label by a single-noun.
We show that our formulation is a special case of the facility location problem, and introduce a simple-yet-effective approach for solving this optimization task at scale.
arXiv Detail & Related papers (2021-04-08T17:54:37Z) - Provably Efficient Reward-Agnostic Navigation with Linear Value
Iteration [143.43658264904863]
We show how iteration under a more standard notion of low inherent Bellman error, typically employed in least-square value-style algorithms, can provide strong PAC guarantees on learning a near optimal value function.
We present a computationally tractable algorithm for the reward-free setting and show how it can be used to learn a near optimal policy for any (linear) reward function.
arXiv Detail & Related papers (2020-08-18T04:34:21Z) - Sequential Transfer in Reinforcement Learning with a Generative Model [48.40219742217783]
We show how to reduce the sample complexity for learning new tasks by transferring knowledge from previously-solved ones.
We derive PAC bounds on its sample complexity which clearly demonstrate the benefits of using this kind of prior knowledge.
We empirically verify our theoretical findings in simple simulated domains.
arXiv Detail & Related papers (2020-07-01T19:53:35Z) - Regularized Online Allocation Problems: Fairness and Beyond [7.433931244705934]
We introduce the emphregularized online allocation problem, a variant that includes a non-linear regularizer acting on the total resource consumption.
In this problem, requests repeatedly arrive over time and, for each request, a decision maker needs to take an action that generates a reward and consumes resources.
The objective is to simultaneously maximize additively separable rewards and the value of a non-separable regularizer subject to the resource constraints.
arXiv Detail & Related papers (2020-07-01T14:24:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.