Learning (Local) Surrogate Loss Functions for Predict-Then-Optimize
Problems
- URL: http://arxiv.org/abs/2203.16067v1
- Date: Wed, 30 Mar 2022 05:46:54 GMT
- Title: Learning (Local) Surrogate Loss Functions for Predict-Then-Optimize
Problems
- Authors: Sanket Shah, Bryan Wilder, Andrew Perrault, Milind Tambe
- Abstract summary: Decision-Focused Learning (DFL) is a paradigm for tailoring a predictive model to a downstream optimisation task.
We present an approach to learn faithful task-specific surrogates which (a) only requires access to a black-box oracle that can solve the optimisation problem and is thus generalizable, and (b) can be convex by gradients and so can be easily optimized over.
- Score: 58.954414264760956
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Decision-Focused Learning (DFL) is a paradigm for tailoring a predictive
model to a downstream optimisation task that uses its predictions, so that it
can perform better on that specific task. The main technical challenge
associated with DFL is that it requires being able to differentiate through
$argmin$ operations to work. However, these $argmin$ optimisations are often
piecewise constant and, as a result, naively differentiating through them would
provide uninformative gradients. Past work has largely focused on getting
around this issue by handcrafting task-specific surrogates to the original
optimisation problem that provide informative gradients when differentiated
through. However, finding these surrogates can be challenging and the need to
handcraft surrogates for each new task limits the usability of DFL. In
addition, even after applying these relaxation techniques, there are no
guarantees that the resulting surrogates are convex and, as a result, training
a predictive model on them may lead to said model getting stuck in local
minimas.
In this paper, we provide an approach to learn faithful task-specific
surrogates which (a) only requires access to a black-box oracle that can solve
the optimisation problem and is thus generalizable, and (b) can be convex by
construction and so can be easily optimized over. To the best of our knowledge,
this is the first work on using learning to find good surrogates for DFL. We
evaluate our approach on a budget allocation problem from the literature and
find that our approach outperforms even the hand-crafted (non-convex) surrogate
loss proposed by the original paper. Taking a step back, we hope that the
generality and simplicity of our approach will help lower the barrier
associated with implementing DFL-based solutions in practice. To that end, we
are currently working on extending our experiments to more domains.
Related papers
- Solving Hidden Monotone Variational Inequalities with Surrogate Losses [23.565183680315073]
We propose a principled surrogate-based approach compatible with deep learning to solve variational inequality (VI) problems.
We demonstrate our surrogate-based approach is effective in min-max optimization and minimizing projected Bellman error.
In the deep reinforcement learning case, we propose a novel variant of TD(0) which is more compute and sample efficient.
arXiv Detail & Related papers (2024-11-07T22:42:08Z) - Directed Exploration in Reinforcement Learning from Linear Temporal Logic [59.707408697394534]
Linear temporal logic (LTL) is a powerful language for task specification in reinforcement learning.
We show that the synthesized reward signal remains fundamentally sparse, making exploration challenging.
We show how better exploration can be achieved by further leveraging the specification and casting its corresponding Limit Deterministic B"uchi Automaton (LDBA) as a Markov reward process.
arXiv Detail & Related papers (2024-08-18T14:25:44Z) - On Leveraging Large Language Models for Enhancing Entity Resolution: A Cost-efficient Approach [7.996010840316654]
We propose an uncertainty reduction framework using Large Language Models (LLMs) to improve entity resolution results.
LLMs capitalize on their advanced linguistic capabilities and a pay-as-you-go'' model that provides significant advantages to those without extensive data science expertise.
We show that our method is efficient and effective, offering promising applications in real-world tasks.
arXiv Detail & Related papers (2024-01-07T09:06:58Z) - Score Function Gradient Estimation to Widen the Applicability of Decision-Focused Learning [17.962860438133312]
Decision-focused learning (DFL) paradigm overcomes limitation by training to directly minimize a task loss, e.g. regret.
We propose an alternative method that makes no such assumptions, it combines smoothing with score function estimation which works on any task loss.
Experiments show that it typically requires more epochs, but that it is on par with specialized methods and performs especially well for the difficult case of problems with uncertainty in the constraints, in terms of solution quality, scalability, or both.
arXiv Detail & Related papers (2023-07-11T12:32:13Z) - Leaving the Nest: Going Beyond Local Loss Functions for
Predict-Then-Optimize [57.22851616806617]
We show that our method achieves state-of-the-art results in four domains from the literature.
Our approach outperforms the best existing method by nearly 200% when the localness assumption is broken.
arXiv Detail & Related papers (2023-05-26T11:17:45Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - A new perspective on classification: optimally allocating limited
resources to uncertain tasks [4.169130102668252]
In credit card fraud detection, for instance, a bank can only assign a small subset of transactions to their fraud investigations team.
We argue that using classification to address task uncertainty is inherently suboptimal as it does not take into account the available capacity.
We present a novel solution using learning to rank by directly optimizing the assignment's expected profit given limited capacity.
arXiv Detail & Related papers (2022-02-09T10:14:45Z) - Few-shot Quality-Diversity Optimization [50.337225556491774]
Quality-Diversity (QD) optimization has been shown to be effective tools in dealing with deceptive minima and sparse rewards in Reinforcement Learning.
We show that, given examples from a task distribution, information about the paths taken by optimization in parameter space can be leveraged to build a prior population, which when used to initialize QD methods in unseen environments, allows for few-shot adaptation.
Experiments carried in both sparse and dense reward settings using robotic manipulation and navigation benchmarks show that it considerably reduces the number of generations that are required for QD optimization in these environments.
arXiv Detail & Related papers (2021-09-14T17:12:20Z) - Prior Guided Feature Enrichment Network for Few-Shot Segmentation [64.91560451900125]
State-of-the-art semantic segmentation methods require sufficient labeled data to achieve good results.
Few-shot segmentation is proposed to tackle this problem by learning a model that quickly adapts to new classes with a few labeled support samples.
Theses frameworks still face the challenge of generalization ability reduction on unseen classes due to inappropriate use of high-level semantic information.
arXiv Detail & Related papers (2020-08-04T10:41:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.