Feasibility-Aware Decision-Focused Learning for Predicting Parameters in the Constraints
- URL: http://arxiv.org/abs/2510.04951v2
- Date: Sun, 26 Oct 2025 08:52:55 GMT
- Title: Feasibility-Aware Decision-Focused Learning for Predicting Parameters in the Constraints
- Authors: Jayanta Mandi, Marianne Defresne, Senne Berden, Tias Guns,
- Abstract summary: Decision-focused learning implements the first stage by training a machine learning (ML) model to optimize the quality of the decisions made using predicted parameters.<n>We derive two novel loss functions based on maximum likelihood estimation.<n>We experimentally demonstrate that adjusting this parameter provides decision-makers control over this trade-off.
- Score: 8.380358508407637
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: When some parameters of a constrained optimization problem (COP) are uncertain, this gives rise to a predict-then-optimize (PtO) problem, comprising two stages: the prediction of the unknown parameters from contextual information and the subsequent optimization using those predicted parameters. Decision-focused learning (DFL) implements the first stage by training a machine learning (ML) model to optimize the quality of the decisions made using the predicted parameters. When the predicted parameters occur in the constraints, they can lead to infeasible solutions. Therefore, it is important to simultaneously manage both feasibility and decision quality. We develop a DFL framework for predicting constraint parameters in a generic COP. While prior works typically assume that the underlying optimization problem is a linear program (LP) or integer LP (ILP), our approach makes no such assumption. We derive two novel loss functions based on maximum likelihood estimation (MLE): the first one penalizes infeasibility (by penalizing predicted parameters that lead to infeasible solutions), while the second one penalizes suboptimal decisions (by penalizing predicted parameters that make the true optimal solution infeasible). We introduce a single tunable parameter to form a weighted average of the two losses, allowing decision-makers to balance suboptimality and feasibility. We experimentally demonstrate that adjusting this parameter provides decision-makers control over this trade-off. Moreover, across several COP instances, we show that adjusting the tunable parameter allows a decision-maker to prioritize either suboptimality or feasibility, outperforming the performance of existing baselines in either objective.
Related papers
- Scaling Exponents Across Parameterizations and Optimizers [94.54718325264218]
We propose a new perspective on parameterization by investigating a key assumption in prior work.
Our empirical investigation includes tens of thousands of models trained with all combinations of threes.
We find that the best learning rate scaling prescription would often have been excluded by the assumptions in prior work.
arXiv Detail & Related papers (2024-07-08T12:32:51Z) - Double Duality: Variational Primal-Dual Policy Optimization for
Constrained Reinforcement Learning [132.7040981721302]
We study the Constrained Convex Decision Process (MDP), where the goal is to minimize a convex functional of the visitation measure.
Design algorithms for a constrained convex MDP faces several challenges, including handling the large state space.
arXiv Detail & Related papers (2024-02-16T16:35:18Z) - End-to-End Learning for Fair Multiobjective Optimization Under
Uncertainty [55.04219793298687]
The Predict-Then-Forecast (PtO) paradigm in machine learning aims to maximize downstream decision quality.
This paper extends the PtO methodology to optimization problems with nondifferentiable Ordered Weighted Averaging (OWA) objectives.
It shows how optimization of OWA functions can be effectively integrated with parametric prediction for fair and robust optimization under uncertainty.
arXiv Detail & Related papers (2024-02-12T16:33:35Z) - Decision-focused predictions via pessimistic bilevel optimization: complexity and algorithms [0.7499722271664147]
We contribute to recent efforts in decision-focused predictions, i.e., to build predictive models.<n>We show computational results of a problem, including its membership in NP.<n>We then formulate various computational techniques to achieve empirical tractability.
arXiv Detail & Related papers (2023-12-29T15:05:00Z) - Predict-Then-Optimize by Proxy: Learning Joint Models of Prediction and
Optimization [59.386153202037086]
Predict-Then- framework uses machine learning models to predict unknown parameters of an optimization problem from features before solving.
This approach can be inefficient and requires handcrafted, problem-specific rules for backpropagation through the optimization step.
This paper proposes an alternative method, in which optimal solutions are learned directly from the observable features by predictive models.
arXiv Detail & Related papers (2023-11-22T01:32:06Z) - Score Function Gradient Estimation to Widen the Applicability of Decision-Focused Learning [17.962860438133312]
Decision-focused learning (DFL) paradigm overcomes limitation by training to directly minimize a task loss, e.g. regret.
We propose an alternative method that makes no such assumptions, it combines smoothing with score function estimation which works on any task loss.
Experiments show that it typically requires more epochs, but that it is on par with specialized methods and performs especially well for the difficult case of problems with uncertainty in the constraints, in terms of solution quality, scalability, or both.
arXiv Detail & Related papers (2023-07-11T12:32:13Z) - Regret Bounds and Experimental Design for Estimate-then-Optimize [9.340611077939828]
In practical applications, data is used to make decisions in two steps: estimation and optimization.
Errors in the estimation step can lead estimate-then-optimize to sub-optimal decisions.
We provide a novel bound on this regret for smooth and unconstrained optimization problems.
arXiv Detail & Related papers (2022-10-27T16:13:48Z) - Generalizing Bayesian Optimization with Decision-theoretic Entropies [102.82152945324381]
We consider a generalization of Shannon entropy from work in statistical decision theory.
We first show that special cases of this entropy lead to popular acquisition functions used in BO procedures.
We then show how alternative choices for the loss yield a flexible family of acquisition functions.
arXiv Detail & Related papers (2022-10-04T04:43:58Z) - Predict+Optimize for Packing and Covering LPs with Unknown Parameters in
Constraints [5.762370982168012]
We propose a novel and practically relevant framework for the Predict+ setting, but with unknown parameters in both the objective and the constraints.
We introduce the notion of a correction function, and an additional penalty term in the loss function, modelling practical scenarios where an estimated optimal solution can be modified into a feasible solution after the true parameters are revealed.
Our approach is inspired by the prior work of Mandi and Guns, though with crucial modifications and re-derivations for our very different setting.
arXiv Detail & Related papers (2022-09-08T09:28:24Z) - Learning MDPs from Features: Predict-Then-Optimize for Sequential
Decision Problems by Reinforcement Learning [52.74071439183113]
We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) solved via reinforcement learning.
Two significant computational challenges arise in applying decision-focused learning to MDPs.
arXiv Detail & Related papers (2021-06-06T23:53:31Z) - Decision Trees for Decision-Making under the Predict-then-Optimize
Framework [7.842152902652215]
We consider the use of decision trees for decision-making problems under the predict-then-optimize framework.
A natural loss function in this framework is to measure the suboptimality of the decisions induced by the predicted input parameters.
We propose a tractable methodology called SPO Trees (SPOTs) for training decision trees under this loss.
arXiv Detail & Related papers (2020-02-29T23:04:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.