ICLN: Input Convex Loss Network for Decision Focused Learning
- URL: http://arxiv.org/abs/2403.01875v1
- Date: Mon, 4 Mar 2024 09:31:56 GMT
- Title: ICLN: Input Convex Loss Network for Decision Focused Learning
- Authors: Haeun Jeon, Hyunglip Bae, Minsu Park, Chanyeong Kim, Woo Chang Kim
- Abstract summary: In decision-making problem under uncertainty, predicting unknown parameters is often considered independent of the optimization part.
We propose Input Convex Loss Network (ICLN), a novel global surrogate loss which can be implemented in a general DFL paradigm.
ICLN learns task loss via Input Convex Neural Networks which is guaranteed to be convex for some inputs, while keeping the global structure for the other inputs.
- Score: 0.562479170374811
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In decision-making problem under uncertainty, predicting unknown parameters
is often considered independent of the optimization part. Decision-focused
Learning (DFL) is a task-oriented framework to integrate prediction and
optimization by adapting predictive model to give better decision for the
corresponding task. Here, an inevitable challenge arises when computing
gradients of the optimal decision with respect to the parameters. Existing
researches cope this issue by smoothly reforming surrogate optimization or
construct surrogate loss function that mimic task loss. However, they are
applied to restricted optimization domain or build functions in a local manner
leading a large computational time. In this paper, we propose Input Convex Loss
Network (ICLN), a novel global surrogate loss which can be implemented in a
general DFL paradigm. ICLN learns task loss via Input Convex Neural Networks
which is guaranteed to be convex for some inputs, while keeping the global
structure for the other inputs. This enables ICLN to admit general DFL through
only a single surrogate loss without any sense for choosing appropriate
parametric forms. We confirm effectiveness and flexibility of ICLN by
evaluating our proposed model with three stochastic decision-making problems.
Related papers
- Two-Stage ML-Guided Decision Rules for Sequential Decision Making under Uncertainty [55.06411438416805]
Sequential Decision Making under Uncertainty (SDMU) is ubiquitous in many domains such as energy, finance, and supply chains.
Some SDMU are naturally modeled as Multistage Problems (MSPs) but the resulting optimizations are notoriously challenging from a computational standpoint.
This paper introduces a novel approach Two-Stage General Decision Rules (TS-GDR) to generalize the policy space beyond linear functions.
The effectiveness of TS-GDR is demonstrated through an instantiation using Deep Recurrent Neural Networks named Two-Stage Deep Decision Rules (TS-LDR)
arXiv Detail & Related papers (2024-05-23T18:19:47Z) - End-to-End Learning for Fair Multiobjective Optimization Under
Uncertainty [55.04219793298687]
The Predict-Then-Forecast (PtO) paradigm in machine learning aims to maximize downstream decision quality.
This paper extends the PtO methodology to optimization problems with nondifferentiable Ordered Weighted Averaging (OWA) objectives.
It shows how optimization of OWA functions can be effectively integrated with parametric prediction for fair and robust optimization under uncertainty.
arXiv Detail & Related papers (2024-02-12T16:33:35Z) - QCQP-Net: Reliably Learning Feasible Alternating Current Optimal Power
Flow Solutions Under Constraints [4.1920378271058425]
We propose an innovated computational learning ACOPF, where the input is mapped to the ACOPF network in a computationally efficient manner.
We show through simulations that our proposed method achieves superior feasibility rate and cost in situations where the existing-based approaches fail.
arXiv Detail & Related papers (2024-01-11T20:17:44Z) - A Multi-Head Ensemble Multi-Task Learning Approach for Dynamical
Computation Offloading [62.34538208323411]
We propose a multi-head ensemble multi-task learning (MEMTL) approach with a shared backbone and multiple prediction heads (PHs)
MEMTL outperforms benchmark methods in both the inference accuracy and mean square error without requiring additional training data.
arXiv Detail & Related papers (2023-09-02T11:01:16Z) - Training Latency Minimization for Model-Splitting Allowed Federated Edge
Learning [16.8717239856441]
We propose a model-splitting allowed FL (SFL) framework to alleviate the shortage of computing power faced by clients in training deep neural networks (DNNs) using federated learning (FL)
Under the synchronized global update setting, the latency to complete a round of global training is determined by the maximum latency for the clients to complete a local training session.
To solve this mixed integer nonlinear programming problem, we first propose a regression method to fit the quantitative-relationship between the cut-layer and other parameters of an AI-model, and thus, transform the TLMP into a continuous problem.
arXiv Detail & Related papers (2023-07-21T12:26:42Z) - Power Control with QoS Guarantees: A Differentiable Projection-based
Unsupervised Learning Framework [14.518558523319518]
Deep neural networks (DNNs) are emerging as a potential solution to solve NP-hard wireless resource allocation problems.
We propose a novel unsupervised learning framework to solve the classical power control problem in a multi-user channel.
We show that the proposed solutions not only improve the data rate but also achieve zero constraint violation probability, compared to the existing computations.
arXiv Detail & Related papers (2023-05-31T14:11:51Z) - Learning MDPs from Features: Predict-Then-Optimize for Sequential
Decision Problems by Reinforcement Learning [52.74071439183113]
We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) solved via reinforcement learning.
Two significant computational challenges arise in applying decision-focused learning to MDPs.
arXiv Detail & Related papers (2021-06-06T23:53:31Z) - Learning to Solve the AC-OPF using Sensitivity-Informed Deep Neural
Networks [52.32646357164739]
We propose a deep neural network (DNN) to solve the solutions of the optimal power flow (ACOPF)
The proposed SIDNN is compatible with a broad range of OPF schemes.
It can be seamlessly integrated in other learning-to-OPF schemes.
arXiv Detail & Related papers (2021-03-27T00:45:23Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.