DC3: A learning method for optimization with hard constraints
- URL: http://arxiv.org/abs/2104.12225v1
- Date: Sun, 25 Apr 2021 18:21:59 GMT
- Title: DC3: A learning method for optimization with hard constraints
- Authors: Priya L. Donti, David Rolnick, J. Zico Kolter
- Abstract summary: We present Deep Constraint Completion and Correction (DC3), an algorithm to address this challenge.
DC3 implicitly completes partial solutions to satisfy equality constraints and unrolls-based corrections to satisfy inequality constraints.
We demonstrate the effectiveness of DC3 in both synthetic optimization tasks and the real-world setting of AC optimal power flow.
- Score: 85.12291213315905
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large optimization problems with hard constraints arise in many settings, yet
classical solvers are often prohibitively slow, motivating the use of deep
networks as cheap "approximate solvers." Unfortunately, naive deep learning
approaches typically cannot enforce the hard constraints of such problems,
leading to infeasible solutions. In this work, we present Deep Constraint
Completion and Correction (DC3), an algorithm to address this challenge.
Specifically, this method enforces feasibility via a differentiable procedure,
which implicitly completes partial solutions to satisfy equality constraints
and unrolls gradient-based corrections to satisfy inequality constraints. We
demonstrate the effectiveness of DC3 in both synthetic optimization tasks and
the real-world setting of AC optimal power flow, where hard constraints encode
the physics of the electrical grid. In both cases, DC3 achieves near-optimal
objective values while preserving feasibility.
Related papers
- WANCO: Weak Adversarial Networks for Constrained Optimization problems [5.257895611010853]
We first transform minimax problems into minimax problems using the augmented Lagrangian method.
We then use two (or several) deep neural networks to represent the primal and dual variables respectively.
The parameters in the neural networks are then trained by an adversarial process.
arXiv Detail & Related papers (2024-07-04T05:37:48Z) - Two-Stage ML-Guided Decision Rules for Sequential Decision Making under Uncertainty [55.06411438416805]
Sequential Decision Making under Uncertainty (SDMU) is ubiquitous in many domains such as energy, finance, and supply chains.
Some SDMU are naturally modeled as Multistage Problems (MSPs) but the resulting optimizations are notoriously challenging from a computational standpoint.
This paper introduces a novel approach Two-Stage General Decision Rules (TS-GDR) to generalize the policy space beyond linear functions.
The effectiveness of TS-GDR is demonstrated through an instantiation using Deep Recurrent Neural Networks named Two-Stage Deep Decision Rules (TS-LDR)
arXiv Detail & Related papers (2024-05-23T18:19:47Z) - Learning Constrained Optimization with Deep Augmented Lagrangian Methods [54.22290715244502]
A machine learning (ML) model is trained to emulate a constrained optimization solver.
This paper proposes an alternative approach, in which the ML model is trained to predict dual solution estimates directly.
It enables an end-to-end training scheme is which the dual objective is as a loss function, and solution estimates toward primal feasibility, emulating a Dual Ascent method.
arXiv Detail & Related papers (2024-03-06T04:43:22Z) - FlowPG: Action-constrained Policy Gradient with Normalizing Flows [14.98383953401637]
Action-constrained reinforcement learning (ACRL) is a popular approach for solving safety-critical resource-alential related decision making problems.
A major challenge in ACRL is to ensure agent taking a valid action satisfying constraints in each step.
arXiv Detail & Related papers (2024-02-07T11:11:46Z) - Self-supervised Equality Embedded Deep Lagrange Dual for Approximate Constrained Optimization [5.412875005737914]
We propose deeprange dual embedding (DeepLDE) as a fast optimal approximators incorporating neural networks (NNs)
We prove the convergence of DeepLDE and primal, nondual-learning method to impose inequality constraints.
We show that the proposed DeepLDE solutions achieve the optimal Lag gap among all the smallest NN-based approaches.
arXiv Detail & Related papers (2023-06-11T13:19:37Z) - An Efficient Learning-Based Solver for Two-Stage DC Optimal Power Flow with Feasibility Guarantees [4.029937264494929]
This paper proposes a learning method to solve the two-stage problem in a more efficient and optimal way.
A technique called the gauge map is incorporated into the learning architecture design to guarantee the learned solutions' feasibility to the network constraints.
arXiv Detail & Related papers (2023-04-03T22:56:08Z) - Adversarially Robust Learning for Security-Constrained Optimal Power
Flow [55.816266355623085]
We tackle the problem of N-k security-constrained optimal power flow (SCOPF)
N-k SCOPF is a core problem for the operation of electrical grids.
Inspired by methods in adversarially robust training, we frame N-k SCOPF as a minimax optimization problem.
arXiv Detail & Related papers (2021-11-12T22:08:10Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.