UNIFY: a Unified Policy Designing Framework for Solving Constrained
Optimization Problems with Machine Learning
- URL: http://arxiv.org/abs/2210.14030v1
- Date: Tue, 25 Oct 2022 14:09:24 GMT
- Title: UNIFY: a Unified Policy Designing Framework for Solving Constrained
Optimization Problems with Machine Learning
- Authors: Mattia Silvestri, Allegra De Filippo, Michele Lombardi, Michela Milano
- Abstract summary: We propose a unified framework to design a solution policy for complex decision-making problems.
Our approach relies on a clever decomposition of the policy in two stages, namely an unconstrained ML model and a CO problem.
We demonstrate the method effectiveness on two practical problems, namely an Energy Management System and the Set Multi-cover with coverage requirements.
- Score: 18.183339583346005
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The interplay between Machine Learning (ML) and Constrained Optimization (CO)
has recently been the subject of increasing interest, leading to a new and
prolific research area covering (e.g.) Decision Focused Learning and
Constrained Reinforcement Learning. Such approaches strive to tackle complex
decision problems under uncertainty over multiple stages, involving both
explicit (cost function, constraints) and implicit knowledge (from data), and
possibly subject to execution time restrictions. While a good degree of success
has been achieved, the existing methods still have limitations in terms of both
applicability and effectiveness. For problems in this class, we propose UNIFY,
a unified framework to design a solution policy for complex decision-making
problems. Our approach relies on a clever decomposition of the policy in two
stages, namely an unconstrained ML model and a CO problem, to take advantage of
the strength of each approach while compensating for its weaknesses. With a
little design effort, UNIFY can generalize several existing approaches, thus
extending their applicability. We demonstrate the method effectiveness on two
practical problems, namely an Energy Management System and the Set Multi-cover
with stochastic coverage requirements. Finally, we highlight some current
challenges of our method and future research directions that can benefit from
the cross-fertilization of the two fields.
Related papers
- BloomWise: Enhancing Problem-Solving capabilities of Large Language Models using Bloom's-Taxonomy-Inspired Prompts [59.83547898874152]
We introduce BloomWise, a new prompting technique, inspired by Bloom's taxonomy, to improve the performance of Large Language Models (LLMs)
The decision regarding the need to employ more sophisticated cognitive skills is based on self-evaluation performed by the LLM.
In extensive experiments across 4 popular math reasoning datasets, we have demonstrated the effectiveness of our proposed approach.
arXiv Detail & Related papers (2024-10-05T09:27:52Z) - Safe and Balanced: A Framework for Constrained Multi-Objective Reinforcement Learning [26.244121960815907]
We propose a primal-based framework that orchestrates policy optimization between multi-objective learning and constraint adherence.
Our method employs a novel natural policy gradient manipulation method to optimize multiple RL objectives.
Empirically, our proposed method also outperforms prior state-of-the-art methods on challenging safe multi-objective reinforcement learning tasks.
arXiv Detail & Related papers (2024-05-26T00:42:10Z) - Near-Optimal Solutions of Constrained Learning Problems [85.48853063302764]
In machine learning systems, the need to curtail their behavior has become increasingly apparent.
This is evidenced by recent advancements towards developing models that satisfy dual robustness variables.
Our results show that rich parametrizations effectively mitigate non-dimensional, finite learning problems.
arXiv Detail & Related papers (2024-03-18T14:55:45Z) - Intrinsically Motivated Hierarchical Policy Learning in Multi-objective
Markov Decision Processes [15.50007257943931]
We propose a novel dual-phase intrinsically motivated reinforcement learning method to address this limitation.
We show experimentally that the proposed method significantly outperforms state-of-the-art multi-objective reinforcement methods in a dynamic robotics environment.
arXiv Detail & Related papers (2023-08-18T02:10:45Z) - Learning with Constraint Learning: New Perspective, Solution Strategy
and Various Applications [45.45917703420217]
We propose a new framework, named Learning with Constraint Learning (LwCL), that can holistically examine challenges.
LwCL is designed as a general hierarchical optimization model that captures the essence of diverse learning and vision problems.
Our proposed framework efficiently addresses a wide range of applications in learning and vision, encompassing three categories and nine different problem types.
arXiv Detail & Related papers (2023-07-28T01:50:27Z) - An Efficient Learning-Based Solver for Two-Stage DC Optimal Power Flow with Feasibility Guarantees [4.029937264494929]
This paper proposes a learning method to solve the two-stage problem in a more efficient and optimal way.
A technique called the gauge map is incorporated into the learning architecture design to guarantee the learned solutions' feasibility to the network constraints.
arXiv Detail & Related papers (2023-04-03T22:56:08Z) - Reinforcement Learning with Stepwise Fairness Constraints [50.538878453547966]
We introduce the study of reinforcement learning with stepwise fairness constraints.
We provide learning algorithms with strong theoretical guarantees in regard to policy optimality and fairness violation.
arXiv Detail & Related papers (2022-11-08T04:06:23Z) - Efficient lifting of symmetry breaking constraints for complex
combinatorial problems [9.156939957189502]
This work extends the learning framework and implementation of a model-based approach for Answer Set Programming.
In particular, we incorporate a new conflict analysis algorithm in the Inductive Logic Programming system ILASP.
arXiv Detail & Related papers (2022-05-14T20:42:13Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Learning What to Defer for Maximum Independent Sets [84.00112106334655]
We propose a novel DRL scheme, coined learning what to defer (LwD), where the agent adaptively shrinks or stretch the number of stages by learning to distribute the element-wise decisions of the solution at each stage.
We apply the proposed framework to the maximum independent set (MIS) problem, and demonstrate its significant improvement over the current state-of-the-art DRL scheme.
arXiv Detail & Related papers (2020-06-17T02:19:31Z) - Probably Approximately Correct Constrained Learning [135.48447120228658]
We develop a generalization theory based on the probably approximately correct (PAC) learning framework.
We show that imposing a learner does not make a learning problem harder in the sense that any PAC learnable class is also a constrained learner.
We analyze the properties of this solution and use it to illustrate how constrained learning can address problems in fair and robust classification.
arXiv Detail & Related papers (2020-06-09T19:59:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.