Learning Optimal Solutions via an LSTM-Optimization Framework
- URL: http://arxiv.org/abs/2207.02937v1
- Date: Wed, 6 Jul 2022 19:38:01 GMT
- Title: Learning Optimal Solutions via an LSTM-Optimization Framework
- Authors: Dogacan Yilmaz, \.I. Esra B\"uy\"uktahtak{\i}n
- Abstract summary: We present a deep learning-optimization framework to tackle dynamic mixed-integer programs.
We develop a bidirectional Long Short Term Memory (LSTM) framework that can process information forward and backward in time.
We demonstrate our approach in predicting the optimal decisions for the single-item capacitated lot-sizing problem.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this study, we present a deep learning-optimization framework to tackle
dynamic mixed-integer programs. Specifically, we develop a bidirectional Long
Short Term Memory (LSTM) framework that can process information forward and
backward in time to learn optimal solutions to sequential decision-making
problems. We demonstrate our approach in predicting the optimal decisions for
the single-item capacitated lot-sizing problem (CLSP), where a binary variable
denotes whether to produce in a period or not. Due to the dynamic nature of the
problem, the CLSP can be treated as a sequence labeling task where a recurrent
neural network can capture the problem's temporal dynamics. Computational
results show that our LSTM-Optimization (LSTM-Opt) framework significantly
reduces the solution time of benchmark CLSP problems without much loss in
feasibility and optimality. For example, the predictions at the 85\% level
reduce the CPLEX solution time by a factor of 9 on average for over 240,000
test instances with an optimality gap of less than 0.05\% and 0.4\%
infeasibility in the test set. Also, models trained using shorter planning
horizons can successfully predict the optimal solution of the instances with
longer planning horizons. For the hardest data set, the LSTM predictions at the
25\% level reduce the solution time of 70 CPU hours to less than 2 CPU minutes
with an optimality gap of 0.8\% and without any infeasibility. The LSTM-Opt
framework outperforms classical ML algorithms, such as the logistic regression
and random forest, in terms of the solution quality, and exact approaches, such
as the ($\ell$, S) and dynamic programming-based inequalities, with respect to
the solution time improvement. Our machine learning approach could be
beneficial in tackling sequential decision-making problems similar to CLSP,
which need to be solved repetitively, frequently, and in a fast manner.
Related papers
- OptiBench Meets ReSocratic: Measure and Improve LLMs for Optimization Modeling [62.19438812624467]
Large language models (LLMs) have exhibited their problem-solving abilities in mathematical reasoning.
We propose OptiBench, a benchmark for End-to-end optimization problem-solving with human-readable inputs and outputs.
arXiv Detail & Related papers (2024-07-13T13:27:57Z) - An Expandable Machine Learning-Optimization Framework to Sequential
Decision-Making [0.0]
We present an integrated prediction-optimization (PredOpt) framework to efficiently solve sequential decision-making problems.
We address the key issues of sequential dependence, infeasibility, and generalization in machine learning (ML) to make predictions for optimal solutions to instances problems.
arXiv Detail & Related papers (2023-11-12T21:54:53Z) - Training Latency Minimization for Model-Splitting Allowed Federated Edge
Learning [16.8717239856441]
We propose a model-splitting allowed FL (SFL) framework to alleviate the shortage of computing power faced by clients in training deep neural networks (DNNs) using federated learning (FL)
Under the synchronized global update setting, the latency to complete a round of global training is determined by the maximum latency for the clients to complete a local training session.
To solve this mixed integer nonlinear programming problem, we first propose a regression method to fit the quantitative-relationship between the cut-layer and other parameters of an AI-model, and thus, transform the TLMP into a continuous problem.
arXiv Detail & Related papers (2023-07-21T12:26:42Z) - M-L2O: Towards Generalizable Learning-to-Optimize by Test-Time Fast
Self-Adaptation [145.7321032755538]
Learning to Optimize (L2O) has drawn increasing attention as it often remarkably accelerates the optimization procedure of complex tasks.
This paper investigates a potential solution to this open challenge by meta-training an L2O that can perform fast test-time self-adaptation to an out-of-distribution task.
arXiv Detail & Related papers (2023-02-28T19:23:20Z) - Learning to repeatedly solve routing problems [5.08128537391027]
We present a learned for the reoptimization of a problem after a minor change in its data.
Given the edges of an original solution, the goal is to predict and fix the ones that have a high chance of remaining in an optimal solution.
This partial prediction of the solution reduces the complexity of the problem and speeds up its resolution.
arXiv Detail & Related papers (2022-12-15T19:33:54Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - Fast Continuous and Integer L-shaped Heuristics Through Supervised
Learning [4.521119623956821]
We propose a methodology to accelerate the solution of mixed-integer linear two-stage programs.
We aim at solving problems where the second stage is highly demanding.
Our core idea is to gain large reductions in online solution time while incurring small reductions in first-stage solution accuracy.
arXiv Detail & Related papers (2022-05-02T13:15:32Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z) - Optimizing Wireless Systems Using Unsupervised and
Reinforced-Unsupervised Deep Learning [96.01176486957226]
Resource allocation and transceivers in wireless networks are usually designed by solving optimization problems.
In this article, we introduce unsupervised and reinforced-unsupervised learning frameworks for solving both variable and functional optimization problems.
arXiv Detail & Related papers (2020-01-03T11:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.