Optimization-Informed Neural Networks
- URL: http://arxiv.org/abs/2210.02113v3
- Date: Sun, 25 Jun 2023 14:34:28 GMT
- Title: Optimization-Informed Neural Networks
- Authors: Dawen Wu, Abdel Lisser
- Abstract summary: We propose optimization-informed neural networks (OINN) to solve constrained nonlinear optimization problems.
In a nutshell, OINN transforms a CNLP into a neural network training problem.
The effectiveness of the proposed approach is demonstrated through a collection of classical problems.
- Score: 0.6853165736531939
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Solving constrained nonlinear optimization problems (CNLPs) is a longstanding
problem that arises in various fields, e.g., economics, computer science, and
engineering. We propose optimization-informed neural networks (OINN), a deep
learning approach to solve CNLPs. By neurodynamic optimization methods, a CNLP
is first reformulated as an initial value problem (IVP) involving an ordinary
differential equation (ODE) system. A neural network model is then used as an
approximate solution for this IVP, with the endpoint being the prediction to
the CNLP. We propose a novel training algorithm that directs the model to hold
the best prediction during training. In a nutshell, OINN transforms a CNLP into
a neural network training problem. By doing so, we can solve CNLPs based on
deep learning infrastructure only, without using standard optimization solvers
or numerical integration solvers. The effectiveness of the proposed approach is
demonstrated through a collection of classical problems, e.g., variational
inequalities, nonlinear complementary problems, and standard CNLPs.
Related papers
- Model Based and Physics Informed Deep Learning Neural Network Structures [7.095119621199481]
Neural Networks (NN) has been used in many areas with great success.
One of the great difficulties is the choice of the NN's structure.
We consider this problem using model based signal and image processing and inverse problems methods.
arXiv Detail & Related papers (2024-08-13T07:28:38Z) - A Stable and Scalable Method for Solving Initial Value PDEs with Neural
Networks [52.5899851000193]
We develop an ODE based IVP solver which prevents the network from getting ill-conditioned and runs in time linear in the number of parameters.
We show that current methods based on this approach suffer from two key issues.
First, following the ODE produces an uncontrolled growth in the conditioning of the problem, ultimately leading to unacceptably large numerical errors.
arXiv Detail & Related papers (2023-04-28T17:28:18Z) - Neural Combinatorial Optimization: a New Player in the Field [69.23334811890919]
This paper presents a critical analysis on the incorporation of algorithms based on neural networks into the classical optimization framework.
A comprehensive study is carried out to analyse the fundamental aspects of such algorithms, including performance, transferability, computational cost and to larger-sized instances.
arXiv Detail & Related papers (2022-05-03T07:54:56Z) - An alternative approach to train neural networks using monotone
variational inequality [22.320632565424745]
We propose an alternative approach to neural network training using the monotone vector field.
Our approach can be used for more efficient fine-tuning of a pre-trained neural network.
arXiv Detail & Related papers (2022-02-17T19:24:20Z) - Adaptive neural domain refinement for solving time-dependent
differential equations [0.0]
A classic approach for solving differential equations with neural networks builds upon neural forms, which employ the differential equation with a discretisation of the solution domain.
It would be desirable to transfer such important and successful strategies to the field of neural network based solutions.
We propose a novel adaptive neural approach to meet this aim for solving time-dependent problems.
arXiv Detail & Related papers (2021-12-23T13:19:07Z) - Acceleration techniques for optimization over trained neural network
ensembles [1.0323063834827415]
We study optimization problems where the objective function is modeled through feedforward neural networks with rectified linear unit activation.
We present a mixed-integer linear program based on existing popular big-$M$ formulations for optimizing over a single neural network.
arXiv Detail & Related papers (2021-12-13T20:50:54Z) - A Meta-Learning Approach to the Optimal Power Flow Problem Under
Topology Reconfigurations [69.73803123972297]
We propose a DNN-based OPF predictor that is trained using a meta-learning (MTL) approach.
The developed OPF-predictor is validated through simulations using benchmark IEEE bus systems.
arXiv Detail & Related papers (2020-12-21T17:39:51Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Resource Allocation via Graph Neural Networks in Free Space Optical
Fronthaul Networks [119.81868223344173]
This paper investigates the optimal resource allocation in free space optical (FSO) fronthaul networks.
We consider the graph neural network (GNN) for the policy parameterization to exploit the FSO network structure.
The primal-dual learning algorithm is developed to train the GNN in a model-free manner, where the knowledge of system models is not required.
arXiv Detail & Related papers (2020-06-26T14:20:48Z) - ODEN: A Framework to Solve Ordinary Differential Equations using
Artificial Neural Networks [0.0]
We prove a specific loss function, which does not require knowledge of the exact solution, to evaluate neural networks' performance.
Neural networks are shown to be proficient at approximating continuous solutions within their training domains.
A user-friendly and adaptable open-source code (ODE$mathcalN$) is provided on GitHub.
arXiv Detail & Related papers (2020-05-28T15:34:10Z) - Self-Directed Online Machine Learning for Topology Optimization [58.920693413667216]
Self-directed Online Learning Optimization integrates Deep Neural Network (DNN) with Finite Element Method (FEM) calculations.
Our algorithm was tested by four types of problems including compliance minimization, fluid-structure optimization, heat transfer enhancement and truss optimization.
It reduced the computational time by 2 5 orders of magnitude compared with directly using methods, and outperformed all state-of-the-art algorithms tested in our experiments.
arXiv Detail & Related papers (2020-02-04T20:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.