Deep learning for gradient flows using the Brezis-Ekeland principle
- URL: http://arxiv.org/abs/2209.14115v1
- Date: Wed, 28 Sep 2022 14:06:32 GMT
- Title: Deep learning for gradient flows using the Brezis-Ekeland principle
- Authors: Laura Carini, Max Jensen, Robert N\"urnberg
- Abstract summary: We propose a deep learning method for the numerical solution of partial differential equations that arise as gradient flows.
The method relies on the Brezis--Ekeland principle, which naturally defines an objective function to be minimized.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a deep learning method for the numerical solution of partial
differential equations that arise as gradient flows. The method relies on the
Brezis--Ekeland principle, which naturally defines an objective function to be
minimized, and so is ideally suited for a machine learning approach using deep
neural networks. We describe our approach in a general framework and illustrate
the method with the help of an example implementation for the heat equation in
space dimensions two to seven.
Related papers
- Practical Aspects on Solving Differential Equations Using Deep Learning: A Primer [0.0]
This primer aims to provide technical and practical insights into the Deep Galerkin method.
We demonstrate how to solve the one-dimensional heat equation step-by-step.
We also show how to apply the Deep Galerkin method to solve systems of ordinary differential equations and integral equations.
arXiv Detail & Related papers (2024-08-21T01:34:20Z) - Forward Gradient-Based Frank-Wolfe Optimization for Memory Efficient Deep Neural Network Training [0.0]
This paper focuses on analyzing the performance of the well-known Frank-Wolfe algorithm.
We show the proposed algorithm does converge to the optimal solution with a sub-linear rate of convergence.
In contrast, the standard Frank-Wolfe algorithm, when provided with access to the Projected Forward Gradient, fails to converge to the optimal solution.
arXiv Detail & Related papers (2024-03-19T07:25:36Z) - Stochastic Gradient Descent for Gaussian Processes Done Right [86.83678041846971]
We show that when emphdone right -- by which we mean using specific insights from optimisation and kernel communities -- gradient descent is highly effective.
We introduce a emphstochastic dual descent algorithm, explain its design in an intuitive manner and illustrate the design choices.
Our method places Gaussian process regression on par with state-of-the-art graph neural networks for molecular binding affinity prediction.
arXiv Detail & Related papers (2023-10-31T16:15:13Z) - $r-$Adaptive Deep Learning Method for Solving Partial Differential
Equations [0.685316573653194]
We introduce an $r-$adaptive algorithm to solve Partial Differential Equations using a Deep Neural Network.
The proposed method restricts to tensor product meshes and optimize the boundary node locations in one dimension, from which we build two- or three-dimensional meshes.
arXiv Detail & Related papers (2022-10-19T21:38:46Z) - An application of the splitting-up method for the computation of a
neural network representation for the solution for the filtering equations [68.8204255655161]
Filtering equations play a central role in many real-life applications, including numerical weather prediction, finance and engineering.
One of the classical approaches to approximate the solution of the filtering equations is to use a PDE inspired method, called the splitting-up method.
We combine this method with a neural network representation to produce an approximation of the unnormalised conditional distribution of the signal process.
arXiv Detail & Related papers (2022-01-10T11:01:36Z) - Deep Learning Approximation of Diffeomorphisms via Linear-Control
Systems [91.3755431537592]
We consider a control system of the form $dot x = sum_i=1lF_i(x)u_i$, with linear dependence in the controls.
We use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points.
arXiv Detail & Related papers (2021-10-24T08:57:46Z) - Learning Linearized Assignment Flows for Image Labeling [70.540936204654]
We introduce a novel algorithm for estimating optimal parameters of linearized assignment flows for image labeling.
We show how to efficiently evaluate this formula using a Krylov subspace and a low-rank approximation.
arXiv Detail & Related papers (2021-08-02T13:38:09Z) - Galerkin Neural Networks: A Framework for Approximating Variational
Equations with Error Control [0.0]
We present a new approach to using neural networks to approximate the solutions of variational equations.
We use a sequence of finite-dimensional subspaces whose basis functions are realizations of a sequence of neural networks.
arXiv Detail & Related papers (2021-05-28T20:25:40Z) - Deep learning: a statistical viewpoint [120.94133818355645]
Deep learning has revealed some major surprises from a theoretical perspective.
In particular, simple gradient methods easily find near-perfect solutions to non-optimal training problems.
We conjecture that specific principles underlie these phenomena.
arXiv Detail & Related papers (2021-03-16T16:26:36Z) - Interpolation Technique to Speed Up Gradients Propagation in Neural ODEs [71.26657499537366]
We propose a simple literature-based method for the efficient approximation of gradients in neural ODE models.
We compare it with the reverse dynamic method to train neural ODEs on classification, density estimation, and inference approximation tasks.
arXiv Detail & Related papers (2020-03-11T13:15:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.