Multigrid-Augmented Deep Learning Preconditioners for the Helmholtz
Equation using Compact Implicit Layers
- URL: http://arxiv.org/abs/2306.17486v3
- Date: Wed, 6 Mar 2024 17:19:09 GMT
- Title: Multigrid-Augmented Deep Learning Preconditioners for the Helmholtz
Equation using Compact Implicit Layers
- Authors: Bar Lerer, Ido Ben-Yair and Eran Treister
- Abstract summary: We present a deep learning-based iterative approach to solve the discrete heterogeneous Helmholtz equation for high wavenumbers.
We construct a multilevel U-Net-like encoder-solver CNN with an implicit layer on the coarsest grid of the U-Net, where convolution kernels are inverted.
Our architecture can be used to generalize over different slowness models of various difficulties and is efficient at solving for many right-hand sides per slowness model.
- Score: 7.56372030029358
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a deep learning-based iterative approach to solve the discrete
heterogeneous Helmholtz equation for high wavenumbers. Combining classical
iterative multigrid solvers and convolutional neural networks (CNNs) via
preconditioning, we obtain a learned neural solver that is faster and scales
better than a standard multigrid solver. Our approach offers three main
contributions over previous neural methods of this kind. First, we construct a
multilevel U-Net-like encoder-solver CNN with an implicit layer on the coarsest
grid of the U-Net, where convolution kernels are inverted. This alleviates the
field of view problem in CNNs and allows better scalability. Second, we improve
upon the previous CNN preconditioner in terms of the number of parameters,
computation time, and convergence rates. Third, we propose a multiscale
training approach that enables the network to scale to problems of previously
unseen dimensions while still maintaining a reasonable training procedure. Our
encoder-solver architecture can be used to generalize over different slowness
models of various difficulties and is efficient at solving for many right-hand
sides per slowness model. We demonstrate the benefits of our novel architecture
with numerical experiments on a variety of heterogeneous two-dimensional
problems at high wavenumbers.
Related papers
- LinSATNet: The Positive Linear Satisfiability Neural Networks [116.65291739666303]
This paper studies how to introduce the popular positive linear satisfiability to neural networks.
We propose the first differentiable satisfiability layer based on an extension of the classic Sinkhorn algorithm for jointly encoding multiple sets of marginal distributions.
arXiv Detail & Related papers (2024-07-18T22:05:21Z) - Solving the Discretised Multiphase Flow Equations with Interface
Capturing on Structured Grids Using Machine Learning Libraries [0.6299766708197884]
This paper solves the discretised multiphase flow equations using tools and methods from machine-learning libraries.
For the first time, finite element discretisations of multiphase flows can be solved using an approach based on (untrained) convolutional neural networks.
arXiv Detail & Related papers (2024-01-12T18:42:42Z) - Neural Basis Functions for Accelerating Solutions to High Mach Euler
Equations [63.8376359764052]
We propose an approach to solving partial differential equations (PDEs) using a set of neural networks.
We regress a set of neural networks onto a reduced order Proper Orthogonal Decomposition (POD) basis.
These networks are then used in combination with a branch network that ingests the parameters of the prescribed PDE to compute a reduced order approximation to the PDE.
arXiv Detail & Related papers (2022-08-02T18:27:13Z) - Multigrid-augmented deep learning preconditioners for the Helmholtz
equation [4.18804572788063]
We present a data-driven approach to solve the discrete heterogeneous Helmholtz equation at high wavenumbers.
We combine classical iterative solvers with convolutional neural networks (CNNs) to form a preconditioner which is applied within a Krylov solver.
arXiv Detail & Related papers (2022-03-14T10:31:11Z) - Distributed Multigrid Neural Solvers on Megavoxel Domains [27.412837974378597]
We consider distributed training of PDE solvers producing full field outputs.
A scalable framework is presented that integrates two distinct advances.
This approach is deployed to train a generalized 3D Poisson solver that scales well to predict output full-field solutions.
arXiv Detail & Related papers (2021-04-29T17:53:22Z) - Meta-Solver for Neural Ordinary Differential Equations [77.8918415523446]
We investigate how the variability in solvers' space can improve neural ODEs performance.
We show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks.
arXiv Detail & Related papers (2021-03-15T17:26:34Z) - Learning optimal multigrid smoothers via neural networks [1.9336815376402723]
We propose an efficient framework for learning optimized smoothers from operator stencils in the form of convolutional neural networks (CNNs)
CNNs are trained on small-scale problems from a given type of PDEs based on a supervised loss function derived from multigrid convergence theories.
Numerical results on anisotropic rotated Laplacian problems demonstrate improved convergence rates and solution time compared with classical hand-crafted relaxation methods.
arXiv Detail & Related papers (2021-02-24T05:02:54Z) - Implicit Convex Regularizers of CNN Architectures: Convex Optimization
of Two- and Three-Layer Networks in Polynomial Time [70.15611146583068]
We study training of Convolutional Neural Networks (CNNs) with ReLU activations.
We introduce exact convex optimization with a complexity with respect to the number of data samples, the number of neurons, and data dimension.
arXiv Detail & Related papers (2020-06-26T04:47:20Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.