Multigrid-augmented deep learning preconditioners for the Helmholtz
equation
- URL: http://arxiv.org/abs/2203.11025v1
- Date: Mon, 14 Mar 2022 10:31:11 GMT
- Title: Multigrid-augmented deep learning preconditioners for the Helmholtz
equation
- Authors: Yael Azulay and Eran Treister
- Abstract summary: We present a data-driven approach to solve the discrete heterogeneous Helmholtz equation at high wavenumbers.
We combine classical iterative solvers with convolutional neural networks (CNNs) to form a preconditioner which is applied within a Krylov solver.
- Score: 4.18804572788063
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we present a data-driven approach to iteratively solve the
discrete heterogeneous Helmholtz equation at high wavenumbers. In our approach,
we combine classical iterative solvers with convolutional neural networks
(CNNs) to form a preconditioner which is applied within a Krylov solver. For
the preconditioner, we use a CNN of type U-Net that operates in conjunction
with multigrid ingredients. Two types of preconditioners are proposed 1) U-Net
as a coarse grid solver, and 2) U-Net as a deflation operator with shifted
Laplacian V-cycles. Following our training scheme and data-augmentation, our
CNN preconditioner can generalize over residuals and a relatively general set
of wave slowness models. On top of that, we also offer an encoder-solver
framework where an "encoder" network generalizes over the medium and sends
context vectors to another "solver" network, which generalizes over the
right-hand-sides. We show that this option is more robust and efficient than
the stand-alone variant. Lastly, we also offer a mini-retraining procedure, to
improve the solver after the model is known. This option is beneficial when
solving multiple right-hand-sides, like in inverse problems. We demonstrate the
efficiency and generalization abilities of our approach on a variety of 2D
problems.
Related papers
- LinSATNet: The Positive Linear Satisfiability Neural Networks [116.65291739666303]
This paper studies how to introduce the popular positive linear satisfiability to neural networks.
We propose the first differentiable satisfiability layer based on an extension of the classic Sinkhorn algorithm for jointly encoding multiple sets of marginal distributions.
arXiv Detail & Related papers (2024-07-18T22:05:21Z) - Physics-guided Full Waveform Inversion using Encoder-Solver Convolutional Neural Networks [7.56372030029358]
Full Waveform Inversion (FWI) is an inverse problem for estimating the wave velocity distribution in a given domain.
We develop a learning process of an encoder-solver preconditioner that is based on convolutional neural networks.
We demonstrate our approach to solving FWI problems using 2D geophysical models with high-frequency data.
arXiv Detail & Related papers (2024-05-27T23:03:21Z) - Learning from Linear Algebra: A Graph Neural Network Approach to Preconditioner Design for Conjugate Gradient Solvers [42.69799418639716]
Deep learning models may be used to precondition residuals during iteration of such linear solvers as the conjugate gradient (CG) method.
Neural network models require an enormous number of parameters to approximate well in this setup.
In our work, we recall well-established preconditioners from linear algebra and use them as a starting point for training the GNN.
arXiv Detail & Related papers (2024-05-24T13:44:30Z) - Multigrid-Augmented Deep Learning Preconditioners for the Helmholtz
Equation using Compact Implicit Layers [7.56372030029358]
We present a deep learning-based iterative approach to solve the discrete heterogeneous Helmholtz equation for high wavenumbers.
We construct a multilevel U-Net-like encoder-solver CNN with an implicit layer on the coarsest grid of the U-Net, where convolution kernels are inverted.
Our architecture can be used to generalize over different slowness models of various difficulties and is efficient at solving for many right-hand sides per slowness model.
arXiv Detail & Related papers (2023-06-30T08:56:51Z) - Sampling binary sparse coding QUBO models using a spiking neuromorphic
processor [3.0586855806896045]
We consider the problem of computing a binary representation of an image.
We aim to find a binary vector minimal set of basis that when added together best reconstruct the given input.
This yields a so-called Quadratic Unconstrained Binary (QUBO) problem.
arXiv Detail & Related papers (2023-06-02T22:47:18Z) - Learning To Dive In Branch And Bound [95.13209326119153]
We propose L2Dive to learn specific diving structurals with graph neural networks.
We train generative models to predict variable assignments and leverage the duality of linear programs to make diving decisions.
arXiv Detail & Related papers (2023-01-24T12:01:45Z) - Neural Basis Functions for Accelerating Solutions to High Mach Euler
Equations [63.8376359764052]
We propose an approach to solving partial differential equations (PDEs) using a set of neural networks.
We regress a set of neural networks onto a reduced order Proper Orthogonal Decomposition (POD) basis.
These networks are then used in combination with a branch network that ingests the parameters of the prescribed PDE to compute a reduced order approximation to the PDE.
arXiv Detail & Related papers (2022-08-02T18:27:13Z) - Learning the Solution Operator of Boundary Value Problems using Graph
Neural Networks [0.0]
We design a general solution operator for two different time-independent PDEs using graph neural networks (GNNs) and spectral graph convolutions.
We train the networks on simulated data from a finite elements solver on a variety of shapes and inhomogeneities.
We find that training on a diverse dataset with lots of variation in the finite element meshes is a key ingredient for achieving good generalization results.
arXiv Detail & Related papers (2022-06-28T15:39:06Z) - Meta-Solver for Neural Ordinary Differential Equations [77.8918415523446]
We investigate how the variability in solvers' space can improve neural ODEs performance.
We show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks.
arXiv Detail & Related papers (2021-03-15T17:26:34Z) - A Flexible Framework for Designing Trainable Priors with Adaptive
Smoothing and Game Encoding [57.1077544780653]
We introduce a general framework for designing and training neural network layers whose forward passes can be interpreted as solving non-smooth convex optimization problems.
We focus on convex games, solved by local agents represented by the nodes of a graph and interacting through regularization functions.
This approach is appealing for solving imaging problems, as it allows the use of classical image priors within deep models that are trainable end to end.
arXiv Detail & Related papers (2020-06-26T08:34:54Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.