A Deep Gradient Correction Method for Iteratively Solving Linear Systems
- URL: http://arxiv.org/abs/2205.10763v1
- Date: Sun, 22 May 2022 06:40:38 GMT
- Title: A Deep Gradient Correction Method for Iteratively Solving Linear Systems
- Authors: Ayano Kaneda, Osman Akar, Jingyu Chen, Victoria Kala, David Hyde,
Joseph Teran
- Abstract summary: We present a novel approach to approximate the solution of large, sparse, symmetric, positive-definite linear systems of equations.
Our algorithm is capable of reducing the linear system residual to a given tolerance in a small number of iterations.
- Score: 5.744903762364991
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a novel deep learning approach to approximate the solution of
large, sparse, symmetric, positive-definite linear systems of equations. These
systems arise from many problems in applied science, e.g., in numerical methods
for partial differential equations. Algorithms for approximating the solution
to these systems are often the bottleneck in problems that require their
solution, particularly for modern applications that require many millions of
unknowns. Indeed, numerical linear algebra techniques have been investigated
for many decades to alleviate this computational burden. Recently, data-driven
techniques have also shown promise for these problems. Motivated by the
conjugate gradients algorithm that iteratively selects search directions for
minimizing the matrix norm of the approximation error, we design an approach
that utilizes a deep neural network to accelerate convergence via data-driven
improvement of the search directions. Our method leverages a carefully chosen
convolutional network to approximate the action of the inverse of the linear
operator up to an arbitrary constant. We train the network using unsupervised
learning with a loss function equal to the $L^2$ difference between an input
and the system matrix times the network evaluation, where the unspecified
constant in the approximate inverse is accounted for. We demonstrate the
efficacy of our approach on spatially discretized Poisson equations with
millions of degrees of freedom arising in computational fluid dynamics
applications. Unlike state-of-the-art learning approaches, our algorithm is
capable of reducing the linear system residual to a given tolerance in a small
number of iterations, independent of the problem size. Moreover, our method
generalizes effectively to various systems beyond those encountered during
training.
Related papers
- A forward differential deep learning-based algorithm for solving high-dimensional nonlinear backward stochastic differential equations [0.6040014326756179]
We present a novel forward differential deep learning-based algorithm for solving high-dimensional nonlinear backward differential equations (BSDEs)
Motivated by the fact that differential deep learning can efficiently approximate the labels and their derivatives with respect to inputs, we transform the BSDE problem into a differential deep learning problem.
The main idea of our algorithm is to discretize the integrals using the Euler-Maruyama method and approximate the unknown discrete solution triple using three deep neural networks.
arXiv Detail & Related papers (2024-08-10T19:34:03Z) - Semi-supervised Learning of Partial Differential Operators and Dynamical
Flows [68.77595310155365]
We present a novel method that combines a hyper-network solver with a Fourier Neural Operator architecture.
We test our method on various time evolution PDEs, including nonlinear fluid flows in one, two, and three spatial dimensions.
The results show that the new method improves the learning accuracy at the time point of supervision point, and is able to interpolate and the solutions to any intermediate time.
arXiv Detail & Related papers (2022-07-28T19:59:14Z) - AMS-Net: Adaptive Multiscale Sparse Neural Network with Interpretable
Basis Expansion for Multiphase Flow Problems [8.991619150027267]
We propose an adaptive sparse learning algorithm that can be applied to learn the physical processes and obtain a sparse representation of the solution given a large snapshot space.
The information of the basis functions are incorporated in the loss function, which minimizes the differences between the downscaled reduced order solutions and reference solutions at multiple time steps.
More numerical tests are performed on two-phase multiscale flow problems to show the capability and interpretability of the proposed method on complicated applications.
arXiv Detail & Related papers (2022-07-24T13:12:43Z) - AI-enhanced iterative solvers for accelerating the solution of large
scale parametrized linear systems of equations [0.0]
This paper exploits up-to-date ML tools and delivers customized iterative solvers of linear equation systems.
The results indicate its superiority over conventional iterative solution schemes.
arXiv Detail & Related papers (2022-07-06T09:47:14Z) - A deep branching solver for fully nonlinear partial differential
equations [0.1474723404975345]
We present a multidimensional deep learning implementation of a branching algorithm for the numerical solution of fully nonlinear PDEs.
This approach is designed to tackle functional nonlinearities involving gradient terms of any orders.
arXiv Detail & Related papers (2022-03-07T09:46:46Z) - Deep Learning Approximation of Diffeomorphisms via Linear-Control
Systems [91.3755431537592]
We consider a control system of the form $dot x = sum_i=1lF_i(x)u_i$, with linear dependence in the controls.
We use the corresponding flow to approximate the action of a diffeomorphism on a compact ensemble of points.
arXiv Detail & Related papers (2021-10-24T08:57:46Z) - Learning Linearized Assignment Flows for Image Labeling [70.540936204654]
We introduce a novel algorithm for estimating optimal parameters of linearized assignment flows for image labeling.
We show how to efficiently evaluate this formula using a Krylov subspace and a low-rank approximation.
arXiv Detail & Related papers (2021-08-02T13:38:09Z) - Linear embedding of nonlinear dynamical systems and prospects for
efficient quantum algorithms [74.17312533172291]
We describe a method for mapping any finite nonlinear dynamical system to an infinite linear dynamical system (embedding)
We then explore an approach for approximating the resulting infinite linear system with finite linear systems (truncation)
arXiv Detail & Related papers (2020-12-12T00:01:10Z) - Solving Sparse Linear Inverse Problems in Communication Systems: A Deep
Learning Approach With Adaptive Depth [51.40441097625201]
We propose an end-to-end trainable deep learning architecture for sparse signal recovery problems.
The proposed method learns how many layers to execute to emit an output, and the network depth is dynamically adjusted for each task in the inference phase.
arXiv Detail & Related papers (2020-10-29T06:32:53Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Enhancement of shock-capturing methods via machine learning [0.0]
We develop an improved finite-volume method for simulating PDEs with discontinuous solutions.
We train a neural network to improve the results of a fifth-order WENO method.
We find that our method outperforms WENO in simulations where the numerical solution becomes overly diffused.
arXiv Detail & Related papers (2020-02-06T21:51:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.