Sparse Cholesky Factorization for Solving Nonlinear PDEs via Gaussian
Processes
- URL: http://arxiv.org/abs/2304.01294v3
- Date: Sat, 9 Mar 2024 02:48:15 GMT
- Title: Sparse Cholesky Factorization for Solving Nonlinear PDEs via Gaussian
Processes
- Authors: Yifan Chen, Houman Owhadi, Florian Sch\"afer
- Abstract summary: We present a sparse Cholesky factorization algorithm for dense kernel matrices.
We numerically illustrate our algorithm's near-linear space/time complexity for a broad class of nonlinear PDEs.
- Score: 3.750429354590631
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, there has been widespread adoption of machine learning-based
approaches to automate the solving of partial differential equations (PDEs).
Among these approaches, Gaussian processes (GPs) and kernel methods have
garnered considerable interest due to their flexibility, robust theoretical
guarantees, and close ties to traditional methods. They can transform the
solving of general nonlinear PDEs into solving quadratic optimization problems
with nonlinear, PDE-induced constraints. However, the complexity bottleneck
lies in computing with dense kernel matrices obtained from pointwise
evaluations of the covariance kernel, and its \textit{partial derivatives}, a
result of the PDE constraint and for which fast algorithms are scarce.
The primary goal of this paper is to provide a near-linear complexity
algorithm for working with such kernel matrices. We present a sparse Cholesky
factorization algorithm for these matrices based on the near-sparsity of the
Cholesky factor under a novel ordering of pointwise and derivative
measurements. The near-sparsity is rigorously justified by directly connecting
the factor to GP regression and exponential decay of basis functions in
numerical homogenization. We then employ the Vecchia approximation of GPs,
which is optimal in the Kullback-Leibler divergence, to compute the approximate
factor. This enables us to compute $\epsilon$-approximate inverse Cholesky
factors of the kernel matrices with complexity $O(N\log^d(N/\epsilon))$ in
space and $O(N\log^{2d}(N/\epsilon))$ in time. We integrate sparse Cholesky
factorizations into optimization algorithms to obtain fast solvers of the
nonlinear PDE. We numerically illustrate our algorithm's near-linear space/time
complexity for a broad class of nonlinear PDEs such as the nonlinear elliptic,
Burgers, and Monge-Amp\`ere equations.
Related papers
- Toward Efficient Kernel-Based Solvers for Nonlinear PDEs [19.975293084297014]
This paper introduces a novel kernel learning framework toward efficiently solving nonlinear partial differential equations (PDEs)
In contrast to the state-of-the-art kernel solver that embeds differential operators within kernels, our approach eliminates these operators from the kernel.
We model the solution using a standard kernel form and differentiate the interpolant to compute the derivatives.
arXiv Detail & Related papers (2024-10-15T01:00:43Z) - A Hybrid Kernel-Free Boundary Integral Method with Operator Learning for Solving Parametric Partial Differential Equations In Complex Domains [0.0]
Kernel-Free Boundary Integral (KFBI) method presents an iterative solution to boundary integral equations arising from elliptic partial differential equations (PDEs)
We propose a hybrid KFBI method, integrating the foundational principles of the KFBI method with the capabilities of deep learning.
arXiv Detail & Related papers (2024-04-23T17:25:35Z) - A Deep-Genetic Algorithm (Deep-GA) Approach for High-Dimensional
Nonlinear Parabolic Partial Differential Equations [0.0]
We propose a new method, called a deep-genetic algorithm (deep-GA) to accelerate the performance of the so-called deep-BSDE method.
Recognizing the sensitivity of the solver to the initial guess selection, we embed a genetic algorithm (GA) into the solver to optimize the selection.
We show that our method provides comparable accuracy with significantly improved computational efficiency.
arXiv Detail & Related papers (2023-11-20T06:35:23Z) - Stable Nonconvex-Nonconcave Training via Linear Interpolation [51.668052890249726]
This paper presents a theoretical analysis of linearahead as a principled method for stabilizing (large-scale) neural network training.
We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear can help by leveraging the theory of nonexpansive operators.
arXiv Detail & Related papers (2023-10-20T12:45:12Z) - Constrained Optimization via Exact Augmented Lagrangian and Randomized
Iterative Sketching [55.28394191394675]
We develop an adaptive inexact Newton method for equality-constrained nonlinear, nonIBS optimization problems.
We demonstrate the superior performance of our method on benchmark nonlinear problems, constrained logistic regression with data from LVM, and a PDE-constrained problem.
arXiv Detail & Related papers (2023-05-28T06:33:37Z) - Fast Computation of Optimal Transport via Entropy-Regularized Extragradient Methods [75.34939761152587]
Efficient computation of the optimal transport distance between two distributions serves as an algorithm that empowers various applications.
This paper develops a scalable first-order optimization-based method that computes optimal transport to within $varepsilon$ additive accuracy.
arXiv Detail & Related papers (2023-01-30T15:46:39Z) - An Operator-Splitting Method for the Gaussian Curvature Regularization
Model with Applications in Surface Smoothing and Imaging [6.860238280163609]
We propose an operator-splitting method for a general Gaussian curvature model.
The proposed method is not sensitive to the choice of parameters, its efficiency and performances being demonstrated.
arXiv Detail & Related papers (2021-08-04T08:59:41Z) - Scalable Variational Gaussian Processes via Harmonic Kernel
Decomposition [54.07797071198249]
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability.
We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections.
Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
arXiv Detail & Related papers (2021-06-10T18:17:57Z) - Solving and Learning Nonlinear PDEs with Gaussian Processes [11.09729362243947]
We introduce a simple, rigorous, and unified framework for solving nonlinear partial differential equations.
The proposed approach provides a natural generalization of collocation kernel methods to nonlinear PDEs and IPs.
For IPs, while the traditional approach has been to iterate between the identifications of parameters in the PDE and the numerical approximation of its solution, our algorithm tackles both simultaneously.
arXiv Detail & Related papers (2021-03-24T03:16:08Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Optimal Randomized First-Order Methods for Least-Squares Problems [56.05635751529922]
This class of algorithms encompasses several randomized methods among the fastest solvers for least-squares problems.
We focus on two classical embeddings, namely, Gaussian projections and subsampled Hadamard transforms.
Our resulting algorithm yields the best complexity known for solving least-squares problems with no condition number dependence.
arXiv Detail & Related papers (2020-02-21T17:45:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.