Towards Neural Sparse Linear Solvers
- URL: http://arxiv.org/abs/2203.06944v1
- Date: Mon, 14 Mar 2022 09:17:02 GMT
- Title: Towards Neural Sparse Linear Solvers
- Authors: Luca Grementieri and Paolo Galeone
- Abstract summary: We propose neural sparse linear solvers to learn approximate solvers for sparse symmetric linear systems.
Our method relies on representing a sparse symmetric linear system as an undirected weighted graph.
We test sparse linear solvers on static linear analysis problems from structural engineering.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Large sparse symmetric linear systems appear in several branches of science
and engineering thanks to the widespread use of the finite element method
(FEM). The fastest sparse linear solvers available implement hybrid iterative
methods. These methods are based on heuristic algorithms to permute rows and
columns or find a preconditioner matrix. In addition, they are inherently
sequential, making them unable to leverage the GPU processing power entirely.
We propose neural sparse linear solvers, a deep learning framework to learn
approximate solvers for sparse symmetric linear systems. Our method relies on
representing a sparse symmetric linear system as an undirected weighted graph.
Such graph representation is inherently permutation-equivariant and
scale-invariant, and it can become the input to a graph neural network trained
to regress the solution. We test neural sparse linear solvers on static linear
analysis problems from structural engineering. Our method is less accurate than
classic algorithms, but it is hardware-independent, fast on GPUs, and
applicable to generic sparse symmetric systems without any additional
hypothesis. Although many limitations remain, this study shows a general
approach to tackle problems involving sparse symmetric matrices using graph
neural networks.
Related papers
- Neural incomplete factorization: learning preconditioners for the conjugate gradient method [2.899792823251184]
We develop a data-driven approach to accelerate the generation of effective preconditioners.
We replace the typically hand-engineered preconditioners by the output of graph neural networks.
Our method generates an incomplete factorization of the matrix and is, therefore, referred to as neural incomplete factorization (NeuralIF)
arXiv Detail & Related papers (2023-05-25T11:45:46Z) - Implicit Parameter-free Online Learning with Truncated Linear Models [51.71216912089413]
parameter-free algorithms are online learning algorithms that do not require setting learning rates.
We propose new parameter-free algorithms that can take advantage of truncated linear models through a new update that has an "implicit" flavor.
Based on a novel decomposition of the regret, the new update is efficient, requires only one gradient at each step, never overshoots the minimum of the truncated model, and retains the favorable parameter-free properties.
arXiv Detail & Related papers (2022-03-19T13:39:49Z) - High-Dimensional Sparse Bayesian Learning without Covariance Matrices [66.60078365202867]
We introduce a new inference scheme that avoids explicit construction of the covariance matrix.
Our approach couples a little-known diagonal estimation result from numerical linear algebra with the conjugate gradient algorithm.
On several simulations, our method scales better than existing approaches in computation time and memory.
arXiv Detail & Related papers (2022-02-25T16:35:26Z) - Quantum Algorithm for Solving a Quadratic Nonlinear System of Equations [0.22940141855172036]
The complexity of our algorithm is $O(rm polylog(n/epsilon))$, which provides an exponential improvement over the optimal classical algorithm in dimension $n$.
Our algorithm exponentially accelerates the solution of QNSE and has wide applications in all kinds of nonlinear problems.
arXiv Detail & Related papers (2021-12-03T00:27:16Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - Optimization-Based Algebraic Multigrid Coarsening Using Reinforcement
Learning [0.0]
A system of linear equations defines a graph on the set of unknowns.
Each level of a multigrid solver requires the selection of an appropriate coarse graph along with restriction and operators that map to and from the coarse representation.
It has been demonstrated that it is possible to directly learn the AMG and restriction operators, given a coarse graph selection.
We propose a sparse method using a reinforcement learning agent based on graph neural networks (GNNs)
arXiv Detail & Related papers (2021-06-03T13:57:32Z) - Multipole Graph Neural Operator for Parametric Partial Differential
Equations [57.90284928158383]
One of the main challenges in using deep learning-based methods for simulating physical systems is formulating physics-based data.
We propose a novel multi-level graph neural network framework that captures interaction at all ranges with only linear complexity.
Experiments confirm our multi-graph network learns discretization-invariant solution operators to PDEs and can be evaluated in linear time.
arXiv Detail & Related papers (2020-06-16T21:56:22Z) - Learning nonlinear dynamical systems from a single trajectory [102.60042167341956]
We introduce algorithms for learning nonlinear dynamical systems of the form $x_t+1=sigma(Thetastarx_t)+varepsilon_t$.
We give an algorithm that recovers the weight matrix $Thetastar$ from a single trajectory with optimal sample complexity and linear running time.
arXiv Detail & Related papers (2020-04-30T10:42:48Z) - Eigendecomposition-Free Training of Deep Networks for Linear
Least-Square Problems [107.3868459697569]
We introduce an eigendecomposition-free approach to training a deep network.
We show that our approach is much more robust than explicit differentiation of the eigendecomposition.
Our method has better convergence properties and yields state-of-the-art results.
arXiv Detail & Related papers (2020-04-15T04:29:34Z) - Learning Algebraic Multigrid Using Graph Neural Networks [34.32501734380907]
We train a single graph neural network to learn a mapping from an entire class of such matrices to prolongation operators.
Experiments on a broad class of problems demonstrate improved convergence rates compared to classical AMG.
arXiv Detail & Related papers (2020-03-12T12:36:48Z) - Multiple Metric Learning for Structured Data [0.0]
We address the problem of merging graph and feature-space information while learning a metric from structured data.
We propose a new graph-based technique for optimizing under metric constraints.
arXiv Detail & Related papers (2020-02-13T19:11:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.