Reconstructing Kernel-based Machine Learning Force Fields with
Super-linear Convergence
- URL: http://arxiv.org/abs/2212.12737v2
- Date: Thu, 20 Apr 2023 14:15:37 GMT
- Title: Reconstructing Kernel-based Machine Learning Force Fields with
Super-linear Convergence
- Authors: Stefan Bl\"ucher and Klaus-Robert M\"uller and Stefan Chmiela
- Abstract summary: We consider the broad class of Nystr"om-type methods to construct preconditioners.
All considered methods aim to identify a representative subset of inducing ( Kernel) columns to approximate the dominant kernel spectrum.
- Score: 0.18416014644193063
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Kernel machines have sustained continuous progress in the field of quantum
chemistry. In particular, they have proven to be successful in the low-data
regime of force field reconstruction. This is because many equivariances and
invariances due to physical symmetries can be incorporated into the kernel
function to compensate for much larger datasets. So far, the scalability of
kernel machines has however been hindered by its quadratic memory and cubical
runtime complexity in the number of training points. While it is known, that
iterative Krylov subspace solvers can overcome these burdens, their convergence
crucially relies on effective preconditioners, which are elusive in practice.
Effective preconditioners need to partially pre-solve the learning problem in a
computationally cheap and numerically robust manner. Here, we consider the
broad class of Nystr\"om-type methods to construct preconditioners based on
successively more sophisticated low-rank approximations of the original kernel
matrix, each of which provides a different set of computational trade-offs. All
considered methods aim to identify a representative subset of inducing (kernel)
columns to approximate the dominant kernel spectrum.
Related papers
- Efficient Parameter Optimisation for Quantum Kernel Alignment: A Sub-sampling Approach in Variational Training [1.3551232282678036]
We present a sub-sampling training approach that uses a subset of the kernel matrix at each training step, thereby reducing the overall computational cost of the training.
In this work, we apply the sub-sampling method to synthetic datasets and a real-world breast cancer dataset and demonstrate considerable reductions in the number of circuits required to train the quantum kernel.
arXiv Detail & Related papers (2024-01-05T16:11:34Z) - Gaussian Process Regression under Computational and Epistemic Misspecification [4.5656369638728656]
In large data applications, computational costs can be reduced using low-rank or sparse approximations of the kernel.
This paper investigates the effect of such kernel approximations on the element error.
arXiv Detail & Related papers (2023-12-14T18:53:32Z) - On the Nystrom Approximation for Preconditioning in Kernel Machines [13.085943975157985]
We show that a Nystrom approximation of the spectral preconditioner is cheaper to compute and store, and has demonstrated success in practical applications.
Specifically, we show that a sample of logarithmic size enables the Nystrom-based approximated preconditioner to accelerate gradient descent nearly as well as the exact preconditioner.
arXiv Detail & Related papers (2023-12-06T06:33:25Z) - Higher-order topological kernels via quantum computation [68.8204255655161]
Topological data analysis (TDA) has emerged as a powerful tool for extracting meaningful insights from complex data.
We propose a quantum approach to defining Betti kernels, which is based on constructing Betti curves with increasing order.
arXiv Detail & Related papers (2023-07-14T14:48:52Z) - Linear Self-Attention Approximation via Trainable Feedforward Kernel [77.34726150561087]
In pursuit of faster computation, Efficient Transformers demonstrate an impressive variety of approaches.
We aim to expand the idea of trainable kernel methods to approximate the self-attention mechanism of the Transformer architecture.
arXiv Detail & Related papers (2022-11-08T08:14:11Z) - Noisy Quantum Kernel Machines [58.09028887465797]
An emerging class of quantum learning machines is that based on the paradigm of quantum kernels.
We study how dissipation and decoherence affect their performance.
We show that decoherence and dissipation can be seen as an implicit regularization for the quantum kernel machines.
arXiv Detail & Related papers (2022-04-26T09:52:02Z) - Meta-Learning Hypothesis Spaces for Sequential Decision-making [79.73213540203389]
We propose to meta-learn a kernel from offline data (Meta-KeL)
Under mild conditions, we guarantee that our estimated RKHS yields valid confidence sets.
We also empirically evaluate the effectiveness of our approach on a Bayesian optimization task.
arXiv Detail & Related papers (2022-02-01T17:46:51Z) - Optimization on manifolds: A symplectic approach [127.54402681305629]
We propose a dissipative extension of Dirac's theory of constrained Hamiltonian systems as a general framework for solving optimization problems.
Our class of (accelerated) algorithms are not only simple and efficient but also applicable to a broad range of contexts.
arXiv Detail & Related papers (2021-07-23T13:43:34Z) - Scalable Variational Gaussian Processes via Harmonic Kernel
Decomposition [54.07797071198249]
We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability.
We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections.
Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.
arXiv Detail & Related papers (2021-06-10T18:17:57Z) - The Fast Kernel Transform [21.001203328543006]
We propose the Fast Kernel Transform (FKT), a general algorithm to compute matrix-vector multiplications for datasets in moderate dimensions with quasilinear complexity.
The FKT is easily applied to a broad class of kernels, including Gaussian, Matern, and Rational Quadratic covariance functions and physically motivated Green's functions.
We illustrate the efficacy and versatility of the FKT by providing timing and accuracy benchmarks and by applying it to scale the neighborhood embedding (t-SNE) and Gaussian processes to large real-world data sets.
arXiv Detail & Related papers (2021-06-08T16:15:47Z) - Kernel k-Means, By All Means: Algorithms and Strong Consistency [21.013169939337583]
Kernel $k$ clustering is a powerful tool for unsupervised learning of non-linear data.
In this paper, we generalize results leveraging a general family of means to combat sub-optimal local solutions.
Our algorithm makes use of majorization-minimization (MM) to better solve this non-linear separation problem.
arXiv Detail & Related papers (2020-11-12T16:07:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.