Rayleigh-Gauss-Newton optimization with enhanced sampling for
variational Monte Carlo
- URL: http://arxiv.org/abs/2106.10558v1
- Date: Sat, 19 Jun 2021 19:05:52 GMT
- Title: Rayleigh-Gauss-Newton optimization with enhanced sampling for
variational Monte Carlo
- Authors: Robert J. Webber, Michael Lindsey
- Abstract summary: We analyze optimization and sampling methods used in Variational Monte Carlo.
We introduce alterations to improve their performance.
In particular, we demonstrate that RGN can be made robust to energy spikes.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational Monte Carlo (VMC) is an approach for computing ground-state
wavefunctions that has recently become more powerful due to the introduction of
neural network-based wavefunction parametrizations. However, efficiently
training neural wavefunctions to converge to an energy minimum remains a
difficult problem. In this work, we analyze optimization and sampling methods
used in VMC and introduce alterations to improve their performance. First,
based on theoretical convergence analysis in a noiseless setting, we motivate a
new optimizer that we call the Rayleigh-Gauss-Newton method, which can improve
upon gradient descent and natural gradient descent to achieve superlinear
convergence. Second, in order to realize this favorable comparison in the
presence of stochastic noise, we analyze the effect of sampling error on VMC
parameter updates and experimentally demonstrate that it can be reduced by the
parallel tempering method. In particular, we demonstrate that RGN can be made
robust to energy spikes that occur when new regions of configuration space
become available to the sampler over the course of optimization. Finally,
putting theory into practice, we apply our enhanced optimization and sampling
methods to the transverse-field Ising and XXZ models on large lattices,
yielding ground-state energy estimates with remarkably high accuracy after just
200-500 parameter updates.
Related papers
- Gradient Normalization with(out) Clipping Ensures Convergence of Nonconvex SGD under Heavy-Tailed Noise with Improved Results [60.92029979853314]
This paper investigates Gradient Normalization without (NSGDC) its gradient reduction variant (NSGDC-VR)
We present significant improvements in the theoretical results for both algorithms.
arXiv Detail & Related papers (2024-10-21T22:40:42Z) - Second-order optimisation strategies for neural network quantum states [1.814143871199829]
We revisit the Kronecker Factored Approximate Curvature, an optimiser that has been used extensively in a variety of simulations.
We reformulate the Variational Monte Carlo approach in a game theory framework, to propose a novel optimiser based on decision geometry.
We find that, on a practical test case for continuous systems, this new optimiser consistently outperforms any of the KFAC improvements in terms of stability, accuracy and speed of convergence.
arXiv Detail & Related papers (2024-01-31T02:34:14Z) - A Kaczmarz-inspired approach to accelerate the optimization of neural network wavefunctions [0.7438129207086058]
We propose the Subsampled Projected Gradient-Increment Natural Descent (SPRING) to reduce this bottleneck.
SPRING combines ideas from the recently introduced minimum-step reconfiguration (MinSR) and the classical randomized Kaczmarz method for solving linear least-squares problems.
We demonstrate that SPRING outperforms both MinSR and the popular Kronecker-Factored Approximate Curvature method (KFAC) across a number of small atoms and molecules.
arXiv Detail & Related papers (2024-01-18T18:23:10Z) - Improving sample efficiency of high dimensional Bayesian optimization
with MCMC [7.241485121318798]
We propose a new method based on Markov Chain Monte Carlo to efficiently sample from an approximated posterior.
We show experimentally that both the Metropolis-Hastings and the Langevin Dynamics version of our algorithm outperform state-of-the-art methods in high-dimensional sequential optimization and reinforcement learning benchmarks.
arXiv Detail & Related papers (2024-01-05T05:56:42Z) - Stochastic Gradient Descent for Gaussian Processes Done Right [86.83678041846971]
We show that when emphdone right -- by which we mean using specific insights from optimisation and kernel communities -- gradient descent is highly effective.
We introduce a emphstochastic dual descent algorithm, explain its design in an intuitive manner and illustrate the design choices.
Our method places Gaussian process regression on par with state-of-the-art graph neural networks for molecular binding affinity prediction.
arXiv Detail & Related papers (2023-10-31T16:15:13Z) - Model-Based Reparameterization Policy Gradient Methods: Theory and
Practical Algorithms [88.74308282658133]
Reization (RP) Policy Gradient Methods (PGMs) have been widely adopted for continuous control tasks in robotics and computer graphics.
Recent studies have revealed that, when applied to long-term reinforcement learning problems, model-based RP PGMs may experience chaotic and non-smooth optimization landscapes.
We propose a spectral normalization method to mitigate the exploding variance issue caused by long model unrolls.
arXiv Detail & Related papers (2023-10-30T18:43:21Z) - Neural Operator Variational Inference based on Regularized Stein
Discrepancy for Deep Gaussian Processes [23.87733307119697]
We introduce Neural Operator Variational Inference (NOVI) for Deep Gaussian Processes.
NOVI uses a neural generator to obtain a sampler and minimizes the Regularized Stein Discrepancy in L2 space between the generated distribution and true posterior.
We demonstrate that the bias introduced by our method can be controlled by multiplying the divergence with a constant, which leads to robust error control and ensures the stability and precision of the algorithm.
arXiv Detail & Related papers (2023-09-22T06:56:35Z) - Neural network quantum state with proximal optimization: a ground-state
searching scheme based on variational Monte Carlo [4.772126473623257]
We introduce a novel objective function with proximal optimization (PO) that enables multiple updates via reusing the mismatched samples.
We investigate the performance of our VMC-PO algorithm for ground-state searching with a 1-dimensional transverse-field Ising model and 2-dimensional Heisenberg antiferromagnet on a square lattice.
arXiv Detail & Related papers (2022-10-29T04:55:39Z) - NerfingMVS: Guided Optimization of Neural Radiance Fields for Indoor
Multi-view Stereo [97.07453889070574]
We present a new multi-view depth estimation method that utilizes both conventional SfM reconstruction and learning-based priors.
We show that our proposed framework significantly outperforms state-of-the-art methods on indoor scenes.
arXiv Detail & Related papers (2021-09-02T17:54:31Z) - Self-Tuning Stochastic Optimization with Curvature-Aware Gradient
Filtering [53.523517926927894]
We explore the use of exact per-sample Hessian-vector products and gradients to construct self-tuning quadratics.
We prove that our model-based procedure converges in noisy gradient setting.
This is an interesting step for constructing self-tuning quadratics.
arXiv Detail & Related papers (2020-11-09T22:07:30Z) - Extrapolation for Large-batch Training in Deep Learning [72.61259487233214]
We show that a host of variations can be covered in a unified framework that we propose.
We prove the convergence of this novel scheme and rigorously evaluate its empirical performance on ResNet, LSTM, and Transformer.
arXiv Detail & Related papers (2020-06-10T08:22:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.