Optimizing Temperature Distributions for Training Neural Quantum States using Parallel Tempering
- URL: http://arxiv.org/abs/2410.23018v1
- Date: Wed, 30 Oct 2024 13:48:35 GMT
- Title: Optimizing Temperature Distributions for Training Neural Quantum States using Parallel Tempering
- Authors: Conor Smith, Quinn T. Campbell, Tameem Albash,
- Abstract summary: We show that temperature optimization can significantly increase the success rate of variational algorithms.
We demonstrate this using two different neural networks, a restricted Boltzmann machine and a feedforward network.
- Score: 0.0
- License:
- Abstract: Parameterized artificial neural networks (ANNs) can be very expressive ansatzes for variational algorithms, reaching state-of-the-art energies on many quantum many-body Hamiltonians. Nevertheless, the training of the ANN can be slow and stymied by the presence of local minima in the parameter landscape. One approach to mitigate this issue is to use parallel tempering methods, and in this work we focus on the role played by the temperature distribution of the parallel tempering replicas. Using an adaptive method that adjusts the temperatures in order to equate the exchange probability between neighboring replicas, we show that this temperature optimization can significantly increase the success rate of the variational algorithm with negligible computational cost by eliminating bottlenecks in the replicas' random walk. We demonstrate this using two different neural networks, a restricted Boltzmann machine and a feedforward network, which we use to study a toy problem based on a permutation invariant Hamiltonian with a pernicious local minimum and the J1-J2 model on a rectangular lattice.
Related papers
- An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - A Stable and Scalable Method for Solving Initial Value PDEs with Neural
Networks [52.5899851000193]
We develop an ODE based IVP solver which prevents the network from getting ill-conditioned and runs in time linear in the number of parameters.
We show that current methods based on this approach suffer from two key issues.
First, following the ODE produces an uncontrolled growth in the conditioning of the problem, ultimately leading to unacceptably large numerical errors.
arXiv Detail & Related papers (2023-04-28T17:28:18Z) - Isometric tensor network representations of two-dimensional thermal
states [0.0]
We use the class of recently introduced tensor network states to represent thermal states of the transverse field Ising model.
We find that this approach offers a different way with low computational complexity to represent states at finite temperatures.
arXiv Detail & Related papers (2023-02-15T19:00:11Z) - Entropic Neural Optimal Transport via Diffusion Processes [105.34822201378763]
We propose a novel neural algorithm for the fundamental problem of computing the entropic optimal transport (EOT) plan between continuous probability distributions.
Our algorithm is based on the saddle point reformulation of the dynamic version of EOT which is known as the Schr"odinger Bridge problem.
In contrast to the prior methods for large-scale EOT, our algorithm is end-to-end and consists of a single learning step.
arXiv Detail & Related papers (2022-11-02T14:35:13Z) - Quantum-Inspired Tempering for Ground State Approximation using
Artificial Neural Networks [0.0]
We propose a parallel tempering method that facilitates escape from local minima.
We show that augmenting the training with quantum parallel tempering becomes useful to finding good approximations to the ground states of problem instances.
arXiv Detail & Related papers (2022-10-20T16:50:32Z) - Conjugate Gradient Method for Generative Adversarial Networks [0.0]
It is not feasible to calculate the Jensen-Shannon divergence of the density function of the data and the density function of the model of deep neural networks.
Generative adversarial networks (GANs) can be used to formulate this problem as a discriminative problem with two models, a generator and a discriminator.
We propose to apply the conjugate gradient method to solve the local Nash equilibrium problem in GANs.
arXiv Detail & Related papers (2022-03-28T04:44:45Z) - An application of the splitting-up method for the computation of a
neural network representation for the solution for the filtering equations [68.8204255655161]
Filtering equations play a central role in many real-life applications, including numerical weather prediction, finance and engineering.
One of the classical approaches to approximate the solution of the filtering equations is to use a PDE inspired method, called the splitting-up method.
We combine this method with a neural network representation to produce an approximation of the unnormalised conditional distribution of the signal process.
arXiv Detail & Related papers (2022-01-10T11:01:36Z) - Dual Training of Energy-Based Models with Overparametrized Shallow
Neural Networks [41.702175127106784]
Energy-based models (EBMs) are generative models that are usually trained via maximum likelihood estimation.
We propose a dual formulation of an EBMs algorithm in which the particles are sometimes restarted at random samples drawn from the data set, and show that performing these restarts corresponds to a score every step.
These results are illustrated in simple numerical experiments.
arXiv Detail & Related papers (2021-07-11T21:43:18Z) - Study on the simulation control of neural network algorithm in thermally
coupled distillation [7.313669465917949]
The neural network algorithm has the advantages of fast learning and can approach nonlinear functions arbitrarily.
This article summarizes the research progress of artificial neural network and the application of neural network in thermally coupled distillation.
arXiv Detail & Related papers (2021-02-06T04:18:04Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - Spatially Adaptive Inference with Stochastic Feature Sampling and
Interpolation [72.40827239394565]
We propose to compute features only at sparsely sampled locations.
We then densely reconstruct the feature map with an efficient procedure.
The presented network is experimentally shown to save substantial computation while maintaining accuracy over a variety of computer vision tasks.
arXiv Detail & Related papers (2020-03-19T15:36:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.