Simulating first-order phase transition with hierarchical autoregressive
networks
- URL: http://arxiv.org/abs/2212.04955v2
- Date: Thu, 25 May 2023 08:11:56 GMT
- Title: Simulating first-order phase transition with hierarchical autoregressive
networks
- Authors: Piotr Bia{\l}as, Paulina Czarnota, Piotr Korcyl, Tomasz Stebel
- Abstract summary: We apply the Hierarchical Autoregressive Neural (HAN) network sampling algorithm to the two-dimensional $Q$-state Potts model.
We quantify the performance of the approach in the vicinity of the first-order phase transition and compare it with that of the Wolff cluster algorithm.
- Score: 0.04588028371034406
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We apply the Hierarchical Autoregressive Neural (HAN) network sampling
algorithm to the two-dimensional $Q$-state Potts model and perform simulations
around the phase transition at $Q=12$. We quantify the performance of the
approach in the vicinity of the first-order phase transition and compare it
with that of the Wolff cluster algorithm. We find a significant improvement as
far as the statistical uncertainty is concerned at a similar numerical effort.
In order to efficiently train large neural networks we introduce the technique
of pre-training. It allows to train some neural networks using smaller system
sizes and then employing them as starting configurations for larger system
sizes. This is possible due to the recursive construction of our hierarchical
approach. Our results serve as a demonstration of the performance of the
hierarchical approach for systems exhibiting bimodal distributions.
Additionally, we provide estimates of the free energy and entropy in the
vicinity of the phase transition with statistical uncertainties of the order of
$10^{-7}$ for the former and $10^{-3}$ for the latter based on a statistics of
$10^6$ configurations.
Related papers
- Efficient Neural Network Approaches for Conditional Optimal Transport with Applications in Bayesian Inference [1.740133468405535]
We present two neural network approaches that approximate the solutions of static and conditional optimal transport (COT) problems.
We demonstrate both algorithms, comparing them with competing state-the-art approaches, using benchmark datasets and simulation-based inverse problems.
arXiv Detail & Related papers (2023-10-25T20:20:09Z) - Optimization Guarantees of Unfolded ISTA and ADMM Networks With Smooth
Soft-Thresholding [57.71603937699949]
We study optimization guarantees, i.e., achieving near-zero training loss with the increase in the number of learning epochs.
We show that the threshold on the number of training samples increases with the increase in the network width.
arXiv Detail & Related papers (2023-09-12T13:03:47Z) - Improved Convergence Guarantees for Shallow Neural Networks [91.3755431537592]
We prove convergence of depth 2 neural networks, trained via gradient descent, to a global minimum.
Our model has the following features: regression with quadratic loss function, fully connected feedforward architecture, RelU activations, Gaussian data instances, adversarial labels.
They strongly suggest that, at least in our model, the convergence phenomenon extends well beyond the NTK regime''
arXiv Detail & Related papers (2022-12-05T14:47:52Z) - Hierarchical autoregressive neural networks for statistical systems [0.05156484100374058]
We propose a hierarchical association of physical degrees of freedom, for instance spins, to neurons which replaces it with the scaling with the linear extent $L$ of the system.
We demonstrate our approach on the two-dimensional Ising model by simulating lattices of various sizes up to $128 times 128$ spins, with time benchmarks reaching lattices of size $512 times 512$.
arXiv Detail & Related papers (2022-03-21T13:55:53Z) - Neural Capacitance: A New Perspective of Neural Network Selection via
Edge Dynamics [85.31710759801705]
Current practice requires expensive computational costs in model training for performance prediction.
We propose a novel framework for neural network selection by analyzing the governing dynamics over synaptic connections (edges) during training.
Our framework is built on the fact that back-propagation during neural network training is equivalent to the dynamical evolution of synaptic connections.
arXiv Detail & Related papers (2022-01-11T20:53:15Z) - Scale Mixtures of Neural Network Gaussian Processes [22.07524388784668]
We introduce a scale mixture of $mathrmNNGP$ for which we introduce a prior on the scale of the last-layer parameters.
We show that with certain scale priors, we obtain heavytailed processes, and we recover Student's $t$ processes in the case of inverse gamma distributions.
We further analyze the neural networks with our prior setting and trained with gradient descents and obtain similar results as for $mathrmNNGP$.
arXiv Detail & Related papers (2021-07-03T11:02:18Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - MSE-Optimal Neural Network Initialization via Layer Fusion [68.72356718879428]
Deep neural networks achieve state-of-the-art performance for a range of classification and inference tasks.
The use of gradient combined nonvolutionity renders learning susceptible to novel problems.
We propose fusing neighboring layers of deeper networks that are trained with random variables.
arXiv Detail & Related papers (2020-01-28T18:25:15Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.