A Deep Dive into the Connections Between the Renormalization Group and
Deep Learning in the Ising Model
- URL: http://arxiv.org/abs/2308.11075v1
- Date: Mon, 21 Aug 2023 22:50:54 GMT
- Title: A Deep Dive into the Connections Between the Renormalization Group and
Deep Learning in the Ising Model
- Authors: Kelsie Taylor
- Abstract summary: Renormalization group (RG) is an essential technique in statistical physics and quantum field theory.
We develop extensive renormalization techniques for the 1D and 2D Ising model to provide a baseline for comparison.
For the 2D Ising model, we successfully generated Ising model samples using the Wolff algorithm, and performed the group flow using a quasi-deterministic method.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The renormalization group (RG) is an essential technique in statistical
physics and quantum field theory, which considers scale-invariant properties of
physical theories and how these theories' parameters change with scaling. Deep
learning is a powerful computational technique that uses multi-layered neural
networks to solve a myriad of complicated problems. Previous research suggests
the possibility that unsupervised deep learning may be a form of RG flow, by
being a layer-by-layer coarse graining of the original data. We examined this
connection on a more rigorous basis for the simple example of Kadanoff block
renormalization of the 2D nearest-neighbor Ising model, with our deep learning
accomplished via Restricted Boltzmann Machines (RBMs). We developed extensive
renormalization techniques for the 1D and 2D Ising model to provide a baseline
for comparison. For the 1D Ising model, we successfully used Adam optimization
on a correlation length loss function to learn the group flow, yielding results
consistent with the analytical model for infinite N. For the 2D Ising model, we
successfully generated Ising model samples using the Wolff algorithm, and
performed the group flow using a quasi-deterministic method, validating these
results by calculating the critical exponent \nu. We then examined RBM learning
of the Ising model layer by layer, finding a blocking structure in the learning
that is qualitatively similar to RG. Lastly, we directly compared the weights
of each layer from the learning to Ising spin renormalization, but found
quantitative inconsistencies for the simple case of nearest-neighbor Ising
models.
Related papers
- Diffusion-Model-Assisted Supervised Learning of Generative Models for
Density Estimation [10.793646707711442]
We present a framework for training generative models for density estimation.
We use the score-based diffusion model to generate labeled data.
Once the labeled data are generated, we can train a simple fully connected neural network to learn the generative model in the supervised manner.
arXiv Detail & Related papers (2023-10-22T23:56:19Z) - Layer-wise Linear Mode Connectivity [52.6945036534469]
Averaging neural network parameters is an intuitive method for the knowledge of two independent models.
It is most prominently used in federated learning.
We analyse the performance of the models that result from averaging single, or groups.
arXiv Detail & Related papers (2023-07-13T09:39:10Z) - Git Re-Basin: Merging Models modulo Permutation Symmetries [3.5450828190071655]
We show how simple algorithms can be used to fit large networks in practice.
We demonstrate the first (to our knowledge) demonstration of zero mode connectivity between independently trained models.
We also discuss shortcomings in the linear mode connectivity hypothesis.
arXiv Detail & Related papers (2022-09-11T10:44:27Z) - Robust lEarned Shrinkage-Thresholding (REST): Robust unrolling for
sparse recover [87.28082715343896]
We consider deep neural networks for solving inverse problems that are robust to forward model mis-specifications.
We design a new robust deep neural network architecture by applying algorithm unfolding techniques to a robust version of the underlying recovery problem.
The proposed REST network is shown to outperform state-of-the-art model-based and data-driven algorithms in both compressive sensing and radar imaging problems.
arXiv Detail & Related papers (2021-10-20T06:15:45Z) - Investigating the Relationship Between Dropout Regularization and Model
Complexity in Neural Networks [0.0]
Dropout Regularization serves to reduce variance in Deep Learning models.
We explore the relationship between the dropout rate and model complexity by training 2,000 neural networks.
We build neural networks that predict the optimal dropout rate given the number of hidden units in each dense layer.
arXiv Detail & Related papers (2021-08-14T23:49:33Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Towards an Understanding of Benign Overfitting in Neural Networks [104.2956323934544]
Modern machine learning models often employ a huge number of parameters and are typically optimized to have zero training loss.
We examine how these benign overfitting phenomena occur in a two-layer neural network setting.
We show that it is possible for the two-layer ReLU network interpolator to achieve a near minimax-optimal learning rate.
arXiv Detail & Related papers (2021-06-06T19:08:53Z) - Provable Model-based Nonlinear Bandit and Reinforcement Learning: Shelve
Optimism, Embrace Virtual Curvature [61.22680308681648]
We show that global convergence is statistically intractable even for one-layer neural net bandit with a deterministic reward.
For both nonlinear bandit and RL, the paper presents a model-based algorithm, Virtual Ascent with Online Model Learner (ViOL)
arXiv Detail & Related papers (2021-02-08T12:41:56Z) - Continual Learning of Generative Models with Limited Data: From
Wasserstein-1 Barycenter to Adaptive Coalescence [22.82926450287203]
Learning generative models is challenging for a network edge node with limited data and computing power.
This study aims to develop a framework which systematically optimize continual learning of generative models.
arXiv Detail & Related papers (2021-01-22T17:15:39Z) - Sparsely constrained neural networks for model discovery of PDEs [0.0]
We present a modular framework that determines the sparsity pattern of a deep-learning based surrogate using any sparse regression technique.
We show how a different network architecture and sparsity estimator improve model discovery accuracy and convergence on several benchmark examples.
arXiv Detail & Related papers (2020-11-09T11:02:40Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.