Lattice Convolutional Networks for Learning Ground States of Quantum
Many-Body Systems
- URL: http://arxiv.org/abs/2206.07370v1
- Date: Wed, 15 Jun 2022 08:24:37 GMT
- Title: Lattice Convolutional Networks for Learning Ground States of Quantum
Many-Body Systems
- Authors: Cong Fu, Xuan Zhang, Huixin Zhang, Hongyi Ling, Shenglong Xu, Shuiwang
Ji
- Abstract summary: We propose lattice convolutions in which a set of proposed operations are used to convert non-square lattices into grid-like augmented lattices.
Based on the proposed lattice convolutions, we design lattice convolutional networks (LCN) that use self-gating and attention mechanisms.
- Score: 33.82764380485598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning methods have been shown to be effective in representing
ground-state wave functions of quantum many-body systems. Existing methods use
convolutional neural networks (CNNs) for square lattices due to their
image-like structures. For non-square lattices, existing method uses graph
neural network (GNN) in which structure information is not precisely captured,
thereby requiring additional hand-crafted sublattice encoding. In this work, we
propose lattice convolutions in which a set of proposed operations are used to
convert non-square lattices into grid-like augmented lattices on which regular
convolution can be applied. Based on the proposed lattice convolutions, we
design lattice convolutional networks (LCN) that use self-gating and attention
mechanisms. Experimental results show that our method achieves performance on
par or better than existing methods on spin 1/2 $J_1$-$J_2$ Heisenberg model
over the square, honeycomb, triangular, and kagome lattices while without using
hand-crafted encoding.
Related papers
- Scalable Graph Compressed Convolutions [68.85227170390864]
We propose a differentiable method that applies permutations to calibrate input graphs for Euclidean convolution.
Based on the graph calibration, we propose the Compressed Convolution Network (CoCN) for hierarchical graph representation learning.
arXiv Detail & Related papers (2024-07-26T03:14:13Z) - Enhancing lattice kinetic schemes for fluid dynamics with Lattice-Equivariant Neural Networks [79.16635054977068]
We present a new class of equivariant neural networks, dubbed Lattice-Equivariant Neural Networks (LENNs)
Our approach develops within a recently introduced framework aimed at learning neural network-based surrogate models Lattice Boltzmann collision operators.
Our work opens towards practical utilization of machine learning-augmented Lattice Boltzmann CFD in real-world simulations.
arXiv Detail & Related papers (2024-05-22T17:23:15Z) - Identification of vortex in unstructured mesh with graph neural networks [0.0]
We present a Graph Neural Network (GNN) based model with U-Net architecture to identify the vortex in CFD results on unstructured meshes.
A vortex auto-labeling method is proposed to label vortex regions in 2D CFD meshes.
arXiv Detail & Related papers (2023-11-11T12:10:16Z) - Permutation Equivariant Neural Functionals [92.0667671999604]
This work studies the design of neural networks that can process the weights or gradients of other neural networks.
We focus on the permutation symmetries that arise in the weights of deep feedforward networks because hidden layer neurons have no inherent order.
In our experiments, we find that permutation equivariant neural functionals are effective on a diverse set of tasks.
arXiv Detail & Related papers (2023-02-27T18:52:38Z) - A Finite Element-Inspired Hypergraph Neural Network: Application to
Fluid Dynamics Simulations [4.984601297028257]
An emerging trend in deep learning research focuses on the applications of graph neural networks (GNNs) for continuum mechanics simulations.
We present a method to construct a hypergraph by connecting the nodes by elements rather than edges.
We term this method a finite element-inspired hypergraph neural network, in short FEIH($phi$)-GNN.
arXiv Detail & Related papers (2022-12-30T04:10:01Z) - Mesh-based graph convolutional neural network models of processes with
complex initial states [0.0]
We propose a graph convolutional neural network that utilizes the discretized representation of the initial microstructure directly, without segmentation or clustering.
We demonstrate the performance of the proposed network and compare it to traditional pixel-based convolution neural network models and feature-based graph convolutional neural networks on three large datasets.
arXiv Detail & Related papers (2021-06-04T03:40:40Z) - Applying Convolutional Neural Networks to Data on Unstructured Meshes
with Space-Filling Curves [0.4588028371034407]
This paper presents the first classical Convolutional Neural Network (CNN) that can be applied directly to data from unstructured finite element meshes or control volume grids.
CNNs have been hugely influential in the areas of image classification and image compression.
Unstructured meshes are frequently used to solve partial differential equations.
arXiv Detail & Related papers (2020-11-24T01:03:50Z) - Primal-Dual Mesh Convolutional Neural Networks [62.165239866312334]
We propose a primal-dual framework drawn from the graph-neural-network literature to triangle meshes.
Our method takes features for both edges and faces of a 3D mesh as input and dynamically aggregates them.
We provide theoretical insights of our approach using tools from the mesh-simplification literature.
arXiv Detail & Related papers (2020-10-23T14:49:02Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z) - Binarized Graph Neural Network [65.20589262811677]
We develop a binarized graph neural network to learn the binary representations of the nodes with binary network parameters.
Our proposed method can be seamlessly integrated into the existing GNN-based embedding approaches.
Experiments indicate that the proposed binarized graph neural network, namely BGN, is orders of magnitude more efficient in terms of both time and space.
arXiv Detail & Related papers (2020-04-19T09:43:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.