Better, Faster Fermionic Neural Networks
- URL: http://arxiv.org/abs/2011.07125v1
- Date: Fri, 13 Nov 2020 20:55:56 GMT
- Title: Better, Faster Fermionic Neural Networks
- Authors: James S. Spencer, David Pfau, Aleksandar Botev, W. M. C. Foulkes
- Abstract summary: We present several improvements to the FermiNet that allow us to set new records for speed and accuracy on challenging systems.
We find that increasing the size of the network is sufficient to reach chemical accuracy on atoms as large as argon.
This enables us to run the FermiNet on the challenging transition of bicyclobutane to butadiene and compare against the PauliNet on the automerization of cyclobutadiene.
- Score: 68.61120920231944
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Fermionic Neural Network (FermiNet) is a recently-developed neural
network architecture that can be used as a wavefunction Ansatz for
many-electron systems, and has already demonstrated high accuracy on small
systems. Here we present several improvements to the FermiNet that allow us to
set new records for speed and accuracy on challenging systems. We find that
increasing the size of the network is sufficient to reach chemical accuracy on
atoms as large as argon. Through a combination of implementing FermiNet in JAX
and simplifying several parts of the network, we are able to reduce the number
of GPU hours needed to train the FermiNet on large systems by an order of
magnitude. This enables us to run the FermiNet on the challenging transition of
bicyclobutane to butadiene and compare against the PauliNet on the
automerization of cyclobutadiene, and we achieve results near the state of the
art for both.
Related papers
- Open-Source Fermionic Neural Networks with Ionic Charge Initialization [0.0]
We integrate the FermiNet model into a standard and widely used open source library, DeepChem.
We propose novel techniques to overcome the difficulties associated with the assignment of excess or lack of electrons for ions.
arXiv Detail & Related papers (2024-01-16T08:51:58Z) - Speed Limits for Deep Learning [67.69149326107103]
Recent advancement in thermodynamics allows bounding the speed at which one can go from the initial weight distribution to the final distribution of the fully trained network.
We provide analytical expressions for these speed limits for linear and linearizable neural networks.
Remarkably, given some plausible scaling assumptions on the NTK spectra and spectral decomposition of the labels -- learning is optimal in a scaling sense.
arXiv Detail & Related papers (2023-07-27T06:59:46Z) - Neural Wave Functions for Superfluids [3.440236962613469]
We study the unitary Fermi gas, a system with strong, short-range, two-body interactions known to possess a superfluid ground state.
We use the recently developed Fermionic neural network (FermiNet) wave function Ansatz for variational Monte Carlo calculations.
arXiv Detail & Related papers (2023-05-11T17:23:29Z) - MF-NeRF: Memory Efficient NeRF with Mixed-Feature Hash Table [62.164549651134465]
We propose MF-NeRF, a memory-efficient NeRF framework that employs a Mixed-Feature hash table to improve memory efficiency and reduce training time while maintaining reconstruction quality.
Our experiments with state-of-the-art Instant-NGP, TensoRF, and DVGO, indicate our MF-NeRF could achieve the fastest training time on the same GPU hardware with similar or even higher reconstruction quality.
arXiv Detail & Related papers (2023-04-25T05:44:50Z) - Graph Neural Networks for Power Allocation in Wireless Networks with
Full Duplex Nodes [10.150768420975155]
Due to mutual interference between users, power allocation problems in wireless networks are often non-trivial.
Graph Graph neural networks (GNNs) have recently emerged as a promising approach tackling these problems and an approach exploits underlying topology of wireless networks.
arXiv Detail & Related papers (2023-03-27T10:59:09Z) - Towards Neural Variational Monte Carlo That Scales Linearly with System
Size [67.09349921751341]
Quantum many-body problems are central to demystifying some exotic quantum phenomena, e.g., high-temperature superconductors.
The combination of neural networks (NN) for representing quantum states, and the Variational Monte Carlo (VMC) algorithm, has been shown to be a promising method for solving such problems.
We propose a NN architecture called Vector-Quantized Neural Quantum States (VQ-NQS) that utilizes vector-quantization techniques to leverage redundancies in the local-energy calculations of the VMC algorithm.
arXiv Detail & Related papers (2022-12-21T19:00:04Z) - MultiScale MeshGraphNets [65.26373813797409]
We propose two complementary approaches to improve the framework from MeshGraphNets.
First, we demonstrate that it is possible to learn accurate surrogate dynamics of a high-resolution system on a much coarser mesh.
Second, we introduce a hierarchical approach (MultiScale MeshGraphNets) which passes messages on two different resolutions.
arXiv Detail & Related papers (2022-10-02T20:16:20Z) - SparseDNN: Fast Sparse Deep Learning Inference on CPUs [1.6244541005112747]
We present SparseDNN, a sparse deep learning inference engine targeting CPUs.
We show that our sparse code generator can achieve significant speedups over state-of-the-art sparse and dense libraries.
arXiv Detail & Related papers (2021-01-20T03:27:35Z) - When Residual Learning Meets Dense Aggregation: Rethinking the
Aggregation of Deep Neural Networks [57.0502745301132]
We propose Micro-Dense Nets, a novel architecture with global residual learning and local micro-dense aggregations.
Our micro-dense block can be integrated with neural architecture search based models to boost their performance.
arXiv Detail & Related papers (2020-04-19T08:34:52Z) - LogicNets: Co-Designed Neural Networks and Circuits for
Extreme-Throughput Applications [6.9276012494882835]
We present a novel method for designing neural network topologies that directly map to a highly efficient FPGA implementation.
We show that the combination of sparsity and low-bit activation quantization results in high-speed circuits with small logic depth and low LUT cost.
arXiv Detail & Related papers (2020-04-06T22:15:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.