Mixed Precision Fermi-Operator Expansion on Tensor Cores From a Machine
Learning Perspective
- URL: http://arxiv.org/abs/2101.06385v1
- Date: Sat, 16 Jan 2021 06:55:20 GMT
- Title: Mixed Precision Fermi-Operator Expansion on Tensor Cores From a Machine
Learning Perspective
- Authors: Joshua Finkelstein, Justin Smith, Susan M. Mniszewski, Kipton Barros,
Christian F. A. Negre, Emanuel H. Rubensson, Anders M. N. Niklasson
- Abstract summary: A performance of over 100 teraFLOPs is achieved for half-precision floating point operations on Nvidia's A100 tensor core units.
A differentiable deep neural network structure is formulated to solve the quantum mechanical electronic structure problem.
- Score: 0.20011494166747584
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a second-order recursive Fermi-operator expansion scheme using
mixed precision floating point operations to perform electronic structure
calculations using tensor core units. A performance of over 100 teraFLOPs is
achieved for half-precision floating point operations on Nvidia's A100 tensor
core units. The second-order recursive Fermi-operator scheme is formulated in
terms of a generalized, differentiable deep neural network structure, which
solves the quantum mechanical electronic structure problem. We demonstrate how
this network can be accelerated by optimizing the weight and bias values to
substantially reduce the number of layers required for convergence. We also
show how this machine learning approach can be used to optimize the
coefficients of the recursive Fermi-operator expansion to accurately represent
fractional occupation numbers of the electronic states at finite temperatures.
Related papers
- Multi-Grid Tensorized Fourier Neural Operator for High-Resolution PDEs [93.82811501035569]
We introduce a new data efficient and highly parallelizable operator learning approach with reduced memory requirement and better generalization.
MG-TFNO scales to large resolutions by leveraging local and global structures of full-scale, real-world phenomena.
We demonstrate superior performance on the turbulent Navier-Stokes equations where we achieve less than half the error with over 150x compression.
arXiv Detail & Related papers (2023-09-29T20:18:52Z) - Tensor Factorized Recursive Hamiltonian Downfolding To Optimize The Scaling Complexity Of The Electronic Correlations Problem on Classical and Quantum Computers [0.15833270109954137]
We present a new variant of post-Hartree-Fock Hamiltonian downfolding-based quantum chemistry methods with optimized scaling for high-cost simulations.
We demonstrate super-quadratic speedups of expensive quantum chemistry algorithms on both classical and quantum computers.
arXiv Detail & Related papers (2023-03-13T12:15:54Z) - D4FT: A Deep Learning Approach to Kohn-Sham Density Functional Theory [79.50644650795012]
We propose a deep learning approach to solve Kohn-Sham Density Functional Theory (KS-DFT)
We prove that such an approach has the same expressivity as the SCF method, yet reduces the computational complexity.
In addition, we show that our approach enables us to explore more complex neural-based wave functions.
arXiv Detail & Related papers (2023-03-01T10:38:10Z) - Efficient Dataset Distillation Using Random Feature Approximation [109.07737733329019]
We propose a novel algorithm that uses a random feature approximation (RFA) of the Neural Network Gaussian Process (NNGP) kernel.
Our algorithm provides at least a 100-fold speedup over KIP and can run on a single GPU.
Our new method, termed an RFA Distillation (RFAD), performs competitively with KIP and other dataset condensation algorithms in accuracy over a range of large-scale datasets.
arXiv Detail & Related papers (2022-10-21T15:56:13Z) - Quantum perturbation theory using Tensor cores and a deep neural network [0.0]
Time-independent quantum response calculations are performed using floating cores.
We demonstrate a peak performance of almost 200 Tflops using the cores of two Nvidia A100 GPUs.
arXiv Detail & Related papers (2022-03-17T21:24:10Z) - Simulating thermal density operators with cluster expansions and tensor
networks [0.0]
We benchmark this cluster tensor network operator (cluster TNO) for one-dimensional systems.
We use this formalism for representing the thermal density operator of a two-dimensional quantum spin system at a certain temperature as a single cluster TNO.
We find through a scaling analysis that the cluster-TNO approximation gives rise to a continuous phase transition in the correct universality class.
arXiv Detail & Related papers (2021-12-02T18:56:44Z) - Factorized Fourier Neural Operators [77.47313102926017]
The Factorized Fourier Neural Operator (F-FNO) is a learning-based method for simulating partial differential equations.
We show that our model maintains an error rate of 2% while still running an order of magnitude faster than a numerical solver.
arXiv Detail & Related papers (2021-11-27T03:34:13Z) - Quantum-based Molecular Dynamics Simulations Using Tensor Cores [2.3551989288556774]
We show how tensor cores can be applied with high efficiency to the Born-Oppenheimer molecular dynamics problem.
The interatomic forces are calculated on-the-fly from an electronic structure that is obtained from a generalized deep neural network.
A canonical ensemble simulation scheme is also presented, where the additional numerical noise in the calculated forces is absorbed into a Langevin-like dynamics.
arXiv Detail & Related papers (2021-07-06T17:11:45Z) - Entanglement of formation of mixed many-body quantum states via Tree
Tensor Operators [0.0]
We use a positive loopless representation for density matrices to encode information on bipartite entanglement.
We observe a finite-size scaling law for the entanglement of formation in 1D critical lattice models at finite temperature for up to 128 spins, extending to mixed states the scaling law for the entanglement entropy.
arXiv Detail & Related papers (2020-11-02T19:00:04Z) - Temporal Attention-Augmented Graph Convolutional Network for Efficient
Skeleton-Based Human Action Recognition [97.14064057840089]
Graphal networks (GCNs) have been very successful in modeling non-Euclidean data structures.
Most GCN-based action recognition methods use deep feed-forward networks with high computational complexity to process all skeletons in an action.
We propose a temporal attention module (TAM) for increasing the efficiency in skeleton-based action recognition.
arXiv Detail & Related papers (2020-10-23T08:01:55Z) - Efficient construction of tensor-network representations of many-body
Gaussian states [59.94347858883343]
We present a procedure to construct tensor-network representations of many-body Gaussian states efficiently and with a controllable error.
These states include the ground and thermal states of bosonic and fermionic quadratic Hamiltonians, which are essential in the study of quantum many-body systems.
arXiv Detail & Related papers (2020-08-12T11:30:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.