Reweighted Time-Evolving Block Decimation for Improved Quantum Dynamics Simulations
- URL: http://arxiv.org/abs/2412.08730v1
- Date: Wed, 11 Dec 2024 19:01:00 GMT
- Title: Reweighted Time-Evolving Block Decimation for Improved Quantum Dynamics Simulations
- Authors: Sayak Guha Roy, Kevin Slagle,
- Abstract summary: We introduce a simple yet significant improvement to the time-evolving block decimation (TEBD) algorithm for simulating the time dynamics of 1D mixed quantum states.
We propose a reweighted TEBD algorithm that deprioritizes high-weight expectation values by a factor of $gamma-n$ during the truncation.
This simple modification makes rTEBD significantly more accurate than the TEBD time-dependent simulation of an MPDO, and competive with and sometimes better than TEBD using MPS.
- Score: 0.0
- License:
- Abstract: We introduce a simple yet significant improvement to the time-evolving block decimation (TEBD) tensor network algorithm for simulating the time dynamics of strongly correlated one-dimensional (1D) mixed quantum states. The efficiency of 1D tensor network methods stems from using a product of matrices to express either: the coefficients of a wavefunction, yielding a matrix product state (MPS); or the expectation values of a density matrix, yielding a matrix product density operator (MPDO). To avoid exponential computational costs, TEBD truncates the matrix dimension while simulating the time evolution. However, when truncating a MPDO, TEBD does not favor the likely more important low-weight expectation values, such as $\langle c_i^\dagger c_j \rangle$, over the exponentially many high-weight expectation values, such as $\langle c_{i_1}^\dagger c^\dagger_{i_2} \cdots c_{i_n} \rangle$ of weight $n$, despite the critical importance of the low-weight expectation values. Motivated by this shortcoming, we propose a reweighted TEBD (rTEBD) algorithm that deprioritizes high-weight expectation values by a factor of $\gamma^{-n}$ during the truncation. This simple modification (which only requires reweighting certain matrices by a factor of $\gamma$ in the MPDO) makes rTEBD significantly more accurate than the TEBD time-dependent simulation of an MPDO, and competive with and sometimes better than TEBD using MPS. Furthermore, by prioritizing low-weight expectation values, rTEBD preserves conserved quantities to high precision.
Related papers
- Large Language Model Evaluation via Matrix Nuclear-Norm [11.878496378814045]
We introduce the Matrix Nuclear-Norm, which serves as a metric to quantify the data compression proficiency of large language models (LLMs)
By employing the ( L_1,2text-norm ) to further approximate the nuclear norm, we can effectively assess the model's information compression capabilities.
The Matrix Nuclear-Norm achieves speeds 8 to 24 times faster than Matrix Entropy for the CEREBRAS-GPT model as sizes increase from 111M to 6.7B.
arXiv Detail & Related papers (2024-10-14T16:15:57Z) - HiRE: High Recall Approximate Top-$k$ Estimation for Efficient LLM
Inference [68.59839755875252]
HiRE comprises of two novel components: (i) a compression scheme to cheaply predict top-$k$ rows/columns with high recall, followed by full computation restricted to the predicted subset, and (ii) DA-TOP-$k$: an efficient multi-device approximate top-$k$ operator.
We demonstrate that on a one billion parameter model, HiRE applied to both the softmax as well as feedforward layers, achieves almost matching pretraining and downstream accuracy, and speeds up inference latency by $1.47times$ on a single TPUv5e device.
arXiv Detail & Related papers (2024-02-14T18:04:36Z) - DBA: Efficient Transformer with Dynamic Bilinear Low-Rank Attention [53.02648818164273]
We present an efficient yet effective attention mechanism, namely the Dynamic Bilinear Low-Rank Attention (DBA)
DBA compresses the sequence length by input-sensitive dynamic projection matrices and achieves linear time and space complexity.
Experiments over tasks with diverse sequence length conditions show that DBA achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-11-24T03:06:36Z) - RSC: Accelerating Graph Neural Networks Training via Randomized Sparse
Computations [56.59168541623729]
Training graph neural networks (GNNs) is time consuming because sparse graph-based operations are hard to be accelerated by hardware.
We explore trading off the computational precision to reduce the time complexity via sampling-based approximation.
We propose Randomized Sparse Computation, which for the first time demonstrate the potential of training GNNs with approximated operations.
arXiv Detail & Related papers (2022-10-19T17:25:33Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Softmax-free Linear Transformers [90.83157268265654]
Vision transformers (ViTs) have pushed the state-of-the-art for visual perception tasks.
Existing methods are either theoretically flawed or empirically ineffective for visual recognition.
We propose a family of Softmax-Free Transformers (SOFT)
arXiv Detail & Related papers (2022-07-05T03:08:27Z) - LUT-GEMM: Quantized Matrix Multiplication based on LUTs for Efficient Inference in Large-Scale Generative Language Models [9.727062803700264]
We introduce LUT-GEMM, an efficient kernel for quantized matrix multiplication.
LUT-GEMM eliminates the resource-intensive dequantization process and reduces computational costs.
We show experimentally that when applied to the OPT-175B model with 3-bit quantization, LUT-GEMM substantially accelerates token generation latency.
arXiv Detail & Related papers (2022-06-20T03:48:17Z) - Unfolding Projection-free SDP Relaxation of Binary Graph Classifier via
GDPA Linearization [59.87663954467815]
Algorithm unfolding creates an interpretable and parsimonious neural network architecture by implementing each iteration of a model-based algorithm as a neural layer.
In this paper, leveraging a recent linear algebraic theorem called Gershgorin disc perfect alignment (GDPA), we unroll a projection-free algorithm for semi-definite programming relaxation (SDR) of a binary graph.
Experimental results show that our unrolled network outperformed pure model-based graph classifiers, and achieved comparable performance to pure data-driven networks but using far fewer parameters.
arXiv Detail & Related papers (2021-09-10T07:01:15Z) - The Power of Log-Sum-Exp: Sequential Density Ratio Matrix Estimation for
Speed-Accuracy Optimization [0.0]
We propose a model for multiclass classification of time series to make a prediction as early and as accurate as possible.
Our overall architecture for early classification, MSPRT-TANDEM, statistically significantly outperforms baseline models on four datasets.
arXiv Detail & Related papers (2021-05-28T07:21:58Z) - Machine Learning Regression for Operator Dynamics [0.0]
We present a solution for efficiently extending the computation of expectation values to long time intervals.
We utilize a multi-layer perceptron (MLP) model as a tool for regression on expectation values calculated within the regime of short time intervals.
arXiv Detail & Related papers (2021-02-23T18:58:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.