Distributed Tensor Network Library for Quantum Computing Emulation
- URL: http://arxiv.org/abs/2505.06119v1
- Date: Fri, 09 May 2025 15:17:42 GMT
- Title: Distributed Tensor Network Library for Quantum Computing Emulation
- Authors: Jakub Adamski, Oliver Thomson Brown,
- Abstract summary: HPC tensor network packages tackle this issue with a procedure called circuit slicing.<n>We present a novel alternative approach, where individual tensors are both broadcast and scattered.<n>We showcase its capabilities on ARCHER2, by emulating two well-known algorithms.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Tensor networks offer an adaptable and efficient approach to emulation of quantum computers. Their usage relies on partitioning circuits into small tensors, which are contracted together to form the final result. While this approach intends to minimise the problem size, exceeding the locally available memory is sometimes unavoidable due to the exponential nature of quantum systems. Most HPC tensor network packages tackle this issue with a procedure called circuit slicing, which distributes the entire network onto multiple ranks, recombining it back when necessary. In this study, we present a novel alternative approach, where individual tensors are both broadcast and scattered to harness multiple levels of parallelism. The technique is abstracted behind a fixed distribution pattern, and actualised in a new portable tensor network library, QTNH, built on top of MPI and ScaLAPACK. We showcase its capabilities on ARCHER2, by emulating two well-known algorithms - the Quantum Fourier Transform and Random Circuit Sampling. This is accomplished by leveraging the implemented operations to realise various contraction strategies, including a unique distributed MPS tensor factorisation approach. We thus demonstrate that our library can be used to advance the accuracy of quantum emulation, while offering a simple and flexible interface to tensor distribution.
Related papers
- Hybrid quantum tensor networks for aeroelastic applications [1.2932412290302258]
We investigate the application of hybrid quantum tensor networks to aeroelastic problems, harnessing the power of Quantum Machine Learning (QML)<n>Our results showcase the ability of hybrid quantum tensor networks to achieve high accuracy in binary classification.<n>We observe promising performance in regressing discrete variables.
arXiv Detail & Related papers (2025-08-07T09:00:18Z) - Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - Quantum tensor network algorithms for evaluation of spectral functions on quantum computers [0.0]
We investigate quantum algorithms derived from tensor networks to simulate the static and dynamic properties of quantum many-body systems.<n>We demonstrate algorithms to prepare ground and excited states on a quantum computer and apply them to molecular nanomagnets (MNMs) as a paradigmatic example.
arXiv Detail & Related papers (2023-09-26T18:01:42Z) - MBQuant: A Novel Multi-Branch Topology Method for Arbitrary Bit-width Network Quantization [51.85834744835766]
We propose MBQuant, a novel method for arbitrary bit-width quantization.
We show that MBQuant achieves significant performance gains compared to existing arbitrary bit-width quantization methods.
arXiv Detail & Related papers (2023-05-14T10:17:09Z) - Vertical Layering of Quantized Neural Networks for Heterogeneous
Inference [57.42762335081385]
We study a new vertical-layered representation of neural network weights for encapsulating all quantized models into a single one.
We can theoretically achieve any precision network for on-demand service while only needing to train and maintain one model.
arXiv Detail & Related papers (2022-12-10T15:57:38Z) - On Optimizing the Communication of Model Parallelism [74.15423270435949]
We study a novel and important communication pattern in large-scale model-parallel deep learning (DL)
In cross-mesh resharding, a sharded tensor needs to be sent from a source device mesh to a destination device mesh.
We propose two contributions to address cross-mesh resharding: an efficient broadcast-based communication system, and an "overlapping-friendly" pipeline schedule.
arXiv Detail & Related papers (2022-11-10T03:56:48Z) - RosneT: A Block Tensor Algebra Library for Out-of-Core Quantum Computing
Simulation [0.18472148461613155]
We present RosneT, a library for distributed, out-of-core block tensor algebra.
We use the PyCOMPSs programming model to transform tensor operations into a collection of tasks handled by the COMPSs runtime.
We report results validating our approach showing good scalability in simulations of Quantum circuits of up to 53 qubits.
arXiv Detail & Related papers (2022-01-17T20:35:40Z) - Parallel Simulation of Quantum Networks with Distributed Quantum State
Management [56.24769206561207]
We identify requirements for parallel simulation of quantum networks and develop the first parallel discrete event quantum network simulator.
Our contributions include the design and development of a quantum state manager that maintains shared quantum information distributed across multiple processes.
We release the parallel SeQUeNCe simulator as an open-source tool alongside the existing sequential version.
arXiv Detail & Related papers (2021-11-06T16:51:17Z) - A Multisite Decomposition of the Tensor Network Path Integrals [0.0]
We extend the tensor network path integral (TNPI) framework to efficiently simulate quantum systems with local dissipative environments.
The MS-TNPI method is useful for studying a variety of extended quantum systems coupled with solvents.
arXiv Detail & Related papers (2021-09-20T17:55:53Z) - Cluster-Promoting Quantization with Bit-Drop for Minimizing Network
Quantization Loss [61.26793005355441]
Cluster-Promoting Quantization (CPQ) finds the optimal quantization grids for neural networks.
DropBits is a new bit-drop technique that revises the standard dropout regularization to randomly drop bits instead of neurons.
We experimentally validate our method on various benchmark datasets and network architectures.
arXiv Detail & Related papers (2021-09-05T15:15:07Z) - Efficient 2D Tensor Network Simulation of Quantum Systems [6.074275058563179]
2D tensor networks such as Projected Entangled States (PEPS) are well-suited for key classes of physical systems and quantum circuits.
We propose new algorithms and software abstractions for PEPS-based methods, accelerating the bottleneck operation of contraction and scalableization of a subnetwork.
arXiv Detail & Related papers (2020-06-26T22:36:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.