Fast and reliable entanglement distribution with quantum repeaters: principles for improving protocols using reinforcement learning
- URL: http://arxiv.org/abs/2303.00777v4
- Date: Mon, 1 Apr 2024 00:19:12 GMT
- Title: Fast and reliable entanglement distribution with quantum repeaters: principles for improving protocols using reinforcement learning
- Authors: Stav Haldar, Pratik J. Barge, Sumeet Khatri, Hwang Lee,
- Abstract summary: Future quantum technologies will rely on networks of shared entanglement between spatially separated nodes.
We provide improved protocols/policies for entanglement distribution along a linear chain of nodes.
- Score: 0.6249768559720122
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Future quantum technologies such as quantum communication, quantum sensing, and distributed quantum computation, will rely on networks of shared entanglement between spatially separated nodes. In this work, we provide improved protocols/policies for entanglement distribution along a linear chain of nodes, both homogeneous and inhomogeneous, that take practical limitations such as photon losses, non-ideal measurements, and quantum memories with short coherence times into account. For a wide range of parameters, our policies improve upon previously known policies, such as the "swap-as-soon-as-possible" policy, with respect to both the waiting time and the fidelity of the end-to-end entanglement. This improvement is greatest for the most practically relevant cases, namely, for short coherence times, high link losses, and highly asymmetric links. To obtain our results, we model entanglement distribution using a Markov decision process, and then we use the Q-learning reinforcement learning (RL) algorithm to discover new policies. These new policies are characterized by dynamic, state-dependent memory cutoffs and collaboration between the nodes. In particular, we quantify this collaboration between the nodes. Our quantifiers tell us how much "global" knowledge of the network every node has. Finally, our understanding of the performance of large quantum networks is currently limited by the computational inefficiency of simulating them using RL or other optimization methods. Thus, in this work, we present a method for nesting policies in order to obtain policies for large repeater chains. By nesting our RL-based policies for small repeater chains, we obtain policies for large repeater chains that improve upon the swap-as-soon-as-possible policy, and thus we pave the way for a scalable method for obtaining policies for long-distance entanglement distribution.
Related papers
- Diffusion-based Reinforcement Learning via Q-weighted Variational Policy Optimization [55.97310586039358]
Diffusion models have garnered widespread attention in Reinforcement Learning (RL) for their powerful expressiveness and multimodality.
We propose a novel model-free diffusion-based online RL algorithm, Q-weighted Variational Policy Optimization (QVPO)
Specifically, we introduce the Q-weighted variational loss, which can be proved to be a tight lower bound of the policy objective in online RL under certain conditions.
We also develop an efficient behavior policy to enhance sample efficiency by reducing the variance of the diffusion policy during online interactions.
arXiv Detail & Related papers (2024-05-25T10:45:46Z) - Analysis of Asynchronous Protocols for Entanglement Distribution in Quantum Networks [9.971549076128268]
We explore two minimal asynchronous protocols for entanglement in quantum networks.
A parallel scheme generating entanglement independently at the link level, and a sequential scheme extending entanglement iteratively from one party to the other.
Our findings suggest the sequential scheme's superiority due to comparable performance with the parallel scheme, coupled with simpler implementation.
arXiv Detail & Related papers (2024-05-03T18:04:11Z) - Intelligent Hybrid Resource Allocation in MEC-assisted RAN Slicing Network [72.2456220035229]
We aim to maximize the SSR for heterogeneous service demands in the cooperative MEC-assisted RAN slicing system.
We propose a recurrent graph reinforcement learning (RGRL) algorithm to intelligently learn the optimal hybrid RA policy.
arXiv Detail & Related papers (2024-05-02T01:36:13Z) - Reducing classical communication costs in multiplexed quantum repeaters using hardware-aware quasi-local policies [5.405186125924916]
We introduce textitquasi-local policies for multiplexed quantum repeater chains.
In quasi-local policies, nodes have increased knowledge of the state of the repeater chain, but not necessarily full, global knowledge.
Our policies also outperform the well-known and widely studied nested purification and doubling swapping policy.
arXiv Detail & Related papers (2024-01-24T01:13:55Z) - Asynchronous Entanglement Routing for the Quantum Internet [0.42855555838080833]
We propose a new set of asynchronous routing protocols for quantum networks.
The protocols update the entanglement-link asynchronous topologyly, identify optimal entanglement-swapping paths, and preserve unused direct-link entanglements.
Our results indicate that asynchronous protocols achieve a larger upper bound with an appropriate setting and significantly higher entanglement rate than existing synchronous approaches.
arXiv Detail & Related papers (2023-12-21T21:14:21Z) - Multi-User Entanglement Distribution in Quantum Networks Using Multipath
Routing [55.2480439325792]
We propose three protocols that increase the entanglement rate of multi-user applications by leveraging multipath routing.
The protocols are evaluated on quantum networks with NISQ constraints, including limited quantum memories and probabilistic entanglement generation.
arXiv Detail & Related papers (2023-03-06T18:06:00Z) - Offline Policy Optimization in RL with Variance Regularizaton [142.87345258222942]
We propose variance regularization for offline RL algorithms, using stationary distribution corrections.
We show that by using Fenchel duality, we can avoid double sampling issues for computing the gradient of the variance regularizer.
The proposed algorithm for offline variance regularization (OVAR) can be used to augment any existing offline policy optimization algorithms.
arXiv Detail & Related papers (2022-12-29T18:25:01Z) - Faster Last-iterate Convergence of Policy Optimization in Zero-Sum
Markov Games [63.60117916422867]
This paper focuses on the most basic setting of competitive multi-agent RL, namely two-player zero-sum Markov games.
We propose a single-loop policy optimization method with symmetric updates from both agents, where the policy is updated via the entropy-regularized optimistic multiplicative weights update (OMWU) method.
Our convergence results improve upon the best known complexities, and lead to a better understanding of policy optimization in competitive Markov games.
arXiv Detail & Related papers (2022-10-03T16:05:43Z) - Optimal entanglement distribution policies in homogeneous repeater
chains with cutoffs [1.9021200954913475]
We study the limits of bipartite entanglement distribution using a chain of quantum repeaters with quantum memories.
We find global-knowledge policies that minimize the expected time to produce end-to-end entanglement.
arXiv Detail & Related papers (2022-07-13T22:25:21Z) - Policies for elementary links in a quantum network [0.0]
An important problem, especially for near-term quantum networks, is to develop optimal entanglement distribution protocols.
We address this problem by initiating the study of quantum network protocols for entanglement distribution using the theory of decision processes.
We show that the previously-studied memory-cutoff protocol can be phrased as a policy within our decision process framework.
arXiv Detail & Related papers (2020-07-07T04:10:41Z) - Conservative Q-Learning for Offline Reinforcement Learning [106.05582605650932]
We show that CQL substantially outperforms existing offline RL methods, often learning policies that attain 2-5 times higher final return.
We theoretically show that CQL produces a lower bound on the value of the current policy and that it can be incorporated into a policy learning procedure with theoretical improvement guarantees.
arXiv Detail & Related papers (2020-06-08T17:53:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.