Reducing classical communication costs in multiplexed quantum repeaters using hardware-aware quasi-local policies
- URL: http://arxiv.org/abs/2401.13168v2
- Date: Thu, 9 May 2024 23:40:59 GMT
- Title: Reducing classical communication costs in multiplexed quantum repeaters using hardware-aware quasi-local policies
- Authors: Stav Haldar, Pratik J. Barge, Xiang Cheng, Kai-Chi Chang, Brian T. Kirby, Sumeet Khatri, Chee Wei Wong, Hwang Lee,
- Abstract summary: We introduce textitquasi-local policies for multiplexed quantum repeater chains.
In quasi-local policies, nodes have increased knowledge of the state of the repeater chain, but not necessarily full, global knowledge.
Our policies also outperform the well-known and widely studied nested purification and doubling swapping policy.
- Score: 5.405186125924916
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Future quantum networks will have nodes equipped with multiple quantum memories, allowing for multiplexing and entanglement distillation strategies in order to increase fidelities and reduce waiting times for end-to-end entanglement distribution. In this work, we introduce \textit{quasi-local} policies for multiplexed quantum repeater chains. In fully-local policies, nodes make decisions based only on knowledge of their own states. In our quasi-local policies, nodes have increased knowledge of the state of the repeater chain, but not necessarily full, global knowledge. Our policies exploit the observation that for most decisions the nodes have to make, they only need to have information about the connected region of the chain they belong to, and not the entire chain. In this way, we not only obtain improved performance over local policies, but we reduce the classical communication (CC) costs inherent to global-knowledge policies. Our policies also outperform the well-known and widely studied nested purification and doubling swapping policy in practically relevant parameter regimes. We also carefully examine the role of entanglement distillation. Via analytical and numerical results, we identify the parameter regimes in which distillation makes sense and is useful. In these regimes, we also address the question: "Should we distill before swapping, or vice versa?" Finally, to provide further practical guidance, we propose an experimental implementation of a multiplexing-based repeater chain, and experimentally demonstrate the key element, a high-dimensional biphoton frequency comb. We then evaluate the anticipated performance of our multiplexing-based policies in such a real-world network through simulation results for two concrete memory platforms, namely rare-earth ions and diamond vacancies.
Related papers
- From $r$ to $Q^*$: Your Language Model is Secretly a Q-Function [50.812404038684505]
We show that we can derive DPO in the token-level MDP as a general inverse Q-learning algorithm, which satisfies the Bellman equation.
We discuss applications of our work, including information elicitation in multi-turn dialogue, reasoning, agentic applications and end-to-end training of multi-model systems.
arXiv Detail & Related papers (2024-04-18T17:37:02Z) - Action-Quantized Offline Reinforcement Learning for Robotic Skill
Learning [68.16998247593209]
offline reinforcement learning (RL) paradigm provides recipe to convert static behavior datasets into policies that can perform better than the policy that collected the data.
In this paper, we propose an adaptive scheme for action quantization.
We show that several state-of-the-art offline RL methods such as IQL, CQL, and BRAC improve in performance on benchmarks when combined with our proposed discretization scheme.
arXiv Detail & Related papers (2023-10-18T06:07:10Z) - Fast and reliable entanglement distribution with quantum repeaters: principles for improving protocols using reinforcement learning [0.6249768559720122]
Future quantum technologies will rely on networks of shared entanglement between spatially separated nodes.
We provide improved protocols/policies for entanglement distribution along a linear chain of nodes.
arXiv Detail & Related papers (2023-03-01T19:05:32Z) - Diversity Through Exclusion (DTE): Niche Identification for
Reinforcement Learning through Value-Decomposition [63.67574523750839]
We propose a generic reinforcement learning (RL) algorithm that performs better than baseline deep Q-learning algorithms in environments with multiple variably-valued niches.
We show that agents trained this way can escape poor-but-attractive local optima to instead converge to harder-to-discover higher value strategies.
arXiv Detail & Related papers (2023-02-02T16:00:19Z) - Symbolic Distillation for Learned TCP Congestion Control [70.27367981153299]
TCP congestion control has achieved tremendous success with deep reinforcement learning (RL) approaches.
Black-box policies lack interpretability and reliability, and often, they need to operate outside the traditional TCP datapath.
This paper proposes a novel two-stage solution to achieve the best of both worlds: first, to train a deep RL agent, then distill its NN policy into white-box, light-weight rules.
arXiv Detail & Related papers (2022-10-24T00:58:16Z) - Optimal entanglement distribution policies in homogeneous repeater
chains with cutoffs [1.9021200954913475]
We study the limits of bipartite entanglement distribution using a chain of quantum repeaters with quantum memories.
We find global-knowledge policies that minimize the expected time to produce end-to-end entanglement.
arXiv Detail & Related papers (2022-07-13T22:25:21Z) - Exact rate analysis for quantum repeaters with imperfect memories and
entanglement swapping as soon as possible [0.0]
We present an exact rate analysis for a secret key that can be shared among two parties employing a linear quantum repeater chain.
We consider additional tools and parameters such as memory cut-offs, multiplexing, initial state and swapping gate fidelities.
arXiv Detail & Related papers (2022-03-19T12:55:56Z) - Rate limits in quantum networks with lossy repeaters [0.6299766708197883]
We quantify how the presence of loss in repeater stations affect the maximum attainable rates for quantum communication.
In the linear chain scenario we show that, by increasing the number of repeater stations, the maximum rate cannot overcome a quantity which solely depends on the loss of a single station.
arXiv Detail & Related papers (2021-10-19T18:00:01Z) - Offline Reinforcement Learning with Implicit Q-Learning [85.62618088890787]
Current offline reinforcement learning methods need to query the value of unseen actions during training to improve the policy.
We propose an offline RL method that never needs to evaluate actions outside of the dataset.
This method enables the learned policy to improve substantially over the best behavior in the data through generalization.
arXiv Detail & Related papers (2021-10-12T17:05:05Z) - Overcoming the repeaterless bound in continuous-variable quantum
communication without quantum memories [0.0]
One of the main problems in quantum communications is how to achieve high rates at long distances.
We introduce a continuous-variable protocol which overcomes the repeaterless bound and scales like the single-repeater bound.
We show that our scheme can be extended to longer repeater chains using quantum memories.
arXiv Detail & Related papers (2021-05-08T04:02:17Z) - Conservative Q-Learning for Offline Reinforcement Learning [106.05582605650932]
We show that CQL substantially outperforms existing offline RL methods, often learning policies that attain 2-5 times higher final return.
We theoretically show that CQL produces a lower bound on the value of the current policy and that it can be incorporated into a policy learning procedure with theoretical improvement guarantees.
arXiv Detail & Related papers (2020-06-08T17:53:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.