Optimising entanglement distribution policies under classical communication constraints assisted by reinforcement learning
- URL: http://arxiv.org/abs/2412.06938v1
- Date: Mon, 09 Dec 2024 19:26:49 GMT
- Title: Optimising entanglement distribution policies under classical communication constraints assisted by reinforcement learning
- Authors: Jan Li, Tim Coopmans, Patrick Emonts, Kenneth Goodenough, Jordi Tura, Evert van Nieuwenburg,
- Abstract summary: Quantum repeaters play a crucial role in the effective distribution of entanglement over long distances.
We introduce and evaluate a fixed local policy, the predictive swap-asap' policy, where nodes only coordinate with nearest neighbours.
Our work showcases the merit of considering policies acting with incomplete information in the realistic case when classical communication effects are significant.
- Score: 0.0
- License:
- Abstract: Quantum repeaters play a crucial role in the effective distribution of entanglement over long distances. The nearest-future type of quantum repeater requires two operations: entanglement generation across neighbouring repeaters and entanglement swapping to promote short-range entanglement to long-range. For many hardware setups, these actions are probabilistic, leading to longer distribution times and incurred errors. Significant efforts have been vested in finding the optimal entanglement-distribution policy, i.e. the protocol specifying when a network node needs to generate or swap entanglement, such that the expected time to distribute long-distance entanglement is minimal. This problem is even more intricate in more realistic scenarios, especially when classical communication delays are taken into account. In this work, we formulate our problem as a Markov decision problem and use reinforcement learning (RL) to optimise over centralised strategies, where one designated node instructs other nodes which actions to perform. Contrary to most RL models, ours can be readily interpreted. Additionally, we introduce and evaluate a fixed local policy, the `predictive swap-asap' policy, where nodes only coordinate with nearest neighbours. Compared to the straightforward generalization of the common swap-asap policy to the scenario with classical communication effects, the `wait-for-broadcast swap-asap' policy, both of the aforementioned entanglement-delivery policies are faster at high success probabilities. Our work showcases the merit of considering policies acting with incomplete information in the realistic case when classical communication effects are significant.
Related papers
- Optimized Generation of Entanglement by Real-Time Ordering of Swapping Operations [4.152766500389854]
Long-distance quantum communication in quantum networks faces significant challenges due to the constraints imposed by the no-cloning theorem.
Most existing quantum communication protocols rely on the a priori distribution of entanglement pairs (EPs)
We present a greedy algorithm that iteratively determines the best route and/or entanglement-swapping operation to perform at each stage based on the current network.
arXiv Detail & Related papers (2024-05-13T06:43:11Z) - Collaborative Ground-Space Communications via Evolutionary Multi-objective Deep Reinforcement Learning [113.48727062141764]
We propose a distributed collaborative beamforming (DCB)-based uplink communication paradigm for enabling ground-space direct communications.
DCB treats the terminals that are unable to establish efficient direct connections with the low Earth orbit (LEO) satellites as distributed antennas.
We propose an evolutionary multi-objective deep reinforcement learning algorithm to obtain the desirable policies.
arXiv Detail & Related papers (2024-04-11T03:13:02Z) - Off-Policy Evaluation for Large Action Spaces via Policy Convolution [60.6953713877886]
Policy Convolution family of estimators uses latent structure within actions to strategically convolve the logging and target policies.
Experiments on synthetic and benchmark datasets demonstrate remarkable mean squared error (MSE) improvements when using PC.
arXiv Detail & Related papers (2023-10-24T01:00:01Z) - Fast and reliable entanglement distribution with quantum repeaters: principles for improving protocols using reinforcement learning [0.6249768559720122]
Future quantum technologies will rely on networks of shared entanglement between spatially separated nodes.
We provide improved protocols/policies for entanglement distribution along a linear chain of nodes.
arXiv Detail & Related papers (2023-03-01T19:05:32Z) - Faster Last-iterate Convergence of Policy Optimization in Zero-Sum
Markov Games [63.60117916422867]
This paper focuses on the most basic setting of competitive multi-agent RL, namely two-player zero-sum Markov games.
We propose a single-loop policy optimization method with symmetric updates from both agents, where the policy is updated via the entropy-regularized optimistic multiplicative weights update (OMWU) method.
Our convergence results improve upon the best known complexities, and lead to a better understanding of policy optimization in competitive Markov games.
arXiv Detail & Related papers (2022-10-03T16:05:43Z) - Optimal entanglement distribution policies in homogeneous repeater
chains with cutoffs [1.9021200954913475]
We study the limits of bipartite entanglement distribution using a chain of quantum repeaters with quantum memories.
We find global-knowledge policies that minimize the expected time to produce end-to-end entanglement.
arXiv Detail & Related papers (2022-07-13T22:25:21Z) - Adaptive advantage in entanglement-assisted communications [0.0]
Entanglement-assisted classical communication protocols usually consist of two successive rounds.
We show that adaptive protocols improve the success probability in Random Access Codes.
We briefly discuss extension of these ideas to scenarios involving quantum communication.
arXiv Detail & Related papers (2022-03-10T13:54:02Z) - Distributed Adaptive Learning Under Communication Constraints [54.22472738551687]
This work examines adaptive distributed learning strategies designed to operate under communication constraints.
We consider a network of agents that must solve an online optimization problem from continual observation of streaming data.
arXiv Detail & Related papers (2021-12-03T19:23:48Z) - Quantum communication complexity beyond Bell nonlocality [87.70068711362255]
Efficient distributed computing offers a scalable strategy for solving resource-demanding tasks.
Quantum resources are well-suited to this task, offering clear strategies that can outperform classical counterparts.
We prove that a new class of communication complexity tasks can be associated to Bell-like inequalities.
arXiv Detail & Related papers (2021-06-11T18:00:09Z) - Improved analytical bounds on delivery times of long-distance
entanglement [0.0]
We provide improved analytical bounds on the average and on the quantiles of the completion time of entanglement distribution protocols.
A canonical example of such a protocol is a nested quantum repeater scheme which consists of heralded entanglement generation and entanglement swaps.
arXiv Detail & Related papers (2021-03-21T18:14:56Z) - Decentralized MCTS via Learned Teammate Models [89.24858306636816]
We present a trainable online decentralized planning algorithm based on decentralized Monte Carlo Tree Search.
We show that deep learning and convolutional neural networks can be employed to produce accurate policy approximators.
arXiv Detail & Related papers (2020-03-19T13:10:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.