咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Optimising entanglement distri... 收藏
arXiv

Optimising entanglement distribution policies under classical communication constraints assisted by reinforcement learning

作     者:Li, Jan Coopmans, Tim Emonts, Patrick Goodenough, Kenneth Tura, Jordi van Nieuwenburg, Evert 

作者机构:aQaL Applied Quantum Algorithms Universiteit Leiden Netherlands Instituut-Lorentz Universiteit Leiden Niels Bohrweg 2 Leiden2333 CA Netherlands LIACS Universiteit Leiden Niels Bohrweg 1 Leiden2333 CA Netherlands QuTech Delft University of Technology Lorentzweg 1 Delft2628 CJ Netherlands EEMCS Delft University of Technology Mekelweg 4 Delft2628 CD Netherlands Manning College of Information and Computer Sciences University of Massachusetts Amherst 140 Governors Dr AmherstMA01002 United States 

出 版 物:《arXiv》 (arXiv)

年 卷 期:2024年

核心收录:

主  题:Reinforcement learning 

摘      要:Quantum repeaters play a crucial role in the effective distribution of entanglement over long distances. The nearest-future type of quantum repeater requires two operations: entanglement generation across neighbouring repeaters and entanglement swapping to promote short-range entanglement to long-range. For many hardware setups, these actions are probabilistic, leading to longer distribution times and incurred errors. Significant efforts have been vested in finding the optimal entanglement-distribution policy, i.e. the protocol specifying when a network node needs to generate or swap entanglement, such that the expected time to distribute long-distance entanglement is minimal. This problem is even more intricate in more realistic scenarios, especially when classical communication delays are taken into account. In this work, we formulate our problem as a Markov decision problem and use reinforcement learning (RL) to optimise over centralised strategies, where one designated node instructs other nodes which actions to perform. Contrary to most RL models, ours can be readily interpreted. Additionally, we introduce and evaluate a fixed local policy, the ‘predictive swap-asap’ policy, where nodes only coordinate with nearest neighbours. Compared to the straightforward generalization of the common swap-asap policy to the scenario with classical communication effects, the ‘wait-for-broadcast swap-asap’ policy, both of the aforementioned entanglement-delivery policies are faster at high success probabilities. Our work showcases the merit of considering policies acting with incomplete information in the realistic case when classical communication effects are significant. © 2024, CC BY.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分