Adaptable Resource Generation Protocols For Quantum Networks

Reinforcement Learning For Fast Quantum Resource Generation Policies

Bachelor Thesis (2024)
Author(s)

E.A. Tacettin (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

B.J. Davies – Mentor (TU Delft - QID/Wehner Group)

G.S. Vardoyan – Mentor (TU Delft - Quantum Computer Science)

R. Hai – Graduation committee member (TU Delft - Web Information Systems)

Faculty
Electrical Engineering, Mathematics and Computer Science
More Info
expand_more
Publication Year
2024
Language
English
Graduation Date
25-06-2024
Awarding Institution
Delft University of Technology
Project
['CSE3000 Research Project']
Programme
['Computer Science and Engineering']
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Quantum networks allow quantum processors to communicate over large distances. These networks often require simultaneously existing multiple entangled pairs of quantum bits (entangled links) as a fundamental resource for communication. Link generation is a sequential and probabilistic process, and successfully generated links are stored in a quantum memory. Links in memory are subject to noise that causes their quality to decay and become unusable. This paper uses reinforcement learning (RL) to investigate dynamic tuning of the entanglement generation protocol to minimise the time to generate multiple links. By comparing a fixed number of actions to a continuous action space, we analyse the importance of finer-grained tunings of the protocol. This is tested in simulated near-term and medium-term network abstractions. The results show that protocol tuning significantly reduces the mean time to generate entangled links, with finer tuning providing greater benefits up to a point. Furthermore, a heuristic is derived from the RL policies which matches and exceeds their performance. Future work can explore more advanced reinforcement learning algorithms to find better policies, as well as using different noise models to make more generally applicable policies.

Files

Research_paper_aksel.pdf
(pdf | 0.437 Mb)
License info not available