Distributed Training for Deep Reinforcement Learning Decoders on the Toric Code

Examensarbete för masterexamen

Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.12380/300977
Download file(s):
File Description SizeFormat 
Master_Thesis_20_Lindeby_Olsson.pdf2.15 MBAdobe PDFView/Open
Bibliographical item details
FieldValue
Type: Examensarbete för masterexamen
Title: Distributed Training for Deep Reinforcement Learning Decoders on the Toric Code
Authors: Olsson, Adam
Lindeby, Gabriel
Abstract: We distribute the training of a deep reinforcement learning-based decoder on the toric code developed by Fitzek et al. [9]. Reinforcement learning agents asynchronously step through multiple environments in parallel and store transitions in a prioritized experience replay buffer. A separate process samples the replay buffer and performs backpropagation on a policy network. With this setup, we managed to improve wall-clock training times with a factor 12 for toric code sizes of d = 5 and d = 7. For d = 9, we were unable to reach optimal performance but improved the decoder’s success rate using a network with a parameter reduction of factor 20. We argue that these results pave the way for optimal decoders, correcting errors close to what is theoretically possible, based on reinforcement learning for toric code sizes ≤ 9. The complete code for the training and toric code environment can be found in the repository https://github.com/Lindeby/toric-RL-decoder and https://github.com/Lindeby/gym_ToricCode.
Keywords: Deep Reinforcement Learning;Distributed;Toric Code;Quantum Error Correction;Ape-X
Issue Date: 2020
Publisher: Chalmers tekniska högskola / Institutionen för fysik
URI: https://hdl.handle.net/20.500.12380/300977
Collection:Examensarbeten för masterexamen // Master Theses



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.