Improving sample-efficiency of model-free reinforcement learning algorithms on image inputs with representation learning

dc.contributor.authorGuberina, Marko
dc.contributor.authorDesta, Betelhem Dejene
dc.contributor.departmentChalmers tekniska högskola / Institutionen för data och informationstekniksv
dc.contributor.examinerStrannegård, Claes
dc.contributor.supervisorGrover, Divya
dc.date.accessioned2022-10-14T13:06:29Z
dc.date.available2022-10-14T13:06:29Z
dc.date.issued2022sv
dc.date.submitted2020
dc.description.abstractReinforcement learning struggles to solve control tasks on directly on images. Performance on identical tasks with access to the underlying states is much better. One avenue to bridge the gap between the two is to leverage unsupervised learning as a means of learning state representations from images, thereby resulting in a better conditioned reinforcement learning problem. Through investigation of related work, characteristics of successful integration of unsupervised learning and reinforcement learning are identified. We hypothesize that joint training of state representations and policies result in higher sample-efficiency if adequate regularization is provided. We further hypothesize that representations which correlate more strongly with the underlying Markov decision process result in additional sample-efficiency. These hypotheses are tested through a simple deterministic generative representation learning model (autoencoder) trained with image reconstruction loss and additional forward and inverse auxiliary losses. While our algorithm does not reach state-of-the-art performance, its modular implementation integrated in the reinforcement learning library Tianshou enables easy use to reinforcement learning practitioners, and thus also accelerates further research. We also identify which aspects of our solution are most important and use them to formulate promising research directions. In our tests we limited ourselves to Atari environments and primarily used Rainbow as the underlying reinforcement learning algorithm.sv
dc.identifier.coursecodeDATX05sv
dc.identifier.urihttps://hdl.handle.net/20.500.12380/305717
dc.language.isoengsv
dc.setspec.uppsokTechnology
dc.subjectsample-efficient reinforcement learningsv
dc.subjectstate representation learningsv
dc.subjectunsupervised learningsv
dc.subjectautoencodersv
dc.titleImproving sample-efficiency of model-free reinforcement learning algorithms on image inputs with representation learningsv
dc.type.degreeExamensarbete för masterexamensv
dc.type.uppsokH
Ladda ner
Original bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
CSE 22-133 Guberina Desta.pdf
Storlek:
1.47 MB
Format:
Adobe Portable Document Format
Beskrivning:
License bundle
Visar 1 - 1 av 1
Hämtar...
Bild (thumbnail)
Namn:
license.txt
Storlek:
1.51 KB
Format:
Item-specific license agreed upon to submission
Beskrivning: