Show simple item record

dc.rights.licenseCC-BY-NC-ND
dc.contributor.advisorvan Ommen, Thijs
dc.contributor.authorKrantas, S.
dc.date.accessioned2021-03-18T19:00:27Z
dc.date.available2021-03-18T19:00:27Z
dc.date.issued2021
dc.identifier.urihttps://studenttheses.uu.nl/handle/20.500.12932/39130
dc.description.abstractThe field of Reinforcement Learning (RL) has been developing extremely fast in the last few years. Its combination with Neural Networks has transformed this area of Machine Learning into a powerful tool that can solve a variety of problems with efficiency and many times without the help of humans. In this thesis, we created an end-to-end learning algorithm (SUMRL) by combining RL and Meta Learning techniques for fast and efficient adaptation to new and unknown tasks. In order to free our agent as much as possible from human supervision, we implemented a technique of Unsupervised Learning (with the help of the Information Theory field) and motivated our agent to create unsupervised skills. These skills were later fed to Meta Learning algorithms as tasks for our agent to meta-train on. We evaluated our algorithm by comparing the meta-testing outcome when the agent was meta-trained with our pretrained skills/tasks and with random tasks. Our experiments showed that although our agent has a good overall performance in both cases, pretrained tasks were not able to clearly return better results than the random tasks at the process of meta-testing.
dc.description.sponsorshipUtrecht University
dc.format.extent12298316
dc.format.mimetypeapplication/pdf
dc.language.isoen
dc.titleSemi Unsupervised Meta Reinforcement Learning
dc.type.contentMaster Thesis
dc.rights.accessrightsOpen Access
dc.subject.courseuuArtificial Intelligence


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record