Reader: Model-based language-instructed reinforcement learning

dc.contributorAalto-yliopistofi
dc.contributorAalto Universityen
dc.contributor.authorDainese, Nicolaen_US
dc.contributor.authorMarttinen, Pekkaen_US
dc.contributor.authorIlin, Alexanderen_US
dc.contributor.departmentDepartment of Computer Scienceen
dc.contributor.editorBouamor, Houdaen_US
dc.contributor.editorPino, Juanen_US
dc.contributor.editorBali, Kalikaen_US
dc.contributor.groupauthorProfessorship Marttinen Pekkaen
dc.contributor.groupauthorComputer Science Professorsen
dc.contributor.groupauthorComputer Science - Artificial Intelligence and Machine Learning (AIML) - Research areaen
dc.contributor.groupauthorComputer Science Professors of Practiceen
dc.date.accessioned2024-01-04T09:04:52Z
dc.date.available2024-01-04T09:04:52Z
dc.date.issued2023en_US
dc.description.abstractWe explore how we can build accurate world models, which are partially specified by language, and how we can plan with them in the face of novelty and uncertainty. We propose the first model-based reinforcement learning approach to tackle the environment Read To Fight Monsters (Zhong et al., 2019), a grounded policy learning problem. In RTFM an agent has to reason over a set of rules and a goal, both described in a language manual, and the observations, while taking into account the uncertainty arising from the stochasticity of the environment, in order to generalize successfully its policy to test episodes. We demonstrate the superior performance and sample efficiency of our model-based approach to the existing model-free SOTA agents in eight variants of RTFM. Furthermore, we show how the agent’s plans can be inspected, which represents progress towards more interpretable agents.en
dc.description.versionPeer revieweden
dc.format.mimetypeapplication/pdfen_US
dc.identifier.citationDainese, N, Marttinen, P & Ilin, A 2023, Reader: Model-based language-instructed reinforcement learning. in H Bouamor, J Pino & K Bali (eds), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, pp. 16583–16599, Conference on Empirical Methods in Natural Language Processing, Singapore, Singapore, 06/12/2023. < https://aclanthology.org/2023.emnlp-main.1032 >en
dc.identifier.isbn979-8-89176-060-8
dc.identifier.otherPURE UUID: a344f405-9106-4628-b6e5-632a66040bb5en_US
dc.identifier.otherPURE ITEMURL: https://research.aalto.fi/en/publications/a344f405-9106-4628-b6e5-632a66040bb5en_US
dc.identifier.otherPURE LINK: https://aclanthology.org/2023.emnlp-main.1032en_US
dc.identifier.otherPURE FILEURL: https://research.aalto.fi/files/130972931/SCI_Dainese_etal_EMNLP_2023.pdfen_US
dc.identifier.urihttps://aaltodoc.aalto.fi/handle/123456789/125486
dc.identifier.urnURN:NBN:fi:aalto-202401041175
dc.language.isoenen
dc.relation.ispartofConference on Empirical Methods in Natural Language Processingen
dc.relation.ispartofseriesProceedings of the 2023 Conference on Empirical Methods in Natural Language Processingen
dc.relation.ispartofseriespp. 16583–16599en
dc.rightsopenAccessen
dc.titleReader: Model-based language-instructed reinforcement learningen
dc.typeA4 Artikkeli konferenssijulkaisussafi
dc.type.versionpublishedVersion

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
SCI_Dainese_etal_EMNLP_2023.pdf
Size:
6.21 MB
Format:
Adobe Portable Document Format