Inference of Strategic Behavior based on Incomplete Observation Data
Loading...
Access rights
openAccess
acceptedVersion
URL
Journal Title
Journal ISSN
Volume Title
A4 Artikkeli konferenssijulkaisussa
This publication is imported from Aalto University research portal.
View publication in the Research portal (opens in new window)
View/Open full text file from the Research portal (opens in new window)
Other link related to publication (opens in new window)
View publication in the Research portal (opens in new window)
View/Open full text file from the Research portal (opens in new window)
Other link related to publication (opens in new window)
Authors
Date
Department
Major/Subject
Mcode
Degree programme
Language
en
Pages
4
Series
NIPS17 Workshop: Learning in the Presence of Strategic Behavior
Abstract
Inferring the goals, preferences and restrictions of strategically behaving agents is a common goal in many situations, and an important requirement for enabling computer systems to better model and understand human users. Inverse reinforcement learning (IRL) is one method for performing this kind of inference based on observations of the agent's behavior. However, traditional IRL methods are only applicable when the observations are in the form of state-action paths -- an assumption which does not hold in many real-world modelling settings. This paper demonstrates that inference is possible even with an arbitrary observation noise model.Description
Other note
Citation
Kangasrääsiö, A & Kaski, S 2017, Inference of Strategic Behavior based on Incomplete Observation Data. in NIPS17 Workshop: Learning in the Presence of Strategic Behavior. Carnegie Mellon University, IEEE Conference on Neural Information Processing Systems, Long Beach, California, United States, 04/12/2017.