Learning Task-Agnostic Action Spaces for Movement Optimization

Loading...
Thumbnail Image

Access rights

openAccess
publishedVersion

URL

Journal Title

Journal ISSN

Volume Title

A1 Alkuperäisartikkeli tieteellisessä aikakauslehdessä

Major/Subject

Mcode

Degree programme

Language

en

Pages

Series

IEEE Transactions on Visualization and Computer Graphics, Volume 28, issue 12, pp. 4700-4712

Abstract

We propose a novel method for exploring the dynamics of physically based animated characters, and learning a task-agnostic action space that makes movement optimization easier. Like several previous papers, we parameterize actions as target states, and learn a short-horizon goal-conditioned low-level control policy that drives the agent's state towards the targets. Our novel contribution is that with our exploration data, we are able to learn the low-level policy in a generic manner and without any reference movement data. Trained once for each agent or simulation environment, the policy improves the efficiency of optimizing both trajectories and high-level policies across multiple tasks and optimization algorithms. We also contribute novel visualizations that show how using target states as actions makes optimized trajectories more robust to disturbances; this manifests as wider optima that are easy to find. Due to its simplicity and generality, our proposed approach should provide a building block that can improve a large variety of movement optimization methods and applications.

Description

Publisher Copyright: CCBY

Other note

Citation

Babadi, A, Van de Panne, M, Liu, C & Hamalainen, P 2022, 'Learning Task-Agnostic Action Spaces for Movement Optimization', IEEE Transactions on Visualization and Computer Graphics, vol. 28, no. 12, pp. 4700-4712. https://doi.org/10.1109/TVCG.2021.3100095