Simulation-Aided Policy Tuning for Black-Box Robot Learning

dc.contributorAalto-yliopistofi
dc.contributorAalto Universityen
dc.contributor.authorHe, Shiming
dc.contributor.authorvon Rohr, Alexander
dc.contributor.authorBaumann, Dominik
dc.contributor.authorXiang, Ji
dc.contributor.authorTrimpe, Sebastian
dc.contributor.departmentDepartment of Electrical Engineering and Automationen
dc.contributor.groupauthorCyber-physical Systemsen
dc.contributor.organizationHangzhou City University
dc.contributor.organizationRWTH Aachen University
dc.contributor.organizationZhejiang University
dc.date.accessioned2025-02-26T09:34:27Z
dc.date.available2025-02-26T09:34:27Z
dc.date.issued2025
dc.description.abstractHow can robots learn and adapt to new tasks and situations with little data? Systematic exploration and simulation are crucial tools for efficient robot learning. We present a novel black-box policy search algorithm focused on data-efficient policy improvements. The algorithm learns directly on the robot and treats simulation as an additional information source to speed up the learning process. At the core of the algorithm, a probabilistic model learns the dependence between the policy parameters and the robot learning objective not only by performing experiments on the robot, but also by leveraging data from a simulator. This substantially reduces interaction time with the robot. Using the model, we can guarantee improvements with high probability for each policy update, thereby facilitating fast, goal-oriented learning. We evaluate our algorithm on simulated fine-tuning tasks and demonstrate the data-efficiency of the proposed dual-information source optimization algorithm. In a real robot learning experiment, we show fast and successful task learning on a robot manipulator with the aid of an imperfect simulator.en
dc.description.versionPeer revieweden
dc.format.extent16
dc.format.mimetypeapplication/pdf
dc.identifier.citationHe, S, von Rohr, A, Baumann, D, Xiang, J & Trimpe, S 2025, 'Simulation-Aided Policy Tuning for Black-Box Robot Learning', IEEE Transactions on Robotics, vol. 41. https://doi.org/10.1109/TRO.2025.3539192en
dc.identifier.doi10.1109/TRO.2025.3539192
dc.identifier.issn1552-3098
dc.identifier.issn1941-0468
dc.identifier.otherPURE UUID: a1cd0d24-9857-4165-a6d4-3f6790e8ec00
dc.identifier.otherPURE ITEMURL: https://research.aalto.fi/en/publications/a1cd0d24-9857-4165-a6d4-3f6790e8ec00
dc.identifier.otherPURE LINK: http://adsabs.harvard.edu/abs/2024arXiv241114246H
dc.identifier.otherPURE FILEURL: https://research.aalto.fi/files/179900495/Simulation-Aided_Policy_Tuning_for_Black-Box_Robot_Learning.pdf
dc.identifier.urihttps://aaltodoc.aalto.fi/handle/123456789/134325
dc.identifier.urnURN:NBN:fi:aalto-202502262591
dc.language.isoenen
dc.publisherIEEE
dc.relation.ispartofseriesIEEE Transactions on Roboticsen
dc.relation.ispartofseriesVolume 41en
dc.rightsopenAccessen
dc.subject.keywordComputer Science - Machine Learning
dc.subject.keywordComputer Science - Robotics
dc.subject.keywordElectrical Engineering and Systems Science - Systems and Control
dc.titleSimulation-Aided Policy Tuning for Black-Box Robot Learningen
dc.typeA1 Alkuperäisartikkeli tieteellisessä aikakauslehdessäfi
dc.type.versionpublishedVersion

Files