Interactive Groupwise Comparison for Reinforcement Learning from Human Feedback

Loading...
Thumbnail Image

Access rights

openAccess
CC BY-NC
publishedVersion

URL

Journal Title

Journal ISSN

Volume Title

A1 Alkuperäisartikkeli tieteellisessä aikakauslehdessä

Date

Major/Subject

Mcode

Degree programme

Language

en

Pages

Series

Computer Graphics Forum

Abstract

Reinforcement learning from human feedback (RLHF) has emerged as a key enabling technology for aligning AI behaviour with human preferences. The traditional way to collect data in RLHF is via pairwise comparisons: human raters are asked to indicate which one of two samples they prefer. We present an interactive visualisation that better exploits the human visual ability to compare and explore whole groups of samples. The interface is comprised of two linked views: 1) an exploration view showing a contextual overview of all sampled behaviours organised in a hierarchical clustering structure; and 2) a comparison view displaying two selected groups of behaviours for user queries. Users can efficiently explore large sets of behaviours by iterating between these two views. Additionally, we devised an active learning approach suggesting groups for comparison. As shown by our evaluation in six simulated robotics tasks, our approach increases the final rewards by 69.34%. It leads to lower error rates and better policies. We open-source the code that can be easily integrated into the RLHF training loop, supporting research on human–AI alignment.

Description

Publisher Copyright: © 2025 The Author(s). Computer Graphics Forum published by Eurographics - The European Association for Computer Graphics and John Wiley & Sons Ltd. | openaire: EC/HE/101141916/EU//Artificial User

Other note

Citation

Kompatscher, J, Shi, D, Varni, G, Weinkauf, T & Oulasvirta, A 2025, 'Interactive Groupwise Comparison for Reinforcement Learning from Human Feedback', Computer Graphics Forum. https://doi.org/10.1111/cgf.70290