Argumentative Reward Learning: Reasoning About Human Preferences

09/28/2022
by   Francis Rhys Ward, et al.
0

We define a novel neuro-symbolic framework, argumentative reward learning, which combines preference-based argumentation with existing approaches to reinforcement learning from human feedback. Our method improves prior work by generalising human preferences, reducing the burden on the user and increasing the robustness of the reward model. We demonstrate this with a number of experiments.

READ FULL TEXT

page 3

page 4

research
10/17/2022

Symbol Guided Hindsight Priors for Reward Learning from Human Preferences

Specifying rewards for reinforcement learned (RL) agents is challenging....
research
10/19/2022

Scaling Laws for Reward Model Overoptimization

In reinforcement learning from human feedback, it is common to optimize ...
research
08/18/2023

Learning Reward Machines through Preference Queries over Sequences

Reward machines have shown great promise at capturing non-Markovian rewa...
research
08/08/2023

Optimizing Algorithms From Pairwise User Preferences

Typical black-box optimization approaches in robotics focus on learning ...
research
06/16/2020

Preference-based Reinforcement Learning with Finite-Time Guarantees

Preference-based Reinforcement Learning (PbRL) replaces reward values in...
research
06/09/2016

Face valuing: Training user interfaces with facial expressions and reinforcement learning

An important application of interactive machine learning is extending or...
research
02/27/2023

Active Reward Learning from Online Preferences

Robot policies need to adapt to human preferences and/or new environment...

Please sign up or login with your details

Forgot password? Click here to reset