Poster

Replacing Rewards with Examples: Example-Based Policy Search via Recursive Classification

Ben Eysenbach · Sergey Levine · Russ Salakhutdinov

Keywords: [ Reinforcement Learning and Planning ] [ Machine Learning ]

[ Abstract ]
[ OpenReview
Wed 8 Dec 4:30 p.m. PST — 6 p.m. PST
 
Oral presentation: Oral Session 5: Reinforcement Learning and Planning
Fri 10 Dec 4 p.m. PST — 5 p.m. PST

Abstract:

Reinforcement learning (RL) algorithms assume that users specify tasks by manually writing down a reward function. However, this process can be laborious and demands considerable technical expertise. Can we devise RL algorithms that instead enable users to specify tasks simply by providing examples of successful outcomes? In this paper, we derive a control algorithm that maximizes the future probability of these successful outcome examples. Prior work has approached similar problems with a two-stage process, first learning a reward function and then optimizing this reward function using another reinforcement learning algorithm. In contrast, our method directly learns a value function from transitions and successful outcomes, without learning this intermediate reward function. Our method therefore requires fewer hyperparameters to tune and lines of code to debug. We show that our method satisfies a new data-driven Bellman equation, where examples take the place of the typical reward function term. Experiments show that our approach outperforms prior methods that learn explicit reward functions.

Chat is not available.