Skip to yearly menu bar Skip to main content


Poster

Reward Mapping for Transfer in Long-Lived Agents

Xiaoxiao Guo · Satinder Singh · Richard L Lewis

Harrah's Special Events Center, 2nd Floor

Abstract:

We consider how to transfer knowledge from previous tasks to a current task in long-lived and bounded agents that must solve a sequence of MDPs over a finite lifetime. A novel aspect of our transfer approach is that we reuse reward functions. While this may seem counterintuitive, we build on the insight of recent work on the optimal rewards problem that guiding an agent's behavior with reward functions other than the task-specifying reward function can help overcome computational bounds of the agent. Specifically, we use good guidance reward functions learned on previous tasks in the sequence to incrementally train a reward mapping function that maps task-specifying reward functions into good initial guidance reward functions for subsequent tasks. We demonstrate that our approach can substantially improve the agent's performance relative to other approaches, including an approach that transfers policies.

Live content is unavailable. Log in and register to view live content