Timezone: »

Explicit Explore-Exploit Algorithms in Continuous State Spaces
Mikael Henaff

Tue Dec 10 05:30 PM -- 07:30 PM (PST) @ East Exhibition Hall B + C #187

We present a new model-based algorithm for reinforcement learning (RL) which consists of explicit exploration and exploitation phases, and is applicable in large or infinite state spaces. The algorithm maintains a set of dynamics models consistent with current experience and explores by finding policies which induce high dis- agreement between their state predictions. It then exploits using the refined set of models or experience gathered during exploration. We show that under realizability and optimal planning assumptions, our algorithm provably finds a near-optimal policy with a number of samples that is polynomial in a structural complexity measure which we show to be low in several natural settings. We then give a practical approximation using neural networks and demonstrate its performance and sample efficiency in practice.

Author Information

Mikael Henaff (Microsoft Research)

More from the Same Authors