Timezone: »

Mask-based Latent Reconstruction for Reinforcement Learning
Tao Yu · Zhizheng Zhang · Cuiling Lan · Yan Lu · Zhibo Chen

Tue Nov 29 09:00 AM -- 11:00 AM (PST) @ Hall J #133

For deep reinforcement learning (RL) from pixels, learning effective state representations is crucial for achieving high performance. However, in practice, limited experience and high-dimensional inputs prevent effective representation learning. To address this, motivated by the success of mask-based modeling in other research fields, we introduce mask-based reconstruction to promote state representation learning in RL. Specifically, we propose a simple yet effective self-supervised method, Mask-based Latent Reconstruction (MLR), to predict complete state representations in the latent space from the observations with spatially and temporally masked pixels. MLR enables better use of context information when learning state representations to make them more informative, which facilitates the training of RL agents. Extensive experiments show that our MLR significantly improves the sample efficiency in RL and outperforms the state-of-the-art sample-efficient RL methods on multiple continuous and discrete control benchmarks. Our code is available at https://github.com/microsoft/Mask-based-Latent-Reconstruction.

Author Information

Tao Yu (USTC)
Zhizheng Zhang (Microsoft Research)
Cuiling Lan (Microsoft)
Yan Lu (Microsoft Research Asia)
Zhibo Chen (University of Science and Technology of China)

More from the Same Authors