Skip to yearly menu bar Skip to main content


Poster

Adaptive Exploration for Data-Efficient General Value Function Evaluations

Arushi Jain · Josiah Hanna · Doina Precup


Abstract: General Value Functions (GVFs) (Sutton et al., 2011) represent predictive knowledge in reinforcement learning. Each GVF computes the expected return for a given policy, based on a unique reward. Existing methods relying on fixed behavior policies or pre-collected data often face data efficiency issues when learning multiple GVFs in parallel using off-policy methods. To address this, we introduce $GVFExplorer$, which adaptively learns a single behavior policy that efficiently collects data for evaluating multiple GVFs in parallel. Our method optimizes the behavior policy by minimizing the total variance in return across GVFs, thereby reducing the required environmental interactions We use an existing temporal-difference-style variance estimator to approximate the return variance. We prove that each behavior policy update decreases the overall mean squared error in GVF predictions. We empirically show our method's performance in tabular and nonlinear function approximation settings, with stationary and non-stationary reward signals, optimizing data usage and reducing prediction errors across multiple GVFs.

Live content is unavailable. Log in and register to view live content