Poster
Small steps no more: Global convergence of stochastic gradient bandits for arbitrary learning rates
Jincheng Mei · Bo Dai · Alekh Agarwal · Sharan Vaswani · Anant Raj · Csaba Szepesvari · Dale Schuurmans
West Ballroom A-D #6701
[
Abstract
]
Thu 12 Dec 11 a.m. PST
— 2 p.m. PST
Abstract:
We provide a new understanding of the stochastic gradient bandit algorithm by showing that it converges to a globally optimal policy almost surely using \emph{any} constant learning rate. This result demonstrates that the stochastic gradient algorithm continues to balance exploration and exploitation appropriately even in scenarios where standard smoothness and noise control assumptions break down. The proofs are based on novel findings about action sampling rates and the relationship between cumulative progress and noise, and extend the current understanding of how simple stochastic gradient methods behave in bandit settings.
Live content is unavailable. Log in and register to view live content