Skip to yearly menu bar Skip to main content


Poster

Small steps no more: Global convergence of stochastic gradient bandits for arbitrary learning rates

Jincheng Mei · Bo Dai · Alekh Agarwal · Sharan Vaswani · Anant Raj · Csaba Szepesvari · Dale Schuurmans

West Ballroom A-D #6701
[ ]
Thu 12 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

We provide a new understanding of the stochastic gradient bandit algorithm by showing that it converges to a globally optimal policy almost surely using \emph{any} constant learning rate. This result demonstrates that the stochastic gradient algorithm continues to balance exploration and exploitation appropriately even in scenarios where standard smoothness and noise control assumptions break down. The proofs are based on novel findings about action sampling rates and the relationship between cumulative progress and noise, and extend the current understanding of how simple stochastic gradient methods behave in bandit settings.

Live content is unavailable. Log in and register to view live content