Bandit Linear Control
Asaf Cassel, Tomer Koren
Spotlight presentation: Orals & Spotlights Track 31: Reinforcement Learning
on 2020-12-10T07:00:00-08:00 - 2020-12-10T07:10:00-08:00
on 2020-12-10T07:00:00-08:00 - 2020-12-10T07:10:00-08:00
Toggle Abstract Paper (in Proceedings / .pdf)
Abstract: We consider the problem of controlling a known linear dynamical system under stochastic noise, adversarially chosen costs, and bandit feedback. Unlike the full feedback setting where the entire cost function is revealed after each decision, here only the cost incurred by the learner is observed. We present a new and efficient algorithm that, for strongly convex and smooth costs, obtains regret that grows with the square root of the time horizon T. We also give extensions of this result to general convex, possibly non-smooth costs, and to non-stochastic system noise. A key component of our algorithm is a new technique for addressing bandit optimization of loss functions with memory.