Skip to yearly menu bar Skip to main content


Poster

Coordinate Descent with Bandit Sampling

Farnood Salehi · Patrick Thiran · Elisa Celis

Room 210 #40

Keywords: [ Convex Optimization ]


Abstract:

Coordinate descent methods minimize a cost function by updating a single decision variable (corresponding to one coordinate) at a time. Ideally, we would update the decision variable that yields the largest marginal decrease in the cost function. However, finding this coordinate would require checking all of them, which is not computationally practical. Therefore, we propose a new adaptive method for coordinate descent. First, we define a lower bound on the decrease of the cost function when a coordinate is updated and, instead of calculating this lower bound for all coordinates, we use a multi-armed bandit algorithm to learn which coordinates result in the largest marginal decrease and simultaneously perform coordinate descent. We show that our approach improves the convergence of the coordinate methods both theoretically and experimentally.

Live content is unavailable. Log in and register to view live content