Skip to yearly menu bar Skip to main content


Poster

Dynamic Local Regret for Non-convex Online Forecasting

Sergul Aydore · Tianhao Zhu · Dean Foster

East Exhibition Hall B + C #20

Keywords: [ Algorithms -> Online Learning; Deep Learning -> Optimization for Deep Networks; Deep Learning ] [ Supervised Deep Networks ] [ Time Series Analysis ] [ Applications ]


Abstract:

We consider online forecasting problems for non-convex machine learning models. Forecasting introduces several challenges such as (i) frequent updates are necessary to deal with concept drift issues since the dynamics of the environment change over time, and (ii) the state of the art models are non-convex models. We address these challenges with a novel regret framework. Standard regret measures commonly do not consider both dynamic environment and non-convex models. We introduce a local regret for non-convex models in a dynamic environment. We present an update rule incurring a cost, according to our proposed local regret, which is sublinear in time T. Our update uses time-smoothed gradients. Using a real-world dataset we show that our time-smoothed approach yields several benefits when compared with state-of-the-art competitors: results are more stable against new data; training is more robust to hyperparameter selection; and our approach is more computationally efficient than the alternatives.

Live content is unavailable. Log in and register to view live content