Skip to yearly menu bar Skip to main content


Poster

A Nonconvex Optimization Framework for Low Rank Matrix Estimation

Tuo Zhao · Zhaoran Wang · Han Liu

210 C #101

Abstract:

We study the estimation of low rank matrices via nonconvex optimization. Compared with convex relaxation, nonconvex optimization exhibits superior empirical performance for large scale instances of low rank matrix estimation. However, the understanding of its theoretical guarantees are limited. In this paper, we define the notion of projected oracle divergence based on which we establish sufficient conditions for the success of nonconvex optimization. We illustrate the consequences of this general framework for matrix sensing and completion. In particular, we prove that a broad class of nonconvex optimization algorithms, including alternating minimization and gradient-type methods, geometrically converge to the global optimum and exactly recover the true low rank matrices under standard conditions.

Live content is unavailable. Log in and register to view live content