Skip to yearly menu bar Skip to main content


Poster

Is Input Sparsity Time Possible for Kernel Low-Rank Approximation?

Cameron Musco · David Woodruff

Pacific Ballroom #208

Keywords: [ Kernel Methods ] [ Matrix and Tensor Factorization ] [ Hardness of Learning and Approximations ] [ Computational Complexity ]


Abstract: Low-rank approximation is a common tool used to accelerate kernel methods: the n×nn×n kernel matrix KK is approximated via a rank-kk matrix ˜K~K which can be stored in much less space and processed more quickly. In this work we study the limits of computationally efficient low-rank kernel approximation. We show that for a broad class of kernels, including the popular Gaussian and polynomial kernels, computing a relative error kk-rank approximation to KK is at least as difficult as multiplying the input data matrix ARn×dARn×d by an arbitrary matrix CRd×kCRd×k. Barring a breakthrough in fast matrix multiplication, when kk is not too large, this requires Ω(nnz(A)k)Ω(nnz(A)k) time where nnz(A)nnz(A) is the number of non-zeros in AA. This lower bound matches, in many parameter regimes, recent work on subquadratic time algorithms for low-rank approximation of general kernels [MM16,MW17], demonstrating that these algorithms are unlikely to be significantly improved, in particular to O(nnz(A))O(nnz(A)) input sparsity runtimes. At the same time there is hope: we show for the first time that O(nnz(A))O(nnz(A)) time approximation is possible for general radial basis function kernels (e.g., the Gaussian kernel) for the closely related problem of low-rank approximation of the kernelized dataset.

Live content is unavailable. Log in and register to view live content