Skip to yearly menu bar Skip to main content


Poster

FALKON: An Optimal Large Scale Kernel Method

Alessandro Rudi · Luigi Carratino · Lorenzo Rosasco

Pacific Ballroom #59

Keywords: [ Kernel Methods ] [ Learning Theory ] [ Regularization ] [ Spaces of Functions and Kernels ]

[ ]
[ Code [ Paper

Abstract: Kernel methods provide a principled way to perform non linear, nonparametric learning. They rely on solid functional analytic foundations and enjoy optimal statistical properties. However, at least in their basic form, they have limited applicability in large scale scenarios because of stringent computational requirements in terms of time and especially memory. In this paper, we take a substantial step in scaling up kernel methods, proposing FALKON, a novel algorithm that allows to efficiently process millions of points. FALKON is derived combining several algorithmic principles, namely stochastic subsampling, iterative solvers and preconditioning. Our theoretical analysis shows that optimal statistical accuracy is achieved requiring essentially $O(n)$ memory and $O(n\sqrt{n})$ time. An extensive experimental analysis on large scale datasets shows that, even with a single machine, FALKON outperforms previous state of the art solutions, which exploit parallel/distributed architectures.

Chat is not available.