Timezone: »

GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration
Jacob Gardner · Geoff Pleiss · Kilian Weinberger · David Bindel · Andrew Wilson

Wed Dec 05 01:55 PM -- 02:00 PM (PST) @ Room 220 E

Despite advances in scalable models, the inference tools used for Gaussian processes (GPs) have yet to fully capitalize on developments in computing hardware. We present an efficient and general approach to GP inference based on Blackbox Matrix-Matrix multiplication (BBMM). BBMM inference uses a modified batched version of the conjugate gradients algorithm to derive all terms for training and inference in a single call. BBMM reduces the asymptotic complexity of exact GP inference from O(n^3) to O(n^2). Adapting this algorithm to scalable approximations and complex GP models simply requires a routine for efficient matrix-matrix multiplication with the kernel and its derivative. In addition, BBMM uses a specialized preconditioner to substantially speed up convergence. In experiments we show that BBMM effectively uses GPU hardware to dramatically accelerate both exact GP inference and scalable approximations. Additionally, we provide GPyTorch, a software platform for scalable GP inference via BBMM, built on PyTorch.

Author Information

Jacob Gardner (Cornell University)
Geoff Pleiss (Cornell University)
Kilian Weinberger (Cornell University)
David Bindel (Cornell University)
Andrew Wilson (Cornell University)
Andrew Wilson

I am a professor of machine learning at New York University.

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors