Poster
in
Workshop: Bayesian Decision-making and Uncertainty: from probabilistic and spatiotemporal modeling to sequential experiment design
Scaling Gaussian Processes for Learning Curve Prediction via Latent Kronecker Structure
Jihao Andreas Lin · Sebastian Ament · Maximilian Balandat · Eytan Bakshy
Keywords: [ learning curve ] [ Kronecker ] [ Gaussian process ] [ latent ] [ AutoML ]
Abstract:
A key task in AutoML is to model learning curves of machine learning models jointly as a function of model hyper-parameters and training progression. While Gaussian processes (GPs) are suitable for this task, naive GPs require O(n3m3)O(n3m3) time and O(n2m2)O(n2m2) space for nn hyper-parameter configurations and O(m)O(m) learning curve observations per hyper-parameter. Efficient inference via Kronecker structure is typically incompatible with early-stopping due to missing learning curve values. We impose latent Kronecker structurelatent Kronecker structure to leverage efficient product kernels while handling missing values. In particular, we interpret the joint covariance matrix of observed values as the projection of a latent Kronecker product. Combined with iterative linear solvers and structured matrix-vector multiplication, our method only requires O(n3+m3)O(n3+m3) time and O(n2+m2)O(n2+m2) space. We show that our GP model can match the performance of a Transformer on a learning curve prediction task.
Chat is not available.