Timezone: »
Despite recent advances in its theoretical understanding, there still remains a significant gap in the ability of existing PAC-Bayesian theories on meta-learning to explain performance improvements in the few-shot learning setting, where the number of training examples in the target tasks is severely limited. This gap originates from an assumption in the existing theories which supposes that the number of training examples in the observed tasks and the number of training examples in the target tasks follow the same distribution, an assumption that rarely holds in practice. By relaxing this assumption, we develop two PAC-Bayesian bounds tailored for the few-shot learning setting and show that two existing meta-learning algorithms (MAML and Reptile) can be derived from our bounds, thereby bridging the gap between practice and PAC-Bayesian theories. Furthermore, we derive a new computationally-efficient PACMAML algorithm, and show it outperforms existing meta-learning algorithms on several few-shot benchmark datasets.
Author Information
Nan Ding (Google)
Xi Chen (Harvard University)
Tomer Levinboim (University of Southern California)
Sebastian Goodman (Google)
Radu Soricut (Google)
More from the Same Authors
-
2017 Poster: Cold-Start Reinforcement Learning with Softmax Policy Gradient »
Nan Ding · Radu Soricut -
2016 Poster: Stochastic Gradient MCMC with Stale Gradients »
Changyou Chen · Nan Ding · Chunyuan Li · Yizhe Zhang · Lawrence Carin -
2015 Poster: Embedding Inference for Structured Multilabel Prediction »
Farzaneh Mirzazadeh · Siamak Ravanbakhsh · Nan Ding · Dale Schuurmans -
2014 Poster: Bayesian Sampling Using Stochastic Gradient Thermostats »
Nan Ding · Youhan Fang · Ryan Babbush · Changyou Chen · Robert D Skeel · Hartmut Neven -
2011 Poster: t-divergence Based Approximate Inference »
Nan Ding · S.V.N. Vishwanathan · Yuan Qi -
2010 Poster: t-logistic regression »
Nan Ding · S.V.N. Vishwanathan