Timezone: »
We prove that ϴ(k d^2 / ε^2) samples are necessary and sufficient for learning a mixture of k Gaussians in R^d, up to error ε in total variation distance. This improves both the known upper bounds and lower bounds for this problem. For mixtures of axis-aligned Gaussians, we show that O(k d / ε^2) samples suffice, matching a known lower bound.
The upper bound is based on a novel technique for distribution learning based on a notion of sample compression. Any class of distributions that allows such a sample compression scheme can also be learned with few samples. Moreover, if a class of distributions has such a compression scheme, then so do the classes of products and mixtures of those distributions. The core of our main result is showing that the class of Gaussians in R^d has an efficient sample compression.
Author Information
Hassan Ashtiani (McMaster University)
Shai Ben-David (University of Waterloo)
Nicholas Harvey (University of British Columbia)
Christopher Liaw (University of British Columbia)
Abbas Mehrabian (Mcgill University)
Yaniv Plan (University of British Columbia)
Related Events (a corresponding poster, oral, or spotlight)
-
2018 Oral: Nearly tight sample complexity bounds for learning mixtures of Gaussians via sample compression schemes »
Wed. Dec 5th 03:05 -- 03:20 PM Room Room 517 CD
More from the Same Authors
-
2021 Spotlight: PLUGIn: A simple algorithm for inverting generative models with recovery guarantees »
Babhru Joshi · Xiaowei Li · Yaniv Plan · Ozgur Yilmaz -
2022 Poster: Benefits of Additive Noise in Composing Classes with Bounded Capacity »
Alireza Fathollah Pour · Hassan Ashtiani -
2022 Spotlight: Benefits of Additive Noise in Composing Classes with Bounded Capacity »
Alireza Fathollah Pour · Hassan Ashtiani -
2022 Spotlight: Lightning Talks 2A-1 »
Caio Kalil Lauand · Ryan Strauss · Yasong Feng · lingyu gu · Alireza Fathollah Pour · Oren Mangoubi · Jianhao Ma · Binghui Li · Hassan Ashtiani · Yongqi Du · Salar Fattahi · Sean Meyn · Jikai Jin · Nisheeth Vishnoi · zengfeng Huang · Junier B Oliva · yuan zhang · Han Zhong · Tianyu Wang · John Hopcroft · Di Xie · Shiliang Pu · Liwei Wang · Robert Qiu · Zhenyu Liao -
2021 Poster: PLUGIn: A simple algorithm for inverting generative models with recovery guarantees »
Babhru Joshi · Xiaowei Li · Yaniv Plan · Ozgur Yilmaz -
2021 Poster: Privately Learning Mixtures of Axis-Aligned Gaussians »
Ishaq Aden-Ali · Hassan Ashtiani · Christopher Liaw -
2020 Poster: Improved Algorithms for Online Submodular Maximization via First-order Regret Bounds »
Nicholas Harvey · Christopher Liaw · Tasuku Soma -
2020 Poster: Regret Bounds without Lipschitz Continuity: Online Learning with Relative-Lipschitz Losses »
Yihan Zhou · Victor Sanches Portella · Mark Schmidt · Nicholas Harvey -
2019 Poster: Disentangled behavioural representations »
Amir Dezfouli · Hassan Ashtiani · Omar Ghattas · Richard Nock · Peter Dayan · Cheng Soon Ong -
2016 Poster: Average-case hardness of RIP certification »
Tengyao Wang · Quentin Berthet · Yaniv Plan