Timezone: »
We establish generalization error bounds for stochastic gradient Langevin dynamics (SGLD) with constant learning rate under the assumptions of dissipativity and smoothness, a setting that has received increased attention in the sampling/optimization literature. Unlike existing bounds for SGLD in non-convex settings, ours are time-independent and decay to zero as the sample size increases. Using the framework of uniform stability, we establish time-independent bounds by exploiting the Wasserstein contraction property of the Langevin diffusion, which also allows us to circumvent the need to bound gradients using Lipschitz-like assumptions. Our analysis also supports variants of SGLD that use different discretization methods, incorporate Euclidean projections, or use non-isotropic noise.
Author Information
Tyler Farghly (University of Oxford)
Patrick Rebeschini (University of Oxford)
More from the Same Authors
-
2023 Poster: A Novel Framework for Policy Mirror Descent with General Parametrization and Linear Convergence »
Carlo Alfano · Rui Yuan · Patrick Rebeschini -
2023 Poster: Optimal Convergence Rate for Exact Policy Mirror Descent in Discounted Markov Decision Processes »
Emmeran Johnson · Ciara Pike-Burke · Patrick Rebeschini -
2021 Poster: Implicit Regularization in Matrix Sensing via Mirror Descent »
Fan Wu · Patrick Rebeschini -
2021 Poster: Distributed Machine Learning with Sparse Heterogeneous Data »
Dominic Richards · Sahand Negahban · Patrick Rebeschini -
2021 Poster: On Optimal Interpolation in Linear Regression »
Eduard Oravkin · Patrick Rebeschini -
2020 Poster: A Continuous-Time Mirror Descent Approach to Sparse Phase Retrieval »
Fan Wu · Patrick Rebeschini -
2020 Spotlight: A Continuous-Time Mirror Descent Approach to Sparse Phase Retrieval »
Fan Wu · Patrick Rebeschini -
2020 Poster: The Statistical Complexity of Early-Stopped Mirror Descent »
Tomas Vaskevicius · Varun Kanade · Patrick Rebeschini -
2020 Spotlight: The Statistical Complexity of Early-Stopped Mirror Descent »
Tomas Vaskevicius · Varun Kanade · Patrick Rebeschini -
2019 Poster: Implicit Regularization for Optimal Sparse Recovery »
Tomas Vaskevicius · Varun Kanade · Patrick Rebeschini -
2019 Poster: Optimal Statistical Rates for Decentralised Non-Parametric Regression with Linear Speed-Up »
Dominic Richards · Patrick Rebeschini -
2019 Poster: Decentralized Cooperative Stochastic Bandits »
David MartÃnez-Rubio · Varun Kanade · Patrick Rebeschini -
2017 Poster: Accelerated consensus via Min-Sum Splitting »
Patrick Rebeschini · Sekhar C Tatikonda