Skip to yearly menu bar Skip to main content


Poster
in
Workshop: INTERPOLATE — First Workshop on Interpolation Regularizers and Beyond

Overparameterization Implicitly Regularizes Input-Space Smoothness

Matteo Gamba · Hossein Azizpour · Mårten Björkman

Keywords: [ Lipschitz constant ] [ deep networks ] [ Double descent ] [ Smoothness ]


Abstract:

Existing bounds on the generalization error of deep networks assume some form of smooth or bounded dependence on the input variable and intermediate activations, falling short of investigating the mechanisms controlling such factors in practice. In this work, we present an empirical study of the Lipschitz constant of networks trained in practice, as the number of model parameters and training epochs vary. We present non-monotonic trends for the Lipschitz constant, strongly correlating with double descent for the test error. Our findings highlight a theoretical shortcoming in modeling input-space smoothness via monotonic bounds.

Chat is not available.