Skip to yearly menu bar Skip to main content


Poster

Layer-wise analysis of deep networks with Gaussian kernels

Grégoire Montavon · Mikio L Braun · Klaus-Robert Müller


Abstract:

Deep networks can potentially express a learning problem more efficiently than local learning machines. While deep networks outperform local learning machines on some problems, it is still unclear how their nice representation emerges from their complex structure. We present an analysis based on Gaussian kernels that measures how the representation of the learning problem evolves layer after layer as the deep network builds higher-level abstract representations of the input. We use this analysis to show empirically that deep networks build progressively better representations of the learning problem and that the best representations are obtained when the deep network discriminates only in the last layers.

Live content is unavailable. Log in and register to view live content