This is the public, feature-limited version of the conference webpage. After Registration and login please visit the full version.

GAIT-prop: A biologically plausible learning rule derived from backpropagation of error

Nasir Ahmad, Marcel A. J. van Gerven, Luca Ambrogioni

Spotlight presentation: Orals & Spotlights Track 08: Deep Learning
on 2020-12-08T07:10:00-08:00 - 2020-12-08T07:20:00-08:00
Poster Session 2 (more posters)
on 2020-12-08T09:00:00-08:00 - 2020-12-08T11:00:00-08:00
Abstract: Traditional backpropagation of error, though a highly successful algorithm for learning in artificial neural network models, includes features which are biologically implausible for learning in real neural circuits. An alternative called target propagation proposes to solve this implausibility by using a top-down model of neural activity to convert an error at the output of a neural network into layer-wise and plausible ‘targets’ for every unit. These targets can then be used to produce weight updates for network training. However, thus far, target propagation has been heuristically proposed without demonstrable equivalence to backpropagation. Here, we derive an exact correspondence between backpropagation and a modified form of target propagation (GAIT-prop) where the target is a small perturbation of the forward pass. Specifically, backpropagation and GAIT-prop give identical updates when synaptic weight matrices are orthogonal. In a series of simple computer vision experiments, we show near-identical performance between backpropagation and GAIT-prop with a soft orthogonality-inducing regularizer.

Preview Video and Chat

To see video, interact with the author and ask questions please use registration and login.