The Forward-Forward Algorithm for Training Deep Neural Networks
Moderator : Kyunghyun Cho
Hall H (level 1)
I will describe a training algorithm for deep neural networks that does not require the neurons to propagate derivatives or remember neural activities. The algorithm can learn multi-level representations of streaming sensory data on the fly without interrupting the processing of the input stream. The algorithm scales much better than reinforcement learning and would be much easier to implement in cortex than backpropagation.