Skip to yearly menu bar Skip to main content


Oral
in
Workshop: Machine Learning with New Compute Paradigms

Bayesian Metaplasticity from Synaptic Uncertainty

Djohan Bonnet · Tifenn HIRTZLIN · Tarcisius Januel · Thomas Dalgaty · Damien Querlioz · Elisa Vianello


Abstract:

Catastrophic forgetting remains a challenge for neural networks, especially in lifelong learning scenarios. In this study, we introduce MEtaplasticity from Synaptic Uncertainty (MESU), inspired by metaplasticity and Bayesian inference principles. MESU harnesses synaptic uncertainty to retain information over time, with its update rule closely approximating the diagonal Newton's method for synaptic updates. Through continual learning experiments on permuted MNIST tasks, we demonstrate MESU's remarkable capability to maintain learning performance across 100 tasks without the need of explicit task boundaries.

Chat is not available.