Timezone: »
Biological agents are known to learn many different tasks over the course of their lives, and to be able to revisit previous tasks and behaviors with little to no loss in performance. In contrast, artificial agents are prone to ‘catastrophic forgetting’ whereby performance on previous tasks deteriorates rapidly as new ones are acquired. This shortcoming has recently been addressed using methods that encourage parameters to stay close to those used for previous tasks. This can be done by (i) using specific parameter regularizers that map out suitable destinations in parameter space, or (ii) guiding the optimization journey by projecting gradients into subspaces that do not interfere with previous tasks. However, these methods often exhibit subpar performance in both feedforward and recurrent neural networks, with recurrent networks being of interest to the study of neural dynamics supporting biological continual learning. In this work, we propose Natural Continual Learning (NCL), a new method that unifies weight regularization and projected gradient descent. NCL uses Bayesian weight regularization to encourage good performance on all tasks at convergence and combines this with gradient projection using the prior precision, which prevents catastrophic forgetting during optimization. Our method outperforms both standard weight regularization techniques and projection based approaches when applied to continual learning problems in feedforward and recurrent networks. Finally, the trained networks evolve task-specific dynamics that are strongly preserved as new tasks are learned, similar to experimental findings in biological circuits.
Author Information
Ta-Chu Kao (Gatsby Unit, UCL)
Kristopher Jensen (University of Cambridge)
Gido van de Ven (University of Cambridge)
Alberto Bernacchia (MediaTek Research)
Guillaume Hennequin (University of Cambridge)
More from the Same Authors
-
2021 : How to distribute data across tasks for meta-learning? »
Alexandru Cioba · Michael Bromberg · Qian Wang · RITWIK NIYOGI · Georgios Batzolis · Jezabel Garcia · Da-shan Shiu · Alberto Bernacchia -
2022 : Gradient Descent: Robustness to Adversarial Corruption »
Fu-Chieh Chang · Farhang Nabiei · Pei-Yuan Wu · Alexandru Cioba · Sattar Vakili · Alberto Bernacchia -
2022 : Q & A »
Gido van de Ven · Tyler Hayes · Dhireesha Kudithipudi -
2022 : Approaches: current strategies for lifelong learning »
Gido van de Ven -
2022 : The problem: currently used settings and benchmarks »
Gido van de Ven -
2022 Tutorial: Lifelong Learning Machines »
Tyler Hayes · Dhireesha Kudithipudi · Gido van de Ven -
2022 : Panel Discussion II: Geometric and topological principles for representations in the brain »
Bruno Olshausen · Kristopher Jensen · Gabriel Kreiman · Manu Madhav · Christian A Shewmake -
2022 : Generative models of non-Euclidean neural population dynamics »
Kristopher Jensen -
2021 : Cyclic orthogonal convolutions for long-range integration of features »
Federica Freddi · Jezabel Garcia · Michael Bromberg · Sepehr Jalali · Da-shan Shiu · Alvin Chua · Alberto Bernacchia -
2021 Poster: Scalable Bayesian GPFA with automatic relevance determination and discrete noise models »
Kristopher Jensen · Ta-Chu Kao · Jasmine Stone · Guillaume Hennequin -
2021 Poster: Optimal Order Simple Regret for Gaussian Process Bandits »
Sattar Vakili · Nacime Bouziani · Sepehr Jalali · Alberto Bernacchia · Da-shan Shiu -
2020 Poster: Manifold GPLVMs for discovering non-Euclidean latent structure in neural data »
Kristopher Jensen · Ta-Chu Kao · Marco Tripodi · Guillaume Hennequin -
2020 Poster: Non-reversible Gaussian processes for identifying latent dynamical structure in neural data »
Virginia Rutten · Alberto Bernacchia · Maneesh Sahani · Guillaume Hennequin -
2020 Oral: Non-reversible Gaussian processes for identifying latent dynamical structure in neural data »
Virginia Rutten · Alberto Bernacchia · Maneesh Sahani · Guillaume Hennequin -
2018 Poster: Exact natural gradient in deep linear networks and its application to the nonlinear case »
Alberto Bernacchia · Mate Lengyel · Guillaume Hennequin