Timezone: »
We explore task-free continual learning (CL), in which a model is trained to avoid catastrophic forgetting in the absence of explicit task boundaries or identities. Among many efforts on task-free CL, a notable family of approaches are memory-based that store and replay a subset of training examples. However, the utility of stored seen examples may diminish over time since CL models are continually updated. Here, we propose Gradient based Memory EDiting (GMED), a framework for editing stored examples in continuous input space via gradient updates, in order to create more "challenging" examples for replay. GMED-edited examples remain similar to their unedited forms, but can yield increased loss in the upcoming model updates, thereby making the future replays more effective in overcoming catastrophic forgetting. By construction, GMED can be seamlessly applied in conjunction with other memory-based CL algorithms to bring further improvement. Experiments validate the effectiveness of GMED, and our best method significantly outperforms baselines and previous state-of-the-art on five out of six datasets.
Author Information
Xisen Jin (University of Southern California)
Arka Sadhu (University of Southern California)
Junyi Du (University of Southern California)
Xiang Ren (University of Southern California)
More from the Same Authors
-
2020 : Poster #2 »
Xiang Ren -
2021 Spotlight: Refining Language Models with Compositional Explanations »
Huihan Yao · Ying Chen · Qinyuan Ye · Xisen Jin · Xiang Ren -
2022 : PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales »
Peifeng Wang · Aaron Chan · Filip Ilievski · Muhao Chen · Xiang Ren -
2022 : Adaptive Pre-training of Language Models for Better Logical Reasoning »
Soumya Sanyal · Yichong Xu · Shuohang Wang · Ziyi Yang · Reid Pryzant · Wenhao Yu · Chenguang Zhu · Xiang Ren -
2022 : SPRINT: Scalable Semantic Policy Pre-training via Language Instruction Relabeling »
Jesse Zhang · Karl Pertsch · Jiahui Zhang · Taewook Nam · Sung Ju Hwang · Xiang Ren · Joseph Lim -
2022 : Information-Theoretic Evaluation of Free-Text Rationales with Conditional $\mathcal{V}$-Information »
Hanjie Chen · Faeze Brahman · Xiang Ren · Yangfeng Ji · Yejin Choi · Swabha Swayamdipta -
2022 : PINTO: Faithful Language Reasoning Using Prompt-Generated Rationales »
Peifeng Wang · Aaron Chan · Filip Ilievski · Muhao Chen · Xiang Ren -
2022 : SPRINT: Scalable Semantic Policy Pre-training via Language Instruction Relabeling »
Jesse Zhang · Karl Pertsch · Jiahui Zhang · Taewook Nam · Sung Ju Hwang · Xiang Ren · Joseph Lim -
2022 Poster: NS3: Neuro-symbolic Semantic Code Search »
Shushan Arakelyan · Anna Hakhverdyan · Miltiadis Allamanis · Luis Garcia · Christophe Hauser · Xiang Ren -
2022 Poster: Unsupervised Cross-Task Generalization via Retrieval Augmentation »
Bill Yuchen Lin · Kangmin Tan · Chris Miller · Beiwen Tian · Xiang Ren -
2021 Poster: SalKG: Learning From Knowledge Graph Explanations for Commonsense Reasoning »
Aaron Chan · Jiashu Xu · Boyuan Long · Soumya Sanyal · Tanishq Gupta · Xiang Ren -
2021 Poster: Refining Language Models with Compositional Explanations »
Huihan Yao · Ying Chen · Qinyuan Ye · Xisen Jin · Xiang Ren -
2018 Poster: Hierarchical Graph Representation Learning with Differentiable Pooling »
Zhitao Ying · Jiaxuan You · Christopher Morris · Xiang Ren · Will Hamilton · Jure Leskovec -
2018 Spotlight: Hierarchical Graph Representation Learning with Differentiable Pooling »
Zhitao Ying · Jiaxuan You · Christopher Morris · Xiang Ren · Will Hamilton · Jure Leskovec