论文标题
基于彩排的持续学习的梯度匹配核心
Gradient-Matching Coresets for Rehearsal-Based Continual Learning
论文作者
论文摘要
持续学习(CL)的目的是有效地使用新数据更新机器学习模型,而不会忘记以前的知识。大多数广泛使用的CL方法依赖于在培训新数据时要重复使用的数据点的排练记忆。策划这样的排练记忆,以维持到目前为止所看到的所有数据的少量,信息丰富的子集,这对于这些方法的成功至关重要。我们为基于彩排的持续学习设计了一种核心选择方法。我们的方法基于梯度匹配的想法:核心引起的梯度应尽可能紧密地匹配原始训练数据集引起的梯度。受神经切线内核理论的启发,我们在模型的初始化分布中执行了这种梯度匹配,从而使我们能够提取核心而无需先训练模型。我们在各种持续学习方案上评估了该方法,并证明它与诸如储层采样等竞争记忆管理策略相比,它可以提高基于排练的CL方法的性能。
The goal of continual learning (CL) is to efficiently update a machine learning model with new data without forgetting previously-learned knowledge. Most widely-used CL methods rely on a rehearsal memory of data points to be reused while training on new data. Curating such a rehearsal memory to maintain a small, informative subset of all the data seen so far is crucial to the success of these methods. We devise a coreset selection method for rehearsal-based continual learning. Our method is based on the idea of gradient matching: The gradients induced by the coreset should match, as closely as possible, those induced by the original training dataset. Inspired by the neural tangent kernel theory, we perform this gradient matching across the model's initialization distribution, allowing us to extract a coreset without having to train the model first. We evaluate the method on a wide range of continual learning scenarios and demonstrate that it improves the performance of rehearsal-based CL methods compared to competing memory management strategies such as reservoir sampling.