论文标题

终身学习的持续评估:确定稳定差距

Continual evaluation for lifelong learning: Identifying the stability gap

论文作者

De Lange, Matthias, van de Ven, Gido, Tuytelaars, Tinne

论文摘要

事实证明,时间依赖于时间依赖的数据生成分布很难对神经网络进行基于梯度的培训,因为贪婪的更新会导致灾难性忘记以前学习的知识。尽管在不断学习以克服这一遗忘的领域取得了进步,但我们表明,一组通用的最新方法仍然因开始学习新任务而遭受重大遗忘,只是这种遗忘是暂时的,随后是绩效恢复的阶段。我们将这种有趣但可能有问题的现象称为稳定差距。由于在每个任务后仅评估持续学习模型的标准实践,因此稳定差距可能一直存在于雷达之下。取而代之的是,我们建立了一个使用触电评估的持续评估框架,并定义了一组新的指标来量化最坏情况性能。从经验上讲,我们证明了经验重播,基于约束的重播,知识依据和参数正则化方法都容易出现稳定差距。并且可以在类,任务和领域的学习基准中观察到稳定差距。此外,一个受控的实验表明,当任务更加不同时,稳定性差距会增加。最后,通过将梯度分解为可塑性和稳定性组件,我们提出了一个概念上的稳定性差距解释。

Time-dependent data-generating distributions have proven to be difficult for gradient-based training of neural networks, as the greedy updates result in catastrophic forgetting of previously learned knowledge. Despite the progress in the field of continual learning to overcome this forgetting, we show that a set of common state-of-the-art methods still suffers from substantial forgetting upon starting to learn new tasks, except that this forgetting is temporary and followed by a phase of performance recovery. We refer to this intriguing but potentially problematic phenomenon as the stability gap. The stability gap had likely remained under the radar due to standard practice in the field of evaluating continual learning models only after each task. Instead, we establish a framework for continual evaluation that uses per-iteration evaluation and we define a new set of metrics to quantify worst-case performance. Empirically we show that experience replay, constraint-based replay, knowledge-distillation, and parameter regularization methods are all prone to the stability gap; and that the stability gap can be observed in class-, task-, and domain-incremental learning benchmarks. Additionally, a controlled experiment shows that the stability gap increases when tasks are more dissimilar. Finally, by disentangling gradients into plasticity and stability components, we propose a conceptual explanation for the stability gap.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源