论文标题

部分可观测时空混沌系统的无模型预测

Uncertainty-aware Contrastive Distillation for Incremental Semantic Segmentation

论文作者

Yang, Guanglei, Fini, Enrico, Xu, Dan, Rota, Paolo, Ding, Mingli, Nabi, Moin, Alameda-Pineda, Xavier, Ricci, Elisa

论文摘要

深度学习中的一个基本和挑战性的问题是灾难性的遗忘,即神经网络在学习新任务时无法从旧任务中获取的知识的趋势。在研究界已经广泛研究了这个问题,并且在过去几年中提出了几种增量学习(IL)方法。尽管较早的计算机视觉作品主要集中在图像分类和对象检测上,但最近引入了某些IL语义分割方法。这些先前的作品表明,尽管具有简单性,但知识蒸馏可以有效地减轻灾难性的遗忘。在本文中,我们遵循了这个研究方向,并受到对比度学习的最新文献的启发,我们提出了一个新颖的蒸馏框架,不确定性意识到的对比蒸馏(\方法)。简而言之,\方法〜是通过引入新颖的蒸馏损失来操作的,该蒸馏损失考虑了迷你批次中的所有图像,从同一类中实现了与所有像素相关的特征之间的相似性,并将与不同类别的像素相对应的那些图像相似。为了减轻灾难性的遗忘,我们将新模型的特征与在上一个增量步骤中学到的冷冻模型提取的功能进行了对比。我们的实验结果证明了所提出的蒸馏技术的优势,该技术可与先前的IL方法协同使用,并在三个常见的基准测试基准上进行递增的语义分割。该代码可在\ url {https://github.com/ygjwd12345/ucd}中获得。

A fundamental and challenging problem in deep learning is catastrophic forgetting, i.e. the tendency of neural networks to fail to preserve the knowledge acquired from old tasks when learning new tasks. This problem has been widely investigated in the research community and several Incremental Learning (IL) approaches have been proposed in the past years. While earlier works in computer vision have mostly focused on image classification and object detection, more recently some IL approaches for semantic segmentation have been introduced. These previous works showed that, despite its simplicity, knowledge distillation can be effectively employed to alleviate catastrophic forgetting. In this paper, we follow this research direction and, inspired by recent literature on contrastive learning, we propose a novel distillation framework, Uncertainty-aware Contrastive Distillation (\method). In a nutshell, \method~is operated by introducing a novel distillation loss that takes into account all the images in a mini-batch, enforcing similarity between features associated to all the pixels from the same classes, and pulling apart those corresponding to pixels from different classes. In order to mitigate catastrophic forgetting, we contrast features of the new model with features extracted by a frozen model learned at the previous incremental step. Our experimental results demonstrate the advantage of the proposed distillation technique, which can be used in synergy with previous IL approaches, and leads to state-of-art performance on three commonly adopted benchmarks for incremental semantic segmentation. The code is available at \url{https://github.com/ygjwd12345/UCD}.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源