论文标题

IIRC:增量隐式分类

IIRC: Incremental Implicitly-Refined Classification

论文作者

Abdelsalam, Mohamed, Faramarzi, Mojtaba, Sodhani, Shagun, Chandar, Sarath

论文摘要

我们介绍了“增量隐式分类的分类(IIRC)”设置,这是类增量学习设置的扩展,其中输入批次的类别具有两个粒度级别。即,每个样品都可以具有“熊”等高级(粗)标签和一个低级(细)标签,例如“北极熊”。一次只提供一个标签,该模型必须弄清楚它是否已经学习了它。这种设置更与现实生活中的场景更加一致,在该场景中,学习者通常会多次与同一实体互动,发现对它们的粒度更大,同时仍然试图不忘记以前的知识。此外,该设置可以评估模型,以解决一些重要的终身学习挑战,这些挑战在现有设置下无法轻易解决。这些挑战可以由一个示例来激发:“如果在一项任务中对班级的模型进行了训练,而在另一个任务中,它会忘记熊的概念,它会正确地推断出北极熊仍然是熊吗?它会将北极熊的标签与其他熊的品种错误地关联吗?”我们开发了标准化的基准测试,该基准可以在IIRC设置上评估模型。我们评估了几种最先进的终身学习算法,并突出了它们的优势和局限性。例如,基于蒸馏的方法的性能相对较好,但很容易预测每个图像的标签太多。我们希望拟议的设置以及基准将为从业者提供有意义的问题。

We introduce the "Incremental Implicitly-Refined Classi-fication (IIRC)" setup, an extension to the class incremental learning setup where the incoming batches of classes have two granularity levels. i.e., each sample could have a high-level (coarse) label like "bear" and a low-level (fine) label like "polar bear". Only one label is provided at a time, and the model has to figure out the other label if it has already learnfed it. This setup is more aligned with real-life scenarios, where a learner usually interacts with the same family of entities multiple times, discovers more granularity about them, while still trying not to forget previous knowledge. Moreover, this setup enables evaluating models for some important lifelong learning challenges that cannot be easily addressed under the existing setups. These challenges can be motivated by the example "if a model was trained on the class bear in one task and on polar bear in another task, will it forget the concept of bear, will it rightfully infer that a polar bear is still a bear? and will it wrongfully associate the label of polar bear to other breeds of bear?". We develop a standardized benchmark that enables evaluating models on the IIRC setup. We evaluate several state-of-the-art lifelong learning algorithms and highlight their strengths and limitations. For example, distillation-based methods perform relatively well but are prone to incorrectly predicting too many labels per image. We hope that the proposed setup, along with the benchmark, would provide a meaningful problem setting to the practitioners

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源