论文标题
从有限的样本中总的来
Aggregative Self-Supervised Feature Learning from a Limited Sample
论文作者
论文摘要
自我监督学习(SSL)是一种有效的方法,可以解决培训数据和注释短缺的问题。 SSL的关键部分是其代理任务,该任务定义了监督信号,并将学习推向有效的功能表示。但是,大多数SSL方法通常集中在单个代理任务上,这极大地限制了学习特征的表达能力,因此使网络泛化能力恶化。在这方面,我们在此提出了两种汇总策略,以各种形式的互补性来提高自我监督的学习特征的稳健性。我们首先提出了一个从有限的样本中的多任务汇总自我监督学习的原则性框架,以形成统一的表示形式,并意图在不同任务之间利用特征互补性。然后,在自我凝集SSL中,我们建议根据基于线性中心的内核对齐度量指标,具有辅助损耗函数的现有代理任务,该指标明确促进了从手头的代理任务中学到的功能,以进一步促进了模型能力。在有限的数据和注释方案下,我们对2D自然图像和3D医学图像分类任务进行的广泛实验证实,提出的聚合策略成功提高了分类精度。
Self-supervised learning (SSL) is an efficient approach that addresses the issue of limited training data and annotation shortage. The key part in SSL is its proxy task that defines the supervisory signals and drives the learning toward effective feature representations. However, most SSL approaches usually focus on a single proxy task, which greatly limits the expressive power of the learned features and therefore deteriorates the network generalization capacity. In this regard, we hereby propose two strategies of aggregation in terms of complementarity of various forms to boost the robustness of self-supervised learned features. We firstly propose a principled framework of multi-task aggregative self-supervised learning from a limited sample to form a unified representation, with an intent of exploiting feature complementarity among different tasks. Then, in self-aggregative SSL, we propose to self-complement an existing proxy task with an auxiliary loss function based on a linear centered kernel alignment metric, which explicitly promotes the exploring of where are uncovered by the features learned from a proxy task at hand to further boost the modeling capability. Our extensive experiments on 2D natural image and 3D medical image classification tasks under limited data and annotation scenarios confirm that the proposed aggregation strategies successfully boost the classification accuracy.