论文标题
朝着可区分性和多样性:在标签不足的情况下最大化批处理
Towards Discriminability and Diversity: Batch Nuclear-norm Maximization under Label Insufficient Situations
论文作者
论文摘要
深层网络的学习在很大程度上依赖于人类宣传的标签的数据。在某些标签不足的情况下,绩效以高数据密度在决策边界上降低。一种常见的解决方案是将香农熵直接最小化,但是熵最小化引起的副作用,即预测多样性的降低,大多被忽略。为了解决此问题,我们重新分配了随机选择的数据批次的分类输出矩阵的结构。我们通过理论分析发现,预测性和多样性可以通过批处理输出矩阵的Frobenius-norm和等级分别测量。此外,核场是Frobenius-norm的上行,并且是基质等级的凸近似。因此,为了提高可区分性和多样性,我们提出了输出矩阵上的批处理核合需最大化(BNM)。 BNM可以在典型的标签学习场景不足之下提高学习,例如半监督学习,领域的适应和开放领域的识别。在这些任务上,广泛的实验结果表明,BNM胜过竞争对手,并且与现有知名方法合作。该代码可从https://github.com/cuishuhao/bnm获得。
The learning of the deep networks largely relies on the data with human-annotated labels. In some label insufficient situations, the performance degrades on the decision boundary with high data density. A common solution is to directly minimize the Shannon Entropy, but the side effect caused by entropy minimization, i.e., reduction of the prediction diversity, is mostly ignored. To address this issue, we reinvestigate the structure of classification output matrix of a randomly selected data batch. We find by theoretical analysis that the prediction discriminability and diversity could be separately measured by the Frobenius-norm and rank of the batch output matrix. Besides, the nuclear-norm is an upperbound of the Frobenius-norm, and a convex approximation of the matrix rank. Accordingly, to improve both discriminability and diversity, we propose Batch Nuclear-norm Maximization (BNM) on the output matrix. BNM could boost the learning under typical label insufficient learning scenarios, such as semi-supervised learning, domain adaptation and open domain recognition. On these tasks, extensive experimental results show that BNM outperforms competitors and works well with existing well-known methods. The code is available at https://github.com/cuishuhao/BNM.