论文标题
DNN作为合作分类器的层次
DNNs as Layers of Cooperating Classifiers
论文作者
论文摘要
在一般情况下,可以描述和预测深神经网络(DNN)的概括能力的强大理论框架仍然难以捉摸。经典尝试产生了复杂性指标,这些指标在很大程度上依赖于全球紧凑性和能力的度量,而对亚组件协作的影响很少研究。我们在完全连接的前馈网络中展示了隐藏节点的激活模式的有趣规律。通过追踪这些模式的起源,我们展示了如何将这些网络视为两个信息处理系统的组合:一种连续和一个离散。我们描述了这两个系统如何自然地来自基于梯度的优化过程,并在单独和协作中演示了两个系统的分类能力。对DNN分类的这种观点为思考概括提供了一种新颖的方式,其中使用了不同的培训数据子集来训练不同的分类器。然后将这些分类器组合在一起以执行分类任务,它们的一致性对于准确的分类至关重要。
A robust theoretical framework that can describe and predict the generalization ability of deep neural networks (DNNs) in general circumstances remains elusive. Classical attempts have produced complexity metrics that rely heavily on global measures of compactness and capacity with little investigation into the effects of sub-component collaboration. We demonstrate intriguing regularities in the activation patterns of the hidden nodes within fully-connected feedforward networks. By tracing the origin of these patterns, we show how such networks can be viewed as the combination of two information processing systems: one continuous and one discrete. We describe how these two systems arise naturally from the gradient-based optimization process, and demonstrate the classification ability of the two systems, individually and in collaboration. This perspective on DNN classification offers a novel way to think about generalization, in which different subsets of the training data are used to train distinct classifiers; those classifiers are then combined to perform the classification task, and their consistency is crucial for accurate classification.