论文标题
通过数据流形的框架镜头的视觉模型的内部表示
Internal Representations of Vision Models Through the Lens of Frames on Data Manifolds
论文作者
论文摘要
尽管过去五年在理解深度学习模型的内部表示方面取得了很大进展,但仍然存在许多问题。当试图了解模型设计选择的影响(例如模型体系结构或训练算法)对隐藏表示形状和动态的影响时,尤其如此。在这项工作中,我们提出了一种新的方法来研究这种表示形式,灵感来自于歧管的切线束上的框架的想法。我们称之为神经框架的构造是通过组装代表数据点的特定类型的驱动器的一组向量来形成的,例如,无限的增强,噪声扰动或生成模型产生的扰动,并研究这些通过网络时如何变化。使用神经框架,我们可以观察到模型在数据点的小社区中进行对处理的方式,逐层,特定的变化模式。我们的结果为许多现象提供了新的观点,例如,通过增强培训产生模型不变性或对抗性训练和模型概括之间的拟议权衡。
While the last five years have seen considerable progress in understanding the internal representations of deep learning models, many questions remain. This is especially true when trying to understand the impact of model design choices, such as model architecture or training algorithm, on hidden representation geometry and dynamics. In this work we present a new approach to studying such representations inspired by the idea of a frame on the tangent bundle of a manifold. Our construction, which we call a neural frame, is formed by assembling a set of vectors representing specific types of perturbations of a data point, for example infinitesimal augmentations, noise perturbations, or perturbations produced by a generative model, and studying how these change as they pass through a network. Using neural frames, we make observations about the way that models process, layer-by-layer, specific modes of variation within a small neighborhood of a datapoint. Our results provide new perspectives on a number of phenomena, such as the manner in which training with augmentation produces model invariance or the proposed trade-off between adversarial training and model generalization.