论文标题
神经网络的表示理论
The Representation Theory of Neural Networks
论文作者
论文摘要
在这项工作中,我们表明神经网络可以通过颤抖表示的数学理论来表示。更具体地说,我们证明神经网络是具有激活函数的颤抖表示,是我们使用网络颤动表示的数学对象。此外,我们表明,网络颤抖会轻轻适应常见的神经网络概念,例如完全连接的层,卷积操作,残差连接,批处理归一化,汇总操作,甚至随机有线神经网络。我们表明,这种数学表示绝不是对神经网络的近似值,因为它与现实完全匹配。这种解释是代数,可以使用代数方法研究。我们还提供了一个颤抖的表示模型,以了解神经网络如何从数据中创建表示形式。我们表明,神经网络将数据保存为Quiver表示,并将其映射到称为Moduli空间的几何空间,该空间是根据网络的基本图表(即其颤动的)给出的。这是由于我们定义的对象以及理解神经网络如何以组合和代数方式计算预测的结果。总体而言,通过颤抖的代表理论代表神经网络会导致9个后果和4个询问未来的研究,我们认为这是一个极大的兴趣,可以更好地了解什么是神经网络以及它们的工作方式。
In this work, we show that neural networks can be represented via the mathematical theory of quiver representations. More specifically, we prove that a neural network is a quiver representation with activation functions, a mathematical object that we represent using a network quiver. Also, we show that network quivers gently adapt to common neural network concepts such as fully-connected layers, convolution operations, residual connections, batch normalization, pooling operations and even randomly wired neural networks. We show that this mathematical representation is by no means an approximation of what neural networks are as it exactly matches reality. This interpretation is algebraic and can be studied with algebraic methods. We also provide a quiver representation model to understand how a neural network creates representations from the data. We show that a neural network saves the data as quiver representations, and maps it to a geometrical space called the moduli space, which is given in terms of the underlying oriented graph of the network, i.e., its quiver. This results as a consequence of our defined objects and of understanding how the neural network computes a prediction in a combinatorial and algebraic way. Overall, representing neural networks through the quiver representation theory leads to 9 consequences and 4 inquiries for future research that we believe are of great interest to better understand what neural networks are and how they work.