论文标题
用于机器学习的剩余矩阵产品状态
Residual Matrix Product State for Machine Learning
论文作者
论文摘要
张量网络源自量子物理,它成为经典和量子机学习的有效工具。然而,张量网络与古典机器学习的复杂神经网络模型之间仍然存在很大的准确差距。在这项工作中,我们结合了矩阵产品状态(MPS),最简单的张量网络结构和残留神经网络的思想,并提出了残留的矩阵产品状态(RESMP)。可以将其视为其图层的网络将“隐藏”特征映射到输出(例如分类),而层的变异参数是样本特征的函数(例如,图像的像素)。这与神经网络不同,在该网络中,图层将特征向输出的特征映射到输出。在效率,稳定性和表达能力方面,地铁可以配备非线性激活和辍学层,并优于最先进的张量网络模型。此外,从多项式扩展的角度来看,在分解和指数机器自然会出现的多项式扩展的角度可以解释。我们的工作有助于连接和杂交神经和张量网络,这对于进一步增强我们对工作机制的理解并提高两种模型的性能至关重要。
Tensor network, which originates from quantum physics, is emerging as an efficient tool for classical and quantum machine learning. Nevertheless, there still exists a considerable accuracy gap between tensor network and the sophisticated neural network models for classical machine learning. In this work, we combine the ideas of matrix product state (MPS), the simplest tensor network structure, and residual neural network and propose the residual matrix product state (ResMPS). The ResMPS can be treated as a network where its layers map the "hidden" features to the outputs (e.g., classifications), and the variational parameters of the layers are the functions of the features of the samples (e.g., pixels of images). This is different from neural network, where the layers map feed-forwardly the features to the output. The ResMPS can equip with the non-linear activations and dropout layers, and outperforms the state-of-the-art tensor network models in terms of efficiency, stability, and expression power. Besides, ResMPS is interpretable from the perspective of polynomial expansion, where the factorization and exponential machines naturally emerge. Our work contributes to connecting and hybridizing neural and tensor networks, which is crucial to further enhance our understand of the working mechanisms and improve the performance of both models.