论文标题
快速学习具有保证可推广性的图形神经网络:一个隐藏的案例
Fast Learning of Graph Neural Networks with Guaranteed Generalizability: One-hidden-layer Case
论文作者
论文摘要
尽管图神经网络(GNN)最近在实践中从图形结构数据中学习取得了长足的进步,但在文献中,它们对普遍性的理论保证仍然难以捉摸。在本文中,我们为GNN提供了一个具有一个隐藏层的GNN的概括分析,以解决回归和二进制分类问题。在假设存在基础真相GNN模型(零概括误差)的假设下,GNN学习的目的是从训练数据中估算基础真相GNN参数。为了实现这一目标,我们提出了一种基于张量初始化和加速梯度下降的学习算法。然后,我们证明,提出的学习算法将回归问题收敛到地面真相GNN模型,并为二进制分类问题足够接近地面真相的模型收敛。此外,在这两种情况下,所提出的学习算法的收敛速率被证明是线性和更快的,比香草梯度下降算法更快。我们进一步探讨了GNN的样品复杂性及其基础图属性之间的关系。最后,我们提供数值实验,以证明我们的分析的有效性以及拟议的学习算法的有效性。
Although graph neural networks (GNNs) have made great progress recently on learning from graph-structured data in practice, their theoretical guarantee on generalizability remains elusive in the literature. In this paper, we provide a theoretically-grounded generalizability analysis of GNNs with one hidden layer for both regression and binary classification problems. Under the assumption that there exists a ground-truth GNN model (with zero generalization error), the objective of GNN learning is to estimate the ground-truth GNN parameters from the training data. To achieve this objective, we propose a learning algorithm that is built on tensor initialization and accelerated gradient descent. We then show that the proposed learning algorithm converges to the ground-truth GNN model for the regression problem, and to a model sufficiently close to the ground-truth for the binary classification problem. Moreover, for both cases, the convergence rate of the proposed learning algorithm is proven to be linear and faster than the vanilla gradient descent algorithm. We further explore the relationship between the sample complexity of GNNs and their underlying graph properties. Lastly, we provide numerical experiments to demonstrate the validity of our analysis and the effectiveness of the proposed learning algorithm for GNNs.