论文标题
训练图神经网络的强盗采样器
Bandit Samplers for Training Graph Neural Networks
论文作者
论文摘要
已经提出了几种具有差异降低的采样算法,以加速图形卷积网络(GCN)的训练。但是,由于最佳采样分布的棘手计算,这些采样算法对GCN进行了次优,并且不适用于更通用的图形神经网络(GNN),其中消息聚集器包含学习的权重,而不是固定的权重,例如图形注意力网络(GAT)。根本原因是,邻居或学到的权重的嵌入在训练过程中正在发生变化,而不是先验的,但仅在采样时才被部分观察到,从而导致最佳方差降低采样器的推导。在本文中,我们将采样方差的优化作为对手匪徒问题,其中奖励与节点嵌入和学习的权重有关,并且可以不断变化。因此,一个好的采样器需要获取有关更多邻居(探索)的差异信息,同时优化即时抽样差异(Exploit)。从理论上讲,我们表明我们的算法渐近地接近3倍以下的最佳差异。我们在多个数据集中显示了方法的效率和有效性。
Several sampling algorithms with variance reduction have been proposed for accelerating the training of Graph Convolution Networks (GCNs). However, due to the intractable computation of optimal sampling distribution, these sampling algorithms are suboptimal for GCNs and are not applicable to more general graph neural networks (GNNs) where the message aggregator contains learned weights rather than fixed weights, such as Graph Attention Networks (GAT). The fundamental reason is that the embeddings of the neighbors or learned weights involved in the optimal sampling distribution are changing during the training and not known a priori, but only partially observed when sampled, thus making the derivation of an optimal variance reduced samplers non-trivial. In this paper, we formulate the optimization of the sampling variance as an adversary bandit problem, where the rewards are related to the node embeddings and learned weights, and can vary constantly. Thus a good sampler needs to acquire variance information about more neighbors (exploration) while at the same time optimizing the immediate sampling variance (exploit). We theoretically show that our algorithm asymptotically approaches the optimal variance within a factor of 3. We show the efficiency and effectiveness of our approach on multiple datasets.