论文标题
稀疏图神经网络
Towards Sparsification of Graph Neural Networks
论文作者
论文摘要
随着实际图的大小扩大,部署了数十亿个参数的较大GNN模型。在此类模型中,高参数计数使图表的训练和推断昂贵且具有挑战性。为了降低GNN的计算和记忆成本,通常采用了输入图中的冗余节点和边缘等优化方法。但是,直接针对模型层稀疏的模型压缩,主要限于用于图像分类和对象检测等任务的传统深层神经网络(DNN)。在本文中,我们利用两种最先进的模型压缩方法(1)训练和修剪,以及(2)稀疏的GNN重量层的训练。我们评估和比较了两种方法的效率,从精确性,训练稀疏性和现实图表上的训练拖失lop方面。我们的实验结果表明,在IA-Email,Wiki-Talk和Stackoverflow数据集上,用于链接预测,稀疏训练和较低的训练拖失板可通过火车和修剪方法实现可比的准确性。在用于节点分类的大脑数据集上,稀疏训练使用较低的数字拖鞋(小于1/7的火车和修剪方法),并在极端模型的稀疏性下保留了更好的精度性能。
As real-world graphs expand in size, larger GNN models with billions of parameters are deployed. High parameter count in such models makes training and inference on graphs expensive and challenging. To reduce the computational and memory costs of GNNs, optimization methods such as pruning the redundant nodes and edges in input graphs have been commonly adopted. However, model compression, which directly targets the sparsification of model layers, has been mostly limited to traditional Deep Neural Networks (DNNs) used for tasks such as image classification and object detection. In this paper, we utilize two state-of-the-art model compression methods (1) train and prune and (2) sparse training for the sparsification of weight layers in GNNs. We evaluate and compare the efficiency of both methods in terms of accuracy, training sparsity, and training FLOPs on real-world graphs. Our experimental results show that on the ia-email, wiki-talk, and stackoverflow datasets for link prediction, sparse training with much lower training FLOPs achieves a comparable accuracy with the train and prune method. On the brain dataset for node classification, sparse training uses a lower number FLOPs (less than 1/7 FLOPs of train and prune method) and preserves a much better accuracy performance under extreme model sparsity.