论文标题
具有容量探索的二进制图卷积网络
Binary Graph Convolutional Network with Capacity Exploration
论文作者
论文摘要
图形神经网络(GNN)的当前成功通常依赖于加载整个属性图进行处理,这可能对有限的内存资源不满意,尤其是当属性图很大时。本文先驱者提出了一个二进制图卷积网络(BI-GCN),该网络将网络参数和输入节点属性二进制,并利用二进制操作,而不是浮点数矩阵乘法,以进行网络压缩和加速。同时,我们还提出了一种基于梯度近似的新近似值方法,以正确训练我们的BI-GCN。根据理论分析,我们的BI-GCN可以平均在三个引用网络上,即Cora,PubMed和Citeseer,在网络参数和输入数据中平均将记忆消耗降低约31倍,并平均降低了〜51X的推理速度。此外,我们引入了一种将我们的二进化方法推广到其他GNN变体的一般方法,并实现相似的效率。尽管拟议的BI-GCN和BI-GNN既简单又有效,但这些压缩网络也可能具有潜在的能力问题,即,它们可能没有足够的存储能力来学习特定任务的足够表示。为了解决这个容量问题,提出了一个熵覆盖假设,以预测BI-GNN隐藏层的宽度的下限。广泛的实验表明,我们的BI-GCN和BI-GNN可以在七个节点分类数据集上与相应的完整精确基准提供可比的性能,并验证了我们的熵涵盖能力问题的有效性。
The current success of Graph Neural Networks (GNNs) usually relies on loading the entire attributed graph for processing, which may not be satisfied with limited memory resources, especially when the attributed graph is large. This paper pioneers to propose a Binary Graph Convolutional Network (Bi-GCN), which binarizes both the network parameters and input node attributes and exploits binary operations instead of floating-point matrix multiplications for network compression and acceleration. Meanwhile, we also propose a new gradient approximation based back-propagation method to properly train our Bi-GCN. According to the theoretical analysis, our Bi-GCN can reduce the memory consumption by an average of ~31x for both the network parameters and input data, and accelerate the inference speed by an average of ~51x, on three citation networks, i.e., Cora, PubMed, and CiteSeer. Besides, we introduce a general approach to generalize our binarization method to other variants of GNNs, and achieve similar efficiencies. Although the proposed Bi-GCN and Bi-GNNs are simple yet efficient, these compressed networks may also possess a potential capacity problem, i.e., they may not have enough storage capacity to learn adequate representations for specific tasks. To tackle this capacity problem, an Entropy Cover Hypothesis is proposed to predict the lower bound of the width of Bi-GNN hidden layers. Extensive experiments have demonstrated that our Bi-GCN and Bi-GNNs can give comparable performances to the corresponding full-precision baselines on seven node classification datasets and verified the effectiveness of our Entropy Cover Hypothesis for solving the capacity problem.