论文标题
GCC:图形神经网络预训练的图形对比度编码
GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training
论文作者
论文摘要
图表学习已成为解决现实世界问题的强大技术。各种下游图表学习任务从其最近的发展中受益,例如节点分类,相似性搜索和图形分类。但是,图表表示学习的先前艺术将重点放在域特定问题上,并为每个图数据集训练专用模型,这通常不可转换为外域数据。受到自然语言处理和计算机视觉预训练的最新进展的启发,我们设计了图形对比编码(GCC)(一个自我监督的图形神经网络预训练框架),以捕获多个网络跨多个网络的通用网络拓扑特性。我们将海湾合作委员会的预培训任务设计为在网络中和跨网络中的子图实例歧视,并利用对比度学习来增强图形神经网络的能力,以学习内在且可转移的结构表示。我们对三个图形学习任务和十个图数据集进行了广泛的实验。结果表明,在各种数据集的集合中进行的海湾合作委员会可以实现其特定于任务和训练有素的从事同行的竞争性或更好的性能。这表明预训练和微调范式为图表学习提供了巨大的潜力。
Graph representation learning has emerged as a powerful technique for addressing real-world problems. Various downstream graph learning tasks have benefited from its recent developments, such as node classification, similarity search, and graph classification. However, prior arts on graph representation learning focus on domain specific problems and train a dedicated model for each graph dataset, which is usually non-transferable to out-of-domain data. Inspired by the recent advances in pre-training from natural language processing and computer vision, we design Graph Contrastive Coding (GCC) -- a self-supervised graph neural network pre-training framework -- to capture the universal network topological properties across multiple networks. We design GCC's pre-training task as subgraph instance discrimination in and across networks and leverage contrastive learning to empower graph neural networks to learn the intrinsic and transferable structural representations. We conduct extensive experiments on three graph learning tasks and ten graph datasets. The results show that GCC pre-trained on a collection of diverse datasets can achieve competitive or better performance to its task-specific and trained-from-scratch counterparts. This suggests that the pre-training and fine-tuning paradigm presents great potential for graph representation learning.