论文标题
图形卷积网络,用于包含缺失功能的图形
Graph Convolutional Networks for Graphs Containing Missing Features
论文作者
论文摘要
图形卷积网络(GCN)在图形分析任务中取得了巨大成功。它通过平滑图表上的节点特征来起作用。当前的GCN模型压倒性地假设节点特征信息已完成。但是,实际图形数据通常不完整,并且包含缺失的功能。传统上,人们必须根据插补技术进行估计并填写未知功能,然后应用GCN。但是,特征填充和图形学习的过程分开,导致性能降低和不稳定的性能。当缺少大量功能时,此问题变得更加严重。我们提出了一种将GCN适应包含缺失功能的图的方法。与传统策略相反,我们的方法将缺少功能和图形学习的处理集成到同一神经网络体系结构中。我们的想法是用高斯混合模型(GMM)表示缺少的数据,并计算GCN的第一个隐藏层中神经元的预期激活,同时保持网络的其他层不变。这使我们能够以端到端的方式学习GMM参数和网络权重参数。值得注意的是,我们的方法不会增加GCN的计算复杂性,并且当功能完成时,它与GCN一致。我们通过广泛的实验证明,我们的方法在节点分类和链接预测任务中显着优于基于插补的方法。我们表明,对于具有完整功能的案例,我们的方法的表现较低。
Graph Convolutional Network (GCN) has experienced great success in graph analysis tasks. It works by smoothing the node features across the graph. The current GCN models overwhelmingly assume that the node feature information is complete. However, real-world graph data are often incomplete and containing missing features. Traditionally, people have to estimate and fill in the unknown features based on imputation techniques and then apply GCN. However, the process of feature filling and graph learning are separated, resulting in degraded and unstable performance. This problem becomes more serious when a large number of features are missing. We propose an approach that adapts GCN to graphs containing missing features. In contrast to traditional strategy, our approach integrates the processing of missing features and graph learning within the same neural network architecture. Our idea is to represent the missing data by Gaussian Mixture Model (GMM) and calculate the expected activation of neurons in the first hidden layer of GCN, while keeping the other layers of the network unchanged. This enables us to learn the GMM parameters and network weight parameters in an end-to-end manner. Notably, our approach does not increase the computational complexity of GCN and it is consistent with GCN when the features are complete. We demonstrate through extensive experiments that our approach significantly outperforms the imputation-based methods in node classification and link prediction tasks. We show that the performance of our approach for the case with a low level of missing features is even superior to GCN for the case with complete features.