论文标题
图形神经网络的自我监督的图形结构完善
Self-Supervised Graph Structure Refinement for Graph Neural Networks
论文作者
论文摘要
旨在学习图形神经网络(GNN)的邻接矩阵(GNNS)的图形结构学习(GSL)在提高GNN的性能方面具有巨大潜力。大多数现有的GSL作品都采用联合学习框架,其中估计的邻接矩阵和GNN参数已针对下游任务进行了优化。但是,由于GSL本质上是一项链接预测任务,其目标可能与下游任务的目标有很大不同。这两个目标的不一致限制了GSL方法学习潜在的最佳图形结构。此外,在邻接矩阵的估计和优化过程中,联合学习框架在时间和空间方面遇到了可扩展性问题。为了减轻这些问题,我们提出了图形结构改进(GSR)框架,并使用预处理前管道。具体而言,训练前阶段旨在通过具有内部和跨视图链接链接预测任务的多视图对比学习框架来全面地估算基础图结构。然后,通过根据预训练模型估计的边缘概率添加和删除边缘来完善图形结构。最后,通过预训练的模型初始化了微调GNN,并针对下游任务进行了优化。通过在微调空间中保持精致的图形结构,GSR避免在微调阶段估算和优化图形结构,从而具有极大的可扩展性和效率。此外,迁移知识和精炼图都可以提高微调GNN。进行了广泛的实验,以评估拟议模型的有效性(在六个基准数据集上的最佳性能),效率和可伸缩性(使用32.8%的GPU记忆使用32.8%的GPU存储器)。
Graph structure learning (GSL), which aims to learn the adjacency matrix for graph neural networks (GNNs), has shown great potential in boosting the performance of GNNs. Most existing GSL works apply a joint learning framework where the estimated adjacency matrix and GNN parameters are optimized for downstream tasks. However, as GSL is essentially a link prediction task, whose goal may largely differ from the goal of the downstream task. The inconsistency of these two goals limits the GSL methods to learn the potential optimal graph structure. Moreover, the joint learning framework suffers from scalability issues in terms of time and space during the process of estimation and optimization of the adjacency matrix. To mitigate these issues, we propose a graph structure refinement (GSR) framework with a pretrain-finetune pipeline. Specifically, The pre-training phase aims to comprehensively estimate the underlying graph structure by a multi-view contrastive learning framework with both intra- and inter-view link prediction tasks. Then, the graph structure is refined by adding and removing edges according to the edge probabilities estimated by the pre-trained model. Finally, the fine-tuning GNN is initialized by the pre-trained model and optimized toward downstream tasks. With the refined graph structure remaining static in the fine-tuning space, GSR avoids estimating and optimizing graph structure in the fine-tuning phase which enjoys great scalability and efficiency. Moreover, the fine-tuning GNN is boosted by both migrating knowledge and refining graphs. Extensive experiments are conducted to evaluate the effectiveness (best performance on six benchmark datasets), efficiency, and scalability (13.8x faster using 32.8% GPU memory compared to the best GSL baseline on Cora) of the proposed model.