论文标题
非全粒图的脱钩的自我监督学习
Decoupled Self-supervised Learning for Non-Homophilous Graphs
论文作者
论文摘要
本文研究了在图表上进行节点表示学习的自我监督学习的问题。大多数现有的自我监督学习方法都认为该图是同质的,其中链接的节点通常属于同一类或具有相似的特征。但是,这种同质性的假设并不总是在现实图表中存在。我们通过为图神经网络开发一个解耦的自我监督学习(DSSL)框架来解决这个问题。 DSSL模仿了节点的生成过程和语义结构的潜在变量建模的链接,该过程将不同邻域之间的不同基础语义解散为自我监督的学习过程。我们的DSSL框架对编码者不可知,不需要预制的增强,因此对不同的图表灵活。为了有效地优化框架,我们得出了自我监督目标的下限的证据,并开发了具有各种推断的可扩展训练算法。我们提供了理论分析,以证明DSSL享有更好的下游性能。对各种图形基准测试的广泛实验表明,与竞争基线相比,我们提出的框架可以实现更好的性能。
This paper studies the problem of conducting self-supervised learning for node representation learning on graphs. Most existing self-supervised learning methods assume the graph is homophilous, where linked nodes often belong to the same class or have similar features. However, such assumptions of homophily do not always hold in real-world graphs. We address this problem by developing a decoupled self-supervised learning (DSSL) framework for graph neural networks. DSSL imitates a generative process of nodes and links from latent variable modeling of the semantic structure, which decouples different underlying semantics between different neighborhoods into the self-supervised learning process. Our DSSL framework is agnostic to the encoders and does not need prefabricated augmentations, thus is flexible to different graphs. To effectively optimize the framework, we derive the evidence lower bound of the self-supervised objective and develop a scalable training algorithm with variational inference. We provide a theoretical analysis to justify that DSSL enjoys the better downstream performance. Extensive experiments on various types of graph benchmarks demonstrate that our proposed framework can achieve better performance compared with competitive baselines.