论文标题

图形神经网络的图形摘要

Graph Summarization with Graph Neural Networks

论文作者

Blasi, Maximilian, Freudenreich, Manuel, Horvath, Johannes, Richerby, David, Scherp, Ansgar

论文摘要

图形汇总的目的是以结构化和紧凑的方式表示大图。基于等价类的图摘要保留了图形顶点的预定义特征,例如$ k $ - 霍普社区(例如顶点标签和边缘标签)。基于这些邻域特征,将顶点分配给等效类。分配的等价类的计算必须是对预定义特征的排列不变操作。这是通过对特征值e进行排序来实现的。 g。,边缘标签在计算上很昂贵,随后将结果放大。图形神经网络(GNN)满足置换不变性要求。我们将图形摘要的问题提出为$ k $ -HOP社区的根顶点的子图分类任务。我们基于流行的消息协议和替代方法来调整不同的GNN体系结构,以执行结构图摘要任务。我们将不同的GNN与标准的多层感知器(MLP)和Bloom滤波器作为非神经方法进行比较。对于我们的实验,我们考虑了大型Web图上的四个流行的图形摘要模型。这类似于具有挑战性的多级顶点分类任务,其类数量从$ 576 $到数十万。我们的结果表明,GNN的性能彼此接近。在四个实验中,三个实验中的三个实验中,非传播的GraphMLP模型的表现优于其他GNN。标准MLP的性能非常好,尤其是在许多课程的情况下。最后,Bloom过滤器的性能优于所有神经体系结构,但数据集的$ 576 $类别。

The goal of graph summarization is to represent large graphs in a structured and compact way. A graph summary based on equivalence classes preserves pre-defined features of a graph's vertex within a $k$-hop neighborhood such as the vertex labels and edge labels. Based on these neighborhood characteristics, the vertex is assigned to an equivalence class. The calculation of the assigned equivalence class must be a permutation invariant operation on the pre-defined features. This is achieved by sorting on the feature values, e. g., the edge labels, which is computationally expensive, and subsequently hashing the result. Graph Neural Networks (GNN) fulfill the permutation invariance requirement. We formulate the problem of graph summarization as a subgraph classification task on the root vertex of the $k$-hop neighborhood. We adapt different GNN architectures, both based on the popular message-passing protocol and alternative approaches, to perform the structural graph summarization task. We compare different GNNs with a standard multi-layer perceptron (MLP) and Bloom filter as non-neural method. For our experiments, we consider four popular graph summary models on a large web graph. This resembles challenging multi-class vertex classification tasks with the numbers of classes ranging from $576$ to multiple hundreds of thousands. Our results show that the performance of GNNs are close to each other. In three out of four experiments, the non-message-passing GraphMLP model outperforms the other GNNs. The performance of the standard MLP is extraordinary good, especially in the presence of many classes. Finally, the Bloom filter outperforms all neural architectures by a large margin, except for the dataset with the fewest number of $576$ classes.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源