论文标题
您可以通过完全不训练权重来获得更好的图形神经网络:寻找未训练的GNN票
You Can Have Better Graph Neural Networks by Not Training Weights at All: Finding Untrained GNNs Tickets
论文作者
论文摘要
最近的作品令人印象深刻地证明,在随机初始化的卷积神经网络(CNN)中存在一个子网,可以在初始化时与完全训练的密集网络的性能相匹配,而无需对网络的权重(即未经训练的网络)进行任何优化。但是,图形神经网络(GNN)中这种未经训练的子网的存在仍然是神秘的。在本文中,我们对发现匹配未训练的GNN的首先探索。以稀疏为核心工具,我们可以在初始化时找到\ textit {未经训练的稀疏子网络},可以匹配\ textit {完全训练的密度} gnns的性能。除了这已经令人鼓舞地发现可比的性能外,我们还表明,所发现的未经训练的子网可能会大大减轻GNN过度光滑的问题,从而成为使得无铃铛和哨声更深入的GNNS的强大工具。我们还观察到,这种稀疏的未经训练的子网在分布外检测和输入扰动的鲁棒性方面具有吸引人的性能。我们在包括开放图基准(OGB)在内的各种流行数据集上广泛使用的GNN体系结构进行了评估。
Recent works have impressively demonstrated that there exists a subnetwork in randomly initialized convolutional neural networks (CNNs) that can match the performance of the fully trained dense networks at initialization, without any optimization of the weights of the network (i.e., untrained networks). However, the presence of such untrained subnetworks in graph neural networks (GNNs) still remains mysterious. In this paper we carry out the first-of-its-kind exploration of discovering matching untrained GNNs. With sparsity as the core tool, we can find \textit{untrained sparse subnetworks} at the initialization, that can match the performance of \textit{fully trained dense} GNNs. Besides this already encouraging finding of comparable performance, we show that the found untrained subnetworks can substantially mitigate the GNN over-smoothing problem, hence becoming a powerful tool to enable deeper GNNs without bells and whistles. We also observe that such sparse untrained subnetworks have appealing performance in out-of-distribution detection and robustness of input perturbations. We evaluate our method across widely-used GNN architectures on various popular datasets including the Open Graph Benchmark (OGB).