论文标题
与GNN的差异私人图分类
Differentially Private Graph Classification with GNNs
论文作者
论文摘要
图形神经网络(GNN)已将自己确立为许多机器学习应用的最新模型,例如对社交网络,蛋白质相互作用和分子的分析。这些数据集中有几个包含对隐私敏感的数据。具有不同隐私的机器学习是一种有前途的技术,可以从敏感数据中获得洞察力,同时提供正式保证隐私保护。但是,由于图的固有结构连接所带来的挑战,迄今为止,GNN的私人培训迄今仍未探索。在这项工作中,我们为图形分类引入了差异隐私,这是机器学习在图形上的关键应用之一。我们的方法适用于多段数据集的深度学习,并且依赖于差异化的私有随机梯度下降(DP-SGD)。我们显示了各种合成和公共数据集的结果,并评估了不同GNN体系结构和培训超参数对模型性能的影响,以差异私有图形分类。最后,我们应用解释性技术来评估是否在私人和非私有环境中学习了类似的表示,并为该领域的未来工作建立了强大的基准。
Graph Neural Networks (GNNs) have established themselves as the state-of-the-art models for many machine learning applications such as the analysis of social networks, protein interactions and molecules. Several among these datasets contain privacy-sensitive data. Machine learning with differential privacy is a promising technique to allow deriving insight from sensitive data while offering formal guarantees of privacy protection. However, the differentially private training of GNNs has so far remained under-explored due to the challenges presented by the intrinsic structural connectivity of graphs. In this work, we introduce differential privacy for graph-level classification, one of the key applications of machine learning on graphs. Our method is applicable to deep learning on multi-graph datasets and relies on differentially private stochastic gradient descent (DP-SGD). We show results on a variety of synthetic and public datasets and evaluate the impact of different GNN architectures and training hyperparameters on model performance for differentially private graph classification. Finally, we apply explainability techniques to assess whether similar representations are learned in the private and non-private settings and establish robust baselines for future work in this area.