论文标题
PRIVFL:移动网络上的高维数据上的实用隐私权回归
PrivFL: Practical Privacy-preserving Federated Regressions on High-dimensional Data over Mobile Networks
论文作者
论文摘要
联合学习(FL)使大量用户能够共同学习共享的机器学习(ML)模型,该模型由集中式服务器协调,数据分布在多个设备上。这种方法使服务器或用户可以使用梯度下降训练和学习ML模型,同时将所有培训数据保留在用户设备上。我们考虑通过用户辍学是常见现象的移动网络训练ML模型。尽管联邦学习旨在降低数据隐私风险,但ML模型隐私并未受到太多关注。 在这项工作中,我们提出了Privfl,这是一种隐私保护系统,用于培训(预测)线性和逻辑回归模型以及在联合环境中的遗漏预测,同时保证数据和模型隐私,并确保对网络中辍学的用户的鲁棒性。我们设计了两个隐私权协议,用于训练基于加性同型加密(HE)方案和聚合协议的线性和逻辑回归模型。利用联合学习的培训算法,我们的培训协议的核心是对Alive用户数据的安全多方全球梯度计算。我们分析了针对半月对手的培训协议的安全性。只要汇总协议在聚合隐私游戏中安全,并且他计划在语义上是安全的,Privfl可以保证用户针对服务器的数据隐私,以及服务器对用户的回归模型隐私。我们演示了PRIVFL在现实世界数据集上的性能,并在联合学习系统中显示了其适用性。
Federated Learning (FL) enables a large number of users to jointly learn a shared machine learning (ML) model, coordinated by a centralized server, where the data is distributed across multiple devices. This approach enables the server or users to train and learn an ML model using gradient descent, while keeping all the training data on users' devices. We consider training an ML model over a mobile network where user dropout is a common phenomenon. Although federated learning was aimed at reducing data privacy risks, the ML model privacy has not received much attention. In this work, we present PrivFL, a privacy-preserving system for training (predictive) linear and logistic regression models and oblivious predictions in the federated setting, while guaranteeing data and model privacy as well as ensuring robustness to users dropping out in the network. We design two privacy-preserving protocols for training linear and logistic regression models based on an additive homomorphic encryption (HE) scheme and an aggregation protocol. Exploiting the training algorithm of federated learning, at the core of our training protocols is a secure multiparty global gradient computation on alive users' data. We analyze the security of our training protocols against semi-honest adversaries. As long as the aggregation protocol is secure under the aggregation privacy game and the additive HE scheme is semantically secure, PrivFL guarantees the users' data privacy against the server, and the server's regression model privacy against the users. We demonstrate the performance of PrivFL on real-world datasets and show its applicability in the federated learning system.