论文标题
ABG:一个多方混合协议框架,用于保护隐私合作学习
ABG: A Multi-Party Mixed Protocol Framework for Privacy-Preserving Cooperative Learning
论文作者
论文摘要
合作学习使两个或更多数据所有者能够共同培训模型,已被广泛采用,以解决机器学习中培训数据不足的问题。如今,机构和组织迫切需要合作培训模型,同时私下保留彼此的数据。为了解决协作学习中隐私保护的问题,安全的外包计算和联合学习是两种典型方法。然而,当这两种方法被合作学习时,这两种方法有很多缺点。对于安全的外包计算,需要引入半honest服务器。一旦外包服务器串通或执行其他主动攻击,将披露数据的隐私。对于联合学习,很难应用于在多方分布垂直分区数据的情况下。在这项工作中,我们提出了一个多方混合协议框架ABG $^n $,该框架有效地实现了算术共享(a),布尔共享(b)和乱码共享(g)之间的任意转换($ n $ n $ n $ - 各方的场景。基于ABG $^n $,我们设计了一个保存隐私的多方合作学习系统,该系统允许不同的数据所有者在数据安全性和保护隐私方面的机器学习中合作。此外,我们为某些典型的机器学习方法(例如逻辑回归和神经网络)设计了特定的隐私计算协议。与以前的工作相比,所提出的方法具有更广泛的应用程序范围,并且不需要依靠其他服务器。最后,我们评估了本地设置和公共云设置上ABG $^n $的性能。实验表明,ABG $^n $的性能出色,尤其是在延迟较低的网络环境中。
Cooperative learning, that enables two or more data owners to jointly train a model, has been widely adopted to solve the problem of insufficient training data in machine learning. Nowadays, there is an urgent need for institutions and organizations to train a model cooperatively while keeping each other's data privately. To address the issue of privacy-preserving in collaborative learning, secure outsourced computation and federated learning are two typical methods. Nevertheless, there are many drawbacks for these two methods when they are leveraged in cooperative learning. For secure outsourced computation, semi-honest servers need to be introduced. Once the outsourced servers collude or perform other active attacks, the privacy of data will be disclosed. For federated learning, it is difficult to apply to the scenarios where vertically partitioned data are distributed over multiple parties. In this work, we propose a multi-party mixed protocol framework, ABG$^n$, which effectively implements arbitrary conversion between Arithmetic sharing (A), Boolean sharing (B) and Garbled-Circuits sharing (G) for $n$-party scenarios. Based on ABG$^n$, we design a privacy-preserving multi-party cooperative learning system, which allows different data owners to cooperate in machine learning in terms of data security and privacy-preserving. Additionally, we design specific privacy-preserving computation protocols for some typical machine learning methods such as logistic regression and neural networks. Compared with previous work, the proposed method has a wider scope of application and does not need to rely on additional servers. Finally, we evaluate the performance of ABG$^n$ on the local setting and on the public cloud setting. The experiments indicate that ABG$^n$ has excellent performance, especially in the network environment with low latency.