论文标题
BEAS:区块链启用异步和安全联合机器学习
BEAS: Blockchain Enabled Asynchronous & Secure Federated Machine Learning
论文作者
论文摘要
联合学习(FL)使多方能够在不透露其私人数据集的情况下分配ML模型。但是,它假定对存储和汇总模型更新的集中聚合器的信任。这使得它容易受到恶意聚合器的梯度篡改和隐私泄漏。恶意政党还可以通过中毒训练数据或模型梯度来将后门引入联合模型。为了解决这些问题,我们提出了BEAS,这是第一个基于区块链的N-Party FL框架,该框架使用梯度修剪提供了严格的培训数据的隐私保证(与现有的基于噪声和基于剪辑的技术相比,差异隐私的改善)。异常检测方案用于最大程度地降低数据振作攻击的风险,以及进一步用于限制模型毒作攻击的疗效的梯度修剪。我们还定义了一种新的方案,以防止在异质学习环境中过早收敛。我们对多个数据集进行了广泛的实验,并有令人鼓舞的结果:BEAS成功防止了数据集重建攻击中的隐私泄漏,并最大程度地降低了中毒攻击的功效。此外,它实现了类似于集中式框架的准确性,并且其通信和计算与参与者数量线性线性规模相似。
Federated Learning (FL) enables multiple parties to distributively train a ML model without revealing their private datasets. However, it assumes trust in the centralized aggregator which stores and aggregates model updates. This makes it prone to gradient tampering and privacy leakage by a malicious aggregator. Malicious parties can also introduce backdoors into the joint model by poisoning the training data or model gradients. To address these issues, we present BEAS, the first blockchain-based framework for N-party FL that provides strict privacy guarantees of training data using gradient pruning (showing improved differential privacy compared to existing noise and clipping based techniques). Anomaly detection protocols are used to minimize the risk of data-poisoning attacks, along with gradient pruning that is further used to limit the efficacy of model-poisoning attacks. We also define a novel protocol to prevent premature convergence in heterogeneous learning environments. We perform extensive experiments on multiple datasets with promising results: BEAS successfully prevents privacy leakage from dataset reconstruction attacks, and minimizes the efficacy of poisoning attacks. Moreover, it achieves an accuracy similar to centralized frameworks, and its communication and computation overheads scale linearly with the number of participants.