论文标题

FLEDEFENDER:在联邦学习中打击目标攻击

FL-Defender: Combating Targeted Attacks in Federated Learning

论文作者

Jebreel, Najeeb, Domingo-Ferrer, Josep

论文摘要

联合学习(FL)使从分发在一组参与工人之间的本地数据中学习全球机器学习模型。这使得i)由于从丰富的联合培训数据中学习而培训更准确的模型,ii)通过不与他人共享工人的本地私人数据来改善隐私。但是,FL的分布性质使其容易受到针对性的中毒攻击的影响,这些攻击会对学习模型的完整性产生负面影响,而不幸的是,很难检测到。现有针对这些攻击的防御措施受到工人数据分布的假设的限制,可能会在主要任务上降低全球模型性能和/或不适合高维模型。在本文中,我们分析了针对FL的靶向攻击,发现与攻击有关的深度学习(DL)模型的最后一层神经元与不相关的神经元具有不同的行为,从而使最后一层梯度有价值攻击检测的有价值特征。因此,我们将\ textIt {fl-defender}作为对抗目标攻击的方法。它由i)组成,通过计算工人的最后一层梯度的工人角度相似性,ii)使用PCA压缩所得的相似性向量来减少冗余信息,并重新对工人的更新基于压缩相似性vectors centroid controdiation contropa,i)由I)组成。在三个具有不同DL模型大小和数据分布的数据集的实验显示了我们方法在防御标签和后门攻击方面的有效性。与几个最先进的防御能力相比,FL Defender取得了最低的攻击成功率,维持全球模型在主要任务上的性能,并导致服务器上的最小计算开销。

Federated learning (FL) enables learning a global machine learning model from local data distributed among a set of participating workers. This makes it possible i) to train more accurate models due to learning from rich joint training data, and ii) to improve privacy by not sharing the workers' local private data with others. However, the distributed nature of FL makes it vulnerable to targeted poisoning attacks that negatively impact the integrity of the learned model while, unfortunately, being difficult to detect. Existing defenses against those attacks are limited by assumptions on the workers' data distribution, may degrade the global model performance on the main task and/or are ill-suited to high-dimensional models. In this paper, we analyze targeted attacks against FL and find that the neurons in the last layer of a deep learning (DL) model that are related to the attacks exhibit a different behavior from the unrelated neurons, making the last-layer gradients valuable features for attack detection. Accordingly, we propose \textit{FL-Defender} as a method to combat FL targeted attacks. It consists of i) engineering more robust discriminative features by calculating the worker-wise angle similarity for the workers' last-layer gradients, ii) compressing the resulting similarity vectors using PCA to reduce redundant information, and iii) re-weighting the workers' updates based on their deviation from the centroid of the compressed similarity vectors. Experiments on three data sets with different DL model sizes and data distributions show the effectiveness of our method at defending against label-flipping and backdoor attacks. Compared to several state-of-the-art defenses, FL-Defender achieves the lowest attack success rates, maintains the performance of the global model on the main task and causes minimal computational overhead on the server.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源