论文标题

对回归学习和相应防御的数据中毒攻击

Data Poisoning Attacks on Regression Learning and Corresponding Defenses

论文作者

Müller, Nicolas Michael, Kowatsch, Daniel, Böttinger, Konstantin

论文摘要

对抗数据中毒是对机器学习的有效攻击,并通过将中毒数据引入培训数据集中威胁模型完整性。到目前为止,即使在许多关键任务系统(例如药物剂量,网络物理系统的控制和管理电源)中使用回归学习,它主要是用于分类的。因此,在本研究中,我们旨在评估数据中毒对回归学习的各个方面,从而超过了先前的广度和深度工作。我们提出了现实的场景,其中数据中毒攻击威胁生产系统并引入新型的黑盒攻击,然后将其应用于现实词的医疗用例。结果,我们观察到,由于仅插入了2%的毒药样本,回归器的平均平方误差(MSE)增加到150%。最后,我们提出了针对新颖和以前的攻击的新防御策略,并在26个数据集上对其进行了彻底评估。由于进行了实验,我们得出的结论是,提出的防御策略有效地减轻了所考虑的攻击。

Adversarial data poisoning is an effective attack against machine learning and threatens model integrity by introducing poisoned data into the training dataset. So far, it has been studied mostly for classification, even though regression learning is used in many mission critical systems (such as dosage of medication, control of cyber-physical systems and managing power supply). Therefore, in the present research, we aim to evaluate all aspects of data poisoning attacks on regression learning, exceeding previous work both in terms of breadth and depth. We present realistic scenarios in which data poisoning attacks threaten production systems and introduce a novel black-box attack, which is then applied to a real-word medical use-case. As a result, we observe that the mean squared error (MSE) of the regressor increases to 150 percent due to inserting only two percent of poison samples. Finally, we present a new defense strategy against the novel and previous attacks and evaluate it thoroughly on 26 datasets. As a result of the conducted experiments, we conclude that the proposed defence strategy effectively mitigates the considered attacks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源