论文标题

了解深神经网络的决策边界:一项实证研究

Understanding the Decision Boundary of Deep Neural Networks: An Empirical Study

论文作者

Mickisch, David, Assion, Felix, Greßner, Florens, Günther, Wiebke, Motta, Mariele

论文摘要

尽管在许多图像分类任务上取得了出色的性能,但最先进的机器学习(ML)分类器仍然容易受到小型输入扰动的影响。尤其是,对抗性例子的存在引起了人们对在安全和安全至关重要环境中部署ML模型的担忧,例如自动驾驶和疾病检测。在过去的几年中,已经发布了许多防御方法,目的是改善对抗性和腐败的鲁棒性。但是,提议的措施仅在非常有限的程度上成功。这种有限的进展部分是由于对深度神经网络的决策边界和决策区域缺乏了解。因此,我们研究了数据的最小距离指向决策界限,以及该边缘如何在深度神经网络的训练中演变。通过对MNIST,Fashion-Mnist和CIFAR-10进行实验,我们观察到决策边界在训练上更接近自然图像。在训练的晚期时期,这种现象甚至仍然完好无损,在训练的晚期,分类器已经获得较低的训练和测试错误率。另一方面,对抗训练似乎有可能防止这种不希望的决策边界融合。

Despite achieving remarkable performance on many image classification tasks, state-of-the-art machine learning (ML) classifiers remain vulnerable to small input perturbations. Especially, the existence of adversarial examples raises concerns about the deployment of ML models in safety- and security-critical environments, like autonomous driving and disease detection. Over the last few years, numerous defense methods have been published with the goal of improving adversarial as well as corruption robustness. However, the proposed measures succeeded only to a very limited extent. This limited progress is partly due to the lack of understanding of the decision boundary and decision regions of deep neural networks. Therefore, we study the minimum distance of data points to the decision boundary and how this margin evolves over the training of a deep neural network. By conducting experiments on MNIST, FASHION-MNIST, and CIFAR-10, we observe that the decision boundary moves closer to natural images over training. This phenomenon even remains intact in the late epochs of training, where the classifier already obtains low training and test error rates. On the other hand, adversarial training appears to have the potential to prevent this undesired convergence of the decision boundary.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源