论文标题
平滑的对抗训练
Smooth Adversarial Training
论文作者
论文摘要
人们普遍认为,网络不能既准确又健壮,因此获得鲁棒性意味着失去准确性。人们通常也认为,除非使网络更大,否则网络架构元素否则在改善对抗性的鲁棒性方面将无关紧要。在这里,我们提供证据,通过仔细研究对抗训练来挑战这些共同的信念。我们的关键观察结果是,由于其非平滑性质,广泛使用的Relu激活功能会显着削弱对抗性训练。因此,我们提出了平滑的对抗训练(SAT),在其中,我们将其取代Relu的平滑近似值以增强对抗性训练。 SAT中平滑激活功能的目的是使其能够找到更难的对抗性示例,并在对抗训练期间计算更好的梯度更新。 与标准的对抗训练相比,SAT可以提高“自由”的对抗性鲁棒性,即,准确性没有下降,计算成本没有增加。例如,在不引入其他计算的情况下,SAT显着将Resnet-50的鲁棒性从33.0%提高到42.3%,同时,ImageNet的精度也提高了0.9%。 SAT还可以很好地与较大的网络合作:它可以帮助有效的L1实现82.2%的精度和58.6%的鲁棒性,以使先前的最新防御能力优于9.5%的准确性,而鲁棒性的精度为11.6%。型号可在https://github.com/cihangxie/smoothardversarialtraining上找到。
It is commonly believed that networks cannot be both accurate and robust, that gaining robustness means losing accuracy. It is also generally believed that, unless making networks larger, network architectural elements would otherwise matter little in improving adversarial robustness. Here we present evidence to challenge these common beliefs by a careful study about adversarial training. Our key observation is that the widely-used ReLU activation function significantly weakens adversarial training due to its non-smooth nature. Hence we propose smooth adversarial training (SAT), in which we replace ReLU with its smooth approximations to strengthen adversarial training. The purpose of smooth activation functions in SAT is to allow it to find harder adversarial examples and compute better gradient updates during adversarial training. Compared to standard adversarial training, SAT improves adversarial robustness for "free", i.e., no drop in accuracy and no increase in computational cost. For example, without introducing additional computations, SAT significantly enhances ResNet-50's robustness from 33.0% to 42.3%, while also improving accuracy by 0.9% on ImageNet. SAT also works well with larger networks: it helps EfficientNet-L1 to achieve 82.2% accuracy and 58.6% robustness on ImageNet, outperforming the previous state-of-the-art defense by 9.5% for accuracy and 11.6% for robustness. Models are available at https://github.com/cihangxie/SmoothAdversarialTraining.