论文标题

ADAUC:针对长尾问题的端到端对抗性AUC优化

AdAUC: End-to-end Adversarial AUC Optimization Against Long-tail Problems

论文作者

Hou, Wenzheng, Xu, Qianqian, Yang, Zhiyong, Bao, Shilong, He, Yuan, Huang, Qingming

论文摘要

众所周知,深度学习模型容易受到对抗性例子的影响。现有的对抗训练的研究已在这一挑战中取得了长足的进步。作为一个典型的特征,他们经常认为班级分布总体平衡。但是,在广泛的应用中,长尾数据集无处不在,其中头等级实例的数量大于尾巴类。在这种情况下,AUC比准确度更合理,因为它对班级分布不敏感。在此激励的情况下,我们提出了一项早期试验,以探索对抗训练方法以优化AUC。主要挑战在于,积极和负面的例子在目标函数中紧密耦合。作为一个直接结果,如果没有数据集进行全面扫描,就无法生成对抗性示例。为了解决这个问题,基于凹入的正规化方案,我们将AUC优化问题重新制定为鞍点问题,该问题将成为实例函数。这导致端到端培训方案。此外,我们提供了提出的算法的收敛保证。我们的分析与现有研究不同,因为该算法被要求通过计算Min-Max问题的梯度来产生对抗性示例。最后,广泛的实验结果表明,在三个长尾数据集中,我们的算法的性能和鲁棒性。

It is well-known that deep learning models are vulnerable to adversarial examples. Existing studies of adversarial training have made great progress against this challenge. As a typical trait, they often assume that the class distribution is overall balanced. However, long-tail datasets are ubiquitous in a wide spectrum of applications, where the amount of head class instances is larger than the tail classes. Under such a scenario, AUC is a much more reasonable metric than accuracy since it is insensitive toward class distribution. Motivated by this, we present an early trial to explore adversarial training methods to optimize AUC. The main challenge lies in that the positive and negative examples are tightly coupled in the objective function. As a direct result, one cannot generate adversarial examples without a full scan of the dataset. To address this issue, based on a concavity regularization scheme, we reformulate the AUC optimization problem as a saddle point problem, where the objective becomes an instance-wise function. This leads to an end-to-end training protocol. Furthermore, we provide a convergence guarantee of the proposed algorithm. Our analysis differs from the existing studies since the algorithm is asked to generate adversarial examples by calculating the gradient of a min-max problem. Finally, the extensive experimental results show the performance and robustness of our algorithm in three long-tail datasets.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源