论文标题
监督稀疏编码的对抗性鲁棒性
Adversarial Robustness of Supervised Sparse Coding
论文作者
论文摘要
最近的几个结果提供了对对抗例子现象的理论见解。但是,由于所研究模型的简单性与在实践中部署的模型的复杂性之间存在差距,现有结果通常受到限制。在这项工作中,我们通过考虑一个涉及学习表示形式的模型,同时提供精确的概括和鲁棒性证书的模型来取得更好的平衡。我们专注于通过结合启动稀疏性编码器与线性分类器结合的假设类别,并在特征空间中(有监督的)表示图的表达性和稳定性之间显示出有趣的相互作用。我们限制了由字典参数列出的假设的强大风险(to $ \ ell_2 $结合的扰动),这些假设在训练数据上实现了轻度编码差距。此外,我们提供了稳健性证书以进行端到端分类。我们通过计算对真实数据的认证准确性来证明我们的分析的适用性,并与其他替代方案进行认证的鲁棒性进行比较。
Several recent results provide theoretical insights into the phenomena of adversarial examples. Existing results, however, are often limited due to a gap between the simplicity of the models studied and the complexity of those deployed in practice. In this work, we strike a better balance by considering a model that involves learning a representation while at the same time giving a precise generalization bound and a robustness certificate. We focus on the hypothesis class obtained by combining a sparsity-promoting encoder coupled with a linear classifier, and show an interesting interplay between the expressivity and stability of the (supervised) representation map and a notion of margin in the feature space. We bound the robust risk (to $\ell_2$-bounded perturbations) of hypotheses parameterized by dictionaries that achieve a mild encoder gap on training data. Furthermore, we provide a robustness certificate for end-to-end classification. We demonstrate the applicability of our analysis by computing certified accuracy on real data, and compare with other alternatives for certified robustness.