论文标题

带有嘈杂标签的深度学习的归一化损失功能

Normalized Loss Functions for Deep Learning with Noisy Labels

论文作者

Ma, Xingjun, Huang, Hanxun, Wang, Yisen, Romano, Simone, Erfani, Sarah, Bailey, James

论文摘要

在存在嘈杂(不正确)标签的情况下,强大的损失功能对于训练准确的深神经网络(DNN)至关重要。已经表明,常用的交叉熵(CE)损失对嘈杂标签并不强大。尽管已经设计了新的损失功能,但它们只是部分强大的。在本文中,我们从理论上显示了一个简单的归一化来表明:任何损失都可以使嘈杂的标签变得可靠。但是,实际上,仅仅是稳健的不足以使损失功能训练准确的DNN。通过调查几个强大的损失功能,我们发现它们遇到了不足的问题。为了解决这个问题,我们提出了一个框架来构建称为主动被动损失(APL)的强大损失功能。 APL结合了两个强大的损失函数,它们相互增强。基准数据集上的实验表明,我们的APL框架创建的新损失函数家族可以始终如一地超过最先进的方法,尤其是在诸如60%或80%不正确标签之类的大噪声速率下。

Robust loss functions are essential for training accurate deep neural networks (DNNs) in the presence of noisy (incorrect) labels. It has been shown that the commonly used Cross Entropy (CE) loss is not robust to noisy labels. Whilst new loss functions have been designed, they are only partially robust. In this paper, we theoretically show by applying a simple normalization that: any loss can be made robust to noisy labels. However, in practice, simply being robust is not sufficient for a loss function to train accurate DNNs. By investigating several robust loss functions, we find that they suffer from a problem of underfitting. To address this, we propose a framework to build robust loss functions called Active Passive Loss (APL). APL combines two robust loss functions that mutually boost each other. Experiments on benchmark datasets demonstrate that the family of new loss functions created by our APL framework can consistently outperform state-of-the-art methods by large margins, especially under large noise rates such as 60% or 80% incorrect labels.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源