论文标题
单峰浓缩损失:序数回归的完全自适应标签分布学习
Unimodal-Concentrated Loss: Fully Adaptive Label Distribution Learning for Ordinal Regression
论文作者
论文摘要
从标签分布中学习,在诸如面部年龄和头部姿势估计等序数回归任务上取得了令人鼓舞的结果,其中,自适应标签分布学习(ALDL)的概念最近引起了人们在理论上的优势方面引起了很多关注。但是,与假设固定形式标签分布的方法相比,ALDL方法没有达到更好的性能。我们认为现有的ALDL算法并不能完全利用序数回归的内在特性。在本文中,我们强调说,在序数回归任务上学习自适应标签分布应遵循三个原则。首先,与地面真相相对应的概率应该是标签分布中最高的。其次,相邻标签的概率应随着距离地面真相的距离的增加而降低,即分布是单模式的。第三,由于难度和歧义的水平不同,标签分布应随着样品的变化而变化,甚至在具有相同标签的不同情况下也有所不同。在这些原则的前提下,我们提出了一种新的自适应标签分布学习的新型损失函数,即单峰浓缩的损失。具体而言,从学习来对策略进行排名的单峰损失将分布限制为单峰。此外,将特定样本的估计误差和预测分布的方差集成到拟议的集中损失中,以使预测的分布在基本真相处最大化,并根据预测的不确定性而变化。与现有损失函数相比,对典型的序数回归任务(包括年龄和头部姿势估计)的广泛实验结果表明,我们所提出的单偶性损失的优越性。
Learning from a label distribution has achieved promising results on ordinal regression tasks such as facial age and head pose estimation wherein, the concept of adaptive label distribution learning (ALDL) has drawn lots of attention recently for its superiority in theory. However, compared with the methods assuming fixed form label distribution, ALDL methods have not achieved better performance. We argue that existing ALDL algorithms do not fully exploit the intrinsic properties of ordinal regression. In this paper, we emphatically summarize that learning an adaptive label distribution on ordinal regression tasks should follow three principles. First, the probability corresponding to the ground-truth should be the highest in label distribution. Second, the probabilities of neighboring labels should decrease with the increase of distance away from the ground-truth, i.e., the distribution is unimodal. Third, the label distribution should vary with samples changing, and even be distinct for different instances with the same label, due to the different levels of difficulty and ambiguity. Under the premise of these principles, we propose a novel loss function for fully adaptive label distribution learning, namely unimodal-concentrated loss. Specifically, the unimodal loss derived from the learning to rank strategy constrains the distribution to be unimodal. Furthermore, the estimation error and the variance of the predicted distribution for a specific sample are integrated into the proposed concentrated loss to make the predicted distribution maximize at the ground-truth and vary according to the predicting uncertainty. Extensive experimental results on typical ordinal regression tasks including age and head pose estimation, show the superiority of our proposed unimodal-concentrated loss compared with existing loss functions.