论文标题

通过相互蒸馏的在线深度度量学习

Online Deep Metric Learning via Mutual Distillation

论文作者

Liu, Gao-Dong, Zhao, Wan-Lei, Zhao, Jie

论文摘要

深度度量学习旨在将输入数据转换为嵌入式空间,在该空间中,相似的样本接近,而不同的样本彼此相距遥远。在实践中,新类别的样本逐渐到达,这需要学到的模型的定期增强。新类别的微调通常会导致旧的表现不佳,这被称为“灾难性遗忘”。现有的解决方案要么从头开始重新培训模型,要么需要在培训期间重播旧样品。在本文中,提出了一个完整的在线深度度量学习框架,该框架是基于一项任务和多任务场景的相互蒸馏而提出的。与师生的框架不同,拟议的方法将旧的和新的学习任务都同样重要。对旧知识或新知识没有偏好。此外,还提出了一种新型的虚拟特征估计方法,以恢复旧模型所提取的假定的特征。它允许在新型号和旧型号之间进行蒸馏,而无需重播旧训练样本或在培训期间持有旧型号。一项全面的研究表明,在不同的骨架的支持下,我们的方法的出色表现。

Deep metric learning aims to transform input data into an embedding space, where similar samples are close while dissimilar samples are far apart from each other. In practice, samples of new categories arrive incrementally, which requires the periodical augmentation of the learned model. The fine-tuning on the new categories usually leads to poor performance on the old, which is known as "catastrophic forgetting". Existing solutions either retrain the model from scratch or require the replay of old samples during the training. In this paper, a complete online deep metric learning framework is proposed based on mutual distillation for both one-task and multi-task scenarios. Different from the teacher-student framework, the proposed approach treats the old and new learning tasks with equal importance. No preference over the old or new knowledge is caused. In addition, a novel virtual feature estimation approach is proposed to recover the features assumed to be extracted by the old models. It allows the distillation between the new and the old models without the replay of old training samples or the holding of old models during the training. A comprehensive study shows the superior performance of our approach with the support of different backbones.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源