论文标题
无监督的深度度量学习,转变的注意力一致性和对比群体损失
Unsupervised Deep Metric Learning with Transformed Attention Consistency and Contrastive Clustering Loss
论文作者
论文摘要
现有的无监督指标学习方法的重点是探索输入图像本身中的自我实施信息。我们观察到,在分析图像时,人眼通常会相互比较图像,而不是单独检查图像。此外,他们经常关注某些关键点,图像区域或对象之间的图像类别,但在类中高度一致。即使图像正在转换,注意力模式也将保持一致。在这一观察结果的推动下,我们开发了一种新的方法来无监督的深度度量学习,其中基于跨图像的自学信息而不是在一个图像中学习网络。为了表征图像比较期间人类注意力的一致模式,我们介绍了转化关注一致性的想法。它假设在视觉上相似的图像,即使经历了不同的图像变换,也应共享相同的一致的视觉注意图。这种一致性会导致成对的自我实施者损失,使我们能够学习一个暹罗深神经网络,以对图像进行编码和比较其转换或匹配的对。为了进一步增强该网络生成的特征的类判别能力,我们将三胞胎损失从监督公制学习中调整为无监督的情况,并引入对比度聚类损失。我们在基准数据集上进行的广泛实验结果表明,我们提出的方法优于当前的最新方法,用于大幅度的无监督度量学习。
Existing approaches for unsupervised metric learning focus on exploring self-supervision information within the input image itself. We observe that, when analyzing images, human eyes often compare images against each other instead of examining images individually. In addition, they often pay attention to certain keypoints, image regions, or objects which are discriminative between image classes but highly consistent within classes. Even if the image is being transformed, the attention pattern will be consistent. Motivated by this observation, we develop a new approach to unsupervised deep metric learning where the network is learned based on self-supervision information across images instead of within one single image. To characterize the consistent pattern of human attention during image comparisons, we introduce the idea of transformed attention consistency. It assumes that visually similar images, even undergoing different image transforms, should share the same consistent visual attention map. This consistency leads to a pairwise self-supervision loss, allowing us to learn a Siamese deep neural network to encode and compare images against their transformed or matched pairs. To further enhance the inter-class discriminative power of the feature generated by this network, we adapt the concept of triplet loss from supervised metric learning to our unsupervised case and introduce the contrastive clustering loss. Our extensive experimental results on benchmark datasets demonstrate that our proposed method outperforms current state-of-the-art methods for unsupervised metric learning by a large margin.