论文标题
弱监督本地化的图像间沟通
Inter-Image Communication for Weakly Supervised Localization
论文作者
论文摘要
弱监督的本地化旨在仅使用图像级监督找到目标对象区域。但是,由于缺乏精细的像素级监督,从分类网络中提取的本地化图通常不准确。在本文中,我们建议利用不同对象的像素级相似性,以互补的方式学习更准确的对象位置。特别是,提出了两种约束,以促使对象特征在同一类别中的一致性。第一个约束是学习从批处理中不同图像随机采样的判别像素之间的随机特征一致性。可以利用嵌入在一个图像中的歧视性信息,以使其与图像间交流的对应物受益。第二个约束是学习整个数据集中对象功能的全局一致性。我们学习每个类别的功能中心,并通过强迫对象功能接近特定于类的中心来实现全局功能一致性。全球中心通过培训过程进行了积极更新。这两个约束可以互相受益,以学习同一类别中一致的像素级特征,并最终提高本地化图的质量。我们在两个流行的基准测试(即ILSVRC和CUB-20011)上进行了广泛的实验。我们的方法在ILSVRC验证集上达到了45.17%的前1位定位错误率,并以很大的边距超过了最新方法。该代码可从https://github.com/xiaomengyc/i2c获得。
Weakly supervised localization aims at finding target object regions using only image-level supervision. However, localization maps extracted from classification networks are often not accurate due to the lack of fine pixel-level supervision. In this paper, we propose to leverage pixel-level similarities across different objects for learning more accurate object locations in a complementary way. Particularly, two kinds of constraints are proposed to prompt the consistency of object features within the same categories. The first constraint is to learn the stochastic feature consistency among discriminative pixels that are randomly sampled from different images within a batch. The discriminative information embedded in one image can be leveraged to benefit its counterpart with inter-image communication. The second constraint is to learn the global consistency of object features throughout the entire dataset. We learn a feature center for each category and realize the global feature consistency by forcing the object features to approach class-specific centers. The global centers are actively updated with the training process. The two constraints can benefit each other to learn consistent pixel-level features within the same categories, and finally improve the quality of localization maps. We conduct extensive experiments on two popular benchmarks, i.e., ILSVRC and CUB-200-2011. Our method achieves the Top-1 localization error rate of 45.17% on the ILSVRC validation set, surpassing the current state-of-the-art method by a large margin. The code is available at https://github.com/xiaomengyc/I2C.