论文标题
在卫星图像上进行自我监督的预处理:关于标签有效的车辆检测的案例研究
Self-Supervised Pretraining on Satellite Imagery: a Case Study on Label-Efficient Vehicle Detection
论文作者
论文摘要
在与防御相关的遥感应用程序(例如卫星图像上的车辆检测)中,有监督的学习需要大量标记的示例才能达到运营性能。由于需要军事专家,因此获得了这些数据,并且一些可观察到的物质本质上很少见。这种有限的标记功能以及由于越来越多的传感器而可用的大量未标记图像,使对象检测在遥感图像上与自我监督学习高度相关。我们研究了在非常高分辨率的光学卫星图像上进行对象检测的内域自学表示学习,但探索却很差。据我们所知,我们首次研究了该任务上标签效率的问题。我们使用大型土地使用分类数据集功能图,以扩展动量对比框架的扩展。然后,我们研究了该模型对预元素专有数据的精细车辆检测和分类的真实世界任务的可转移性,该数据旨在代表战略现场监视的操作用例。我们表明,我们的内域自我监督学习模型与Imagenet预处理具有竞争力,并且在低标签方面表现优于它。
In defense-related remote sensing applications, such as vehicle detection on satellite imagery, supervised learning requires a huge number of labeled examples to reach operational performances. Such data are challenging to obtain as it requires military experts, and some observables are intrinsically rare. This limited labeling capability, as well as the large number of unlabeled images available due to the growing number of sensors, make object detection on remote sensing imagery highly relevant for self-supervised learning. We study in-domain self-supervised representation learning for object detection on very high resolution optical satellite imagery, that is yet poorly explored. For the first time to our knowledge, we study the problem of label efficiency on this task. We use the large land use classification dataset Functional Map of the World to pretrain representations with an extension of the Momentum Contrast framework. We then investigate this model's transferability on a real-world task of fine-grained vehicle detection and classification on Preligens proprietary data, which is designed to be representative of an operational use case of strategic site surveillance. We show that our in-domain self-supervised learning model is competitive with ImageNet pretraining, and outperforms it in the low-label regime.