论文标题
使用对抗空间金字塔网络的遥感图像的道路细分
Road Segmentation for Remote Sensing Images using Adversarial Spatial Pyramid Networks
论文作者
论文摘要
对于广泛的应用,遥感图像中的道路提取非常重要。由于具有复杂的背景和高密度,因此大多数现有方法无法准确提取看起来正确和完整的道路网络。此外,他们遭受了培训数据不足或手动注释的高成本。为了解决这些问题,我们引入了一种新模型,以将结构化域的适应性应用于合成图像的产生和路段。我们将特征金字塔网络纳入生成对抗网络中,以最大程度地减少源和目标域之间的差异。学会了发电机来产生质量的合成图像,并且歧视者试图区分它们。我们还提出了一个特征金字塔网络,该网络通过从网络的所有层中提取有效特征来描述不同尺度对象,从而改善了所提出的模型的性能。确实,引入了一种新颖的规模架构,以从多层次特征地图中学习并改善功能的语义。为了优化,该模型是通过联合重建损耗函数训练的,该功能最大程度地减少了假图像与真实图像之间的差异。在三个数据集上进行的广泛实验证明了拟议方法在准确性和效率方面的出色性能。特别是,我们的模型在马萨诸塞州的数据集上实现了最先进的78.86 IOU,该数据集具有14.89亿参数和86.78b拖鞋,比在评估中使用的最好的表演者中的最高表现者中的拖鞋少4倍,但准确性较低(+3.47%IOU)。
Road extraction in remote sensing images is of great importance for a wide range of applications. Because of the complex background, and high density, most of the existing methods fail to accurately extract a road network that appears correct and complete. Moreover, they suffer from either insufficient training data or high costs of manual annotation. To address these problems, we introduce a new model to apply structured domain adaption for synthetic image generation and road segmentation. We incorporate a feature pyramid network into generative adversarial networks to minimize the difference between the source and target domains. A generator is learned to produce quality synthetic images, and the discriminator attempts to distinguish them. We also propose a feature pyramid network that improves the performance of the proposed model by extracting effective features from all the layers of the network for describing different scales objects. Indeed, a novel scale-wise architecture is introduced to learn from the multi-level feature maps and improve the semantics of the features. For optimization, the model is trained by a joint reconstruction loss function, which minimizes the difference between the fake images and the real ones. A wide range of experiments on three datasets prove the superior performance of the proposed approach in terms of accuracy and efficiency. In particular, our model achieves state-of-the-art 78.86 IOU on the Massachusetts dataset with 14.89M parameters and 86.78B FLOPs, with 4x fewer FLOPs but higher accuracy (+3.47% IOU) than the top performer among state-of-the-art approaches used in the evaluation.