论文标题
空中图像语义分割的对抗性损失
Adversarial Loss for Semantic Segmentation of Aerial Imagery
论文作者
论文摘要
从空中图像中自动建筑物提取在城市规划,灾难管理和变更检测中有多种应用。近年来,几项作品采用了深层卷积神经网络(CNN)来进行提取,因为它们产生了丰富的特征,这些功能在照明条件,阴影等方面不变。大多数深度学习细分方法在不了解背景的情况下优化了每个像素的损失。这通常会导致可能导致丢失或未精制区域的输出。在这项工作中,我们提出了一种新型的损失功能,结合了对抗性和横向损失,这些损失学会了了解语义分割的局部和全球环境。在DeepLab V3+网络上部署的新提出的损失功能在马萨诸塞州建筑物数据集中获得了最新结果。损耗函数改善了结构并完善建筑物的边缘,而无需任何常用的后处理方法,例如条件随机场。我们还进行消融研究以了解对抗损失的影响。最后,与以前的94.88%的最佳F1相比,这项提出的方法在马萨诸塞州建筑物数据集上的轻松F1得分为95.59%。
Automatic building extraction from aerial imagery has several applications in urban planning, disaster management, and change detection. In recent years, several works have adopted deep convolutional neural networks (CNNs) for building extraction, since they produce rich features that are invariant against lighting conditions, shadows, etc. Although several advances have been made, building extraction from aerial imagery still presents multiple challenges. Most of the deep learning segmentation methods optimize the per-pixel loss with respect to the ground truth without knowledge of the context. This often leads to imperfect outputs that may lead to missing or unrefined regions. In this work, we propose a novel loss function combining both adversarial and cross-entropy losses that learn to understand both local and global contexts for semantic segmentation. The newly proposed loss function deployed on the DeepLab v3+ network obtains state-of-the-art results on the Massachusetts buildings dataset. The loss function improves the structure and refines the edges of buildings without requiring any of the commonly used post-processing methods, such as Conditional Random Fields. We also perform ablation studies to understand the impact of the adversarial loss. Finally, the proposed method achieves a relaxed F1 score of 95.59% on the Massachusetts buildings dataset compared to the previous best F1 of 94.88%.