论文标题
在语义细分中从量表不变的示例中学习域的适应
Learning from Scale-Invariant Examples for Domain Adaptation in Semantic Segmentation
论文作者
论文摘要
语义分割模型的无监督域适应(UDA)的自学学习方法遭受了预测和选择合理优质的伪标签的挑战。在本文中,我们提出了一种利用语义分割模型的规模不变特性的新方法,以进行自我监督域的适应性。我们的算法是基于一个合理的假设,即,无论对象和物体的大小(给定上下文),语义标记应不变。我们表明,该约束在目标域的图像上违反了,因此可以用来将标记在不同规模的贴片之间传输标签。具体而言,我们表明,与呈现的原始尺寸图像相比,当带有缩放目标域的缩放贴片时,语义分割模型在呈缩放量的目标斑块时产生高熵。这些规模不变的示例是从目标域最自信的图像中提取的。提出了动态类特异性熵阈值机制,以滤除不可靠的伪标记。此外,我们还结合了焦点损失,以解决自我监督学习中阶级失衡问题的问题。已经进行了广泛的实验,结果表明利用规模不变标签,我们的表现优于现有的基于自我监督的最新域适应方法。具体而言,我们通过VGG16-FCN8基线网络获得了GTA5的铅和合成的GTA5铅的1.3%和3.8%。
Self-supervised learning approaches for unsupervised domain adaptation (UDA) of semantic segmentation models suffer from challenges of predicting and selecting reasonable good quality pseudo labels. In this paper, we propose a novel approach of exploiting scale-invariance property of the semantic segmentation model for self-supervised domain adaptation. Our algorithm is based on a reasonable assumption that, in general, regardless of the size of the object and stuff (given context) the semantic labeling should be unchanged. We show that this constraint is violated over the images of the target domain, and hence could be used to transfer labels in-between differently scaled patches. Specifically, we show that semantic segmentation model produces output with high entropy when presented with scaled-up patches of target domain, in comparison to when presented original size images. These scale-invariant examples are extracted from the most confident images of the target domain. Dynamic class specific entropy thresholding mechanism is presented to filter out unreliable pseudo-labels. Furthermore, we also incorporate the focal loss to tackle the problem of class imbalance in self-supervised learning. Extensive experiments have been performed, and results indicate that exploiting the scale-invariant labeling, we outperform existing self-supervised based state-of-the-art domain adaptation methods. Specifically, we achieve 1.3% and 3.8% of lead for GTA5 to Cityscapes and SYNTHIA to Cityscapes with VGG16-FCN8 baseline network.