论文标题
改善图像样式转移的潜在空间
Improving the Latent Space of Image Style Transfer
论文作者
论文摘要
现有的神经风格转移研究已经研究了,以匹配内容和样式图像的深度特征之间的统计信息,这些图像是由预先训练的VGG提取的,并在合成的艺术图像方面取得了重大改进。但是,在某些情况下,预先训练的编码器的特征统计数据可能与我们所感知的视觉样式不一致。例如,不同样式图像之间的样式距离小于相同样式的样式。在这种不适当的潜在空间中,现有方法的目标函数将在错误的方向上进行优化,从而导致不良的风格化结果。此外,预先训练的编码器提取的功能中缺乏内容细节也导致内容泄漏问题。为了在样式转移使用的潜在空间中解决这些问题,我们提出了两个对比训练方案,以获取更适合此任务的精制编码器。样式的对比损失使风格化的结果更接近相同的视觉样式图像,并将其从内容图像中推开。内容对比损失使编码器能够保留更多可用的详细信息。我们可以将我们的培训计划直接添加到某些现有样式转移方法中,并显着改善其结果。广泛的实验结果证明了我们方法的有效性和优势。
Existing neural style transfer researches have studied to match statistical information between the deep features of content and style images, which were extracted by a pre-trained VGG, and achieved significant improvement in synthesizing artistic images. However, in some cases, the feature statistics from the pre-trained encoder may not be consistent with the visual style we perceived. For example, the style distance between images of different styles is less than that of the same style. In such an inappropriate latent space, the objective function of the existing methods will be optimized in the wrong direction, resulting in bad stylization results. In addition, the lack of content details in the features extracted by the pre-trained encoder also leads to the content leak problem. In order to solve these issues in the latent space used by style transfer, we propose two contrastive training schemes to get a refined encoder that is more suitable for this task. The style contrastive loss pulls the stylized result closer to the same visual style image and pushes it away from the content image. The content contrastive loss enables the encoder to retain more available details. We can directly add our training scheme to some existing style transfer methods and significantly improve their results. Extensive experimental results demonstrate the effectiveness and superiority of our methods.