论文标题
利用合成数据在不利条件下学习视频稳定
Leveraging Synthetic Data to Learn Video Stabilization Under Adverse Conditions
论文作者
论文摘要
视频稳定在提高视频质量方面起着核心作用。但是,尽管这些方法取得了长足的进展,但它们主要是在标准天气和照明条件下进行的,并且在不利条件下的性能可能差。在本文中,我们提出了一种综合感知的不利天气鲁棒算法,用于视频稳定,该算法不需要真实数据,并且只能在合成数据上接受培训。我们还提出了Silver,这是一种新颖的渲染引擎,可通过自动地面真相提取程序生成所需的训练数据。我们的方法使用我们特殊生成的合成数据来训练仿射转换矩阵估计器,避免了当前方法面临的特征提取问题。此外,由于在不利条件下没有视频稳定数据集,因此我们提出了新颖的VSAC105REAL数据集以进行评估。我们将方法与使用两个基准测试的五种最先进的视频稳定算法进行了比较。我们的结果表明,当前方法在至少一个天气条件下的表现较差,即使在具有合成数据的小数据集中训练,我们就在考虑所有天气状况时就可以在稳定性平均得分,失真得分,成功率和平均种植比方面取得最佳性能。因此,我们的视频稳定模型在现实世界的视频上很好地概括了,并且不需要大规模的合成训练数据来收敛。
Video stabilization plays a central role to improve videos quality. However, despite the substantial progress made by these methods, they were, mainly, tested under standard weather and lighting conditions, and may perform poorly under adverse conditions. In this paper, we propose a synthetic-aware adverse weather robust algorithm for video stabilization that does not require real data and can be trained only on synthetic data. We also present Silver, a novel rendering engine to generate the required training data with an automatic ground-truth extraction procedure. Our approach uses our specially generated synthetic data for training an affine transformation matrix estimator avoiding the feature extraction issues faced by current methods. Additionally, since no video stabilization datasets under adverse conditions are available, we propose the novel VSAC105Real dataset for evaluation. We compare our method to five state-of-the-art video stabilization algorithms using two benchmarks. Our results show that current approaches perform poorly in at least one weather condition, and that, even training in a small dataset with synthetic data only, we achieve the best performance in terms of stability average score, distortion score, success rate, and average cropping ratio when considering all weather conditions. Hence, our video stabilization model generalizes well on real-world videos and does not require large-scale synthetic training data to converge.