论文标题
来回查看:带有明确时间差异建模的视频超分辨率
Look Back and Forth: Video Super-Resolution with Explicit Temporal Difference Modeling
论文作者
论文摘要
时间建模对于视频超分辨率至关重要。大多数视频超分辨率方法采用光流或可变形卷积以明确运动补偿。但是,这种时间建模技术会增加模型的复杂性,并可能在闭塞或复杂运动的情况下失败,从而导致严重的失真和伪影。在本文中,我们建议探讨LR和HR空间中明确的时间差异建模的作用。我们建议根据差异水平将连续帧直接馈入VSR模型,而是建议计算帧之间的时间差,并根据差异将这些像素分为两个子集。它们分别使用两个不同接受场的两个分支进行处理,以更好地提取互补信息。为了进一步增强超分辨率结果,不仅提取了空间残差特征,而且还计算了高频域中连续帧之间的差异。它允许模型利用中间的SR在将来和过去都会导致改进当前的SR输出。可以缓存在不同时间步骤的差异,以便可以将更多时间的信息传播到当前框架以进行改进。在几个视频超分辨率基准数据集上进行的实验证明了该方法的有效性及其在最新方法上的有利性能。
Temporal modeling is crucial for video super-resolution. Most of the video super-resolution methods adopt the optical flow or deformable convolution for explicitly motion compensation. However, such temporal modeling techniques increase the model complexity and might fail in case of occlusion or complex motion, resulting in serious distortion and artifacts. In this paper, we propose to explore the role of explicit temporal difference modeling in both LR and HR space. Instead of directly feeding consecutive frames into a VSR model, we propose to compute the temporal difference between frames and divide those pixels into two subsets according to the level of difference. They are separately processed with two branches of different receptive fields in order to better extract complementary information. To further enhance the super-resolution result, not only spatial residual features are extracted, but the difference between consecutive frames in high-frequency domain is also computed. It allows the model to exploit intermediate SR results in both future and past to refine the current SR output. The difference at different time steps could be cached such that information from further distance in time could be propagated to the current frame for refinement. Experiments on several video super-resolution benchmark datasets demonstrate the effectiveness of the proposed method and its favorable performance against state-of-the-art methods.