论文标题
动态纹理视频细分的无监督学习共识模型
Unsupervised Learning Consensus Model for Dynamic Texture Videos Segmentation
论文作者
论文摘要
动态纹理(DT)细分和视频处理一般,目前由基于深层神经网络的方法广泛主导,这些方法需要部署大量层。尽管这种参数方法对动态纹理分割显示出了出色的性能,但所有当前的深度学习方法都遭受与缺乏足够的参考注释相关的重要主要弱点,以训练模型并使其发挥作用。这项研究探讨了在没有培训数据来细分新视频的情况下可以使用的无监督分割方法。我们提出了一个有效的无监督学习共识模型,用于分割动态纹理(ULCM)。该模型旨在合并包含多个和弱质量区域的不同分割图,以实现更准确的分割结果。组合过程所需的不同标记字段是通过应用于输入视频的简化分组方案获得的(基于三个正交平面:XY,YT和XT)。在提出的模型中,所需的局部二进制模式(LBP)直方图的值集围绕要分类的像素,用作表示视频中复制的空间和时间信息的特征。在具有挑战性的合成数据集上进行的实验表明,与当前需要参数估计或训练步骤的当前动态纹理分割方法相反,ULCM明显更快,更易于代码,简单且参数有限。基于YUP ++数据集的进一步定性实验证明了ULCM的有效和竞争性。
Dynamic texture (DT) segmentation, and video processing in general, is currently widely dominated by methods based on deep neural networks that require the deployment of a large number of layers. Although this parametric approach has shown superior performances for the dynamic texture segmentation, all current deep learning methods suffer from a significant main weakness related to the lack of a sufficient reference annotation to train models and to make them functional. This study explores the unsupervised segmentation approach that can be used in the absence of training data to segment new videos. We present an effective unsupervised learning consensus model for the segmentation of dynamic texture (ULCM). This model is designed to merge different segmentation maps that contain multiple and weak quality regions in order to achieve a more accurate final result of segmentation. The diverse labeling fields required for the combination process are obtained by a simplified grouping scheme applied to an input video (on the basis of a three orthogonal planes: xy, yt and xt). In the proposed model, the set of values of the requantized local binary patterns (LBP) histogram around the pixel to be classified are used as features which represent both the spatial and temporal information replicated in the video. Experiments conducted on the challenging SynthDB dataset show that, contrary to current dynamic texture segmentation approaches that either require parameter estimation or a training step, ULCM is significantly faster, easier to code, simple and has limited parameters. Further qualitative experiments based on the YUP++ dataset prove the efficiently and competitively of the ULCM.