论文标题

通过图形操作搜索自适应互动建模

Adaptive Interaction Modeling via Graph Operations Search

论文作者

Li, Haoxin, Zheng, Wei-Shi, Tao, Yu, Hu, Haifeng, Lai, Jian-Huang

论文摘要

交互建模对于视频动作分析很重要。最近,一些作品设计了特定的结构,以模拟视频中的交互。但是,它们的结构是手动设计和非自适应的,需要结构设计工作,更重要的是无法自适应地建模。在本文中,我们自动化结构设计的过程,以学习相互作用建模的自适应结构。我们建议使用可区分的体系结构搜索机制搜索网络结构,该机制学会构建不同视频的自适应结构,以促进自适应互动建模。为此,我们首先使用几个基本的图形操作来设计搜索空间,这些操作明确捕获了视频中的不同关系。我们在实验上证明,我们的体系结构搜索框架学会了构建自适应互动建模结构,从而提供了对结构和某些交互特征之间关系的更多了解,并释放了结构设计工作的要求。此外,我们表明搜索空间中设计的基本图形操作能够建模视频中的不同交互。两个交互数据集的实验表明,我们的方法可以通过最先进的方法实现竞争性能。

Interaction modeling is important for video action analysis. Recently, several works design specific structures to model interactions in videos. However, their structures are manually designed and non-adaptive, which require structures design efforts and more importantly could not model interactions adaptively. In this paper, we automate the process of structures design to learn adaptive structures for interaction modeling. We propose to search the network structures with differentiable architecture search mechanism, which learns to construct adaptive structures for different videos to facilitate adaptive interaction modeling. To this end, we first design the search space with several basic graph operations that explicitly capture different relations in videos. We experimentally demonstrate that our architecture search framework learns to construct adaptive interaction modeling structures, which provides more understanding about the relations between the structures and some interaction characteristics, and also releases the requirement of structures design efforts. Additionally, we show that the designed basic graph operations in the search space are able to model different interactions in videos. The experiments on two interaction datasets show that our method achieves competitive performance with state-of-the-arts.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源