论文标题
类别级别对象姿势通过神经分析进行估计
Category Level Object Pose Estimation via Neural Analysis-by-Synthesis
论文作者
论文摘要
许多对象构成估计算法依赖于按合法的分析框架,该框架需要明确表示单个对象实例。在本文中,我们将基于梯度的拟合过程与一个参数神经图像合成模块相结合,该模块能够隐式表示整个对象类别的外观,形状和姿势,从而使每个对象实例不必要地需要明确的CAD模型。图像合成网络旨在有效地跨越姿势配置空间,以便可以将模型容量共同捕获形状和局部外观(即纹理)变化。在推理时,通过基于外观的损失将综合图像与目标进行比较,并且误差信号通过网络反向传播到输入参数。保持网络参数固定,这允许以关节方式进行对象姿势,形状和外观的迭代优化,我们在实验上表明该方法可以单独从2D图像中恢复具有高精度的对象的方向。当提供深度测量值时,为了克服规模的歧义,该方法可以准确地恢复完整的6DOF姿势。
Many object pose estimation algorithms rely on the analysis-by-synthesis framework which requires explicit representations of individual object instances. In this paper we combine a gradient-based fitting procedure with a parametric neural image synthesis module that is capable of implicitly representing the appearance, shape and pose of entire object categories, thus rendering the need for explicit CAD models per object instance unnecessary. The image synthesis network is designed to efficiently span the pose configuration space so that model capacity can be used to capture the shape and local appearance (i.e., texture) variations jointly. At inference time the synthesized images are compared to the target via an appearance based loss and the error signal is backpropagated through the network to the input parameters. Keeping the network parameters fixed, this allows for iterative optimization of the object pose, shape and appearance in a joint manner and we experimentally show that the method can recover orientation of objects with high accuracy from 2D images alone. When provided with depth measurements, to overcome scale ambiguities, the method can accurately recover the full 6DOF pose successfully.