论文标题
FS6D:很少射击6D姿势估计新物体
FS6D: Few-Shot 6D Pose Estimation of Novel Objects
论文作者
论文摘要
6D对象姿势估计网络的能力有限,因为近距离假设及其对高保真对象CAD模型的依赖,因此可以扩展到大量对象实例。在这项工作中,我们研究了一个新的开放式问题。少数6D对象提出估计:通过一些支持视图估算未知对象的6D姿势,而无需额外的培训。为了解决该问题,我们指出了充分探索给定的支持视图和查询场景补丁之间的外观和几何关系的重要性,并通过将密集的RGBD原型与变压器提取和匹配,提出了匹配的密集原型匹配框架。此外,我们表明,来自各种外观和形状的先验对于问题设置下的概括能力至关重要,因此提出了用于网络预训练的大规模RGBD感性数据集(Shapenet6d)。还引入了一种简单有效的在线纹理混合方法,以消除合成数据集中的域间隙,该数据集以低成本丰富了外观多样性。最后,我们讨论了该问题的可能解决方案,并在流行的数据集上建立基准,以促进未来的研究。项目页面位于\ url {https://fs6d.github.io/}。
6D object pose estimation networks are limited in their capability to scale to large numbers of object instances due to the close-set assumption and their reliance on high-fidelity object CAD models. In this work, we study a new open set problem; the few-shot 6D object poses estimation: estimating the 6D pose of an unknown object by a few support views without extra training. To tackle the problem, we point out the importance of fully exploring the appearance and geometric relationship between the given support views and query scene patches and propose a dense prototypes matching framework by extracting and matching dense RGBD prototypes with transformers. Moreover, we show that the priors from diverse appearances and shapes are crucial to the generalization capability under the problem setting and thus propose a large-scale RGBD photorealistic dataset (ShapeNet6D) for network pre-training. A simple and effective online texture blending approach is also introduced to eliminate the domain gap from the synthesis dataset, which enriches appearance diversity at a low cost. Finally, we discuss possible solutions to this problem and establish benchmarks on popular datasets to facilitate future research. The project page is at \url{https://fs6d.github.io/}.