论文标题

通过生物力学约束弱监督的3D手姿势估计

Weakly Supervised 3D Hand Pose Estimation via Biomechanical Constraints

论文作者

Spurr, Adrian, Iqbal, Umar, Molchanov, Pavlo, Hilliges, Otmar, Kautz, Jan

论文摘要

由于固有的规模和深度歧义,从2D图像中估算3D手姿势是一个困难的,反面的问题。当前的最新方法训练具有3D地面数据的完全监督的深层神经网络。但是,获取3D注释很昂贵,通常需要校准的多视图设置或劳动密集型手动注释。尽管2D关键点的注释更容易获得,但如何有效利用此类弱监督的数据来改善3D手姿势预测的任务仍然是一个重要的开放问题。关键难度源于以下事实:直接应用额外的2D监督主要受益于2D代理目标,但对减轻深度和规模歧义的影响很小。面对这一挑战,我们提出了一系列新颖的损失。我们通过广泛的实验表明,我们提出的约束大大降低了深度歧义,并允许网络更有效地利用其他2D注释的图像。例如,在具有挑战性的Freihand数据集上,使用其他2D注释而没有我们提出的生物力学约束,将深度误差仅减少$ 15 \%$,而当使用拟议的生物力学约束时,该错误将显着降低$ 50 \%$。

Estimating 3D hand pose from 2D images is a difficult, inverse problem due to the inherent scale and depth ambiguities. Current state-of-the-art methods train fully supervised deep neural networks with 3D ground-truth data. However, acquiring 3D annotations is expensive, typically requiring calibrated multi-view setups or labor intensive manual annotations. While annotations of 2D keypoints are much easier to obtain, how to efficiently leverage such weakly-supervised data to improve the task of 3D hand pose prediction remains an important open question. The key difficulty stems from the fact that direct application of additional 2D supervision mostly benefits the 2D proxy objective but does little to alleviate the depth and scale ambiguities. Embracing this challenge we propose a set of novel losses. We show by extensive experiments that our proposed constraints significantly reduce the depth ambiguity and allow the network to more effectively leverage additional 2D annotated images. For example, on the challenging freiHAND dataset using additional 2D annotation without our proposed biomechanical constraints reduces the depth error by only $15\%$, whereas the error is reduced significantly by $50\%$ when the proposed biomechanical constraints are used.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源