论文标题
基于CNN的点光度光度立体声问题的方法
A CNN Based Approach for the Point-Light Photometric Stereo Problem
论文作者
论文摘要
在不同的光源下,使用几个图像重建对象的3D形状是一项非常具有挑战性的任务,尤其是当考虑到诸如光传播和衰减之类的现实假设时,考虑了观察几何形状和镜面反射。许多解决光度立体声(PS)问题的作品通常会放松上述大部分假设。特别是他们忽略了镜面反射和全球照明效应。在这项工作中,我们提出了一种基于CNN的方法,能够通过利用深度神经网络的最新改进来处理这些现实的假设,以用于远场光度立体声并将其适应点光设置。我们通过采用点光PS的迭代程序来实现这一目标,以实现两个主要步骤的形状估计。首先,我们训练每像素CNN,从反射样品中预测表面垂直。其次,我们通过整合正常场来计算深度,以迭代估算光方向和衰减,该方向和衰减用于补偿输入图像以计算下一次迭代的反射样品。 我们的方法毫无疑问地超过了勤奋的现实世界数据集上的最新方法。此外,为了衡量我们的近场点光源PS数据的方法的性能,我们介绍了Luces的第一个现实世界中的“用于近场点光源光源光度立体声”的数据集,这是点光源的效果和观点的效果。我们的方法也表现出在该数据集上的竞争。数据和测试代码可在项目页面上找到。
Reconstructing the 3D shape of an object using several images under different light sources is a very challenging task, especially when realistic assumptions such as light propagation and attenuation, perspective viewing geometry and specular light reflection are considered. Many of works tackling Photometric Stereo (PS) problems often relax most of the aforementioned assumptions. Especially they ignore specular reflection and global illumination effects. In this work, we propose a CNN-based approach capable of handling these realistic assumptions by leveraging recent improvements of deep neural networks for far-field Photometric Stereo and adapt them to the point light setup. We achieve this by employing an iterative procedure of point-light PS for shape estimation which has two main steps. Firstly we train a per-pixel CNN to predict surface normals from reflectance samples. Secondly, we compute the depth by integrating the normal field in order to iteratively estimate light directions and attenuation which is used to compensate the input images to compute reflectance samples for the next iteration. Our approach sigificantly outperforms the state-of-the-art on the DiLiGenT real world dataset. Furthermore, in order to measure the performance of our approach for near-field point-light source PS data, we introduce LUCES the first real-world 'dataset for near-fieLd point light soUrCe photomEtric Stereo' of 14 objects of different materials were the effects of point light sources and perspective viewing are a lot more significant. Our approach also outperforms the competition on this dataset as well. Data and test code are available at the project page.