论文标题
重新审视多视图光度立体声
Multi-View Photometric Stereo Revisited
论文作者
论文摘要
多视图光度立体声(MVP)是一种从图像中详细且精确的3D获取的首选方法。尽管流行的MVP方法可以提供出色的结果,但它们通常很复杂,并且限于各向同性材料对象。为了解决此类局限性,我们提出了一种简单,实用的MVP方法,该方法适用于各向同性以及其他对象材料类型(例如各向异性和光泽)。本文中提出的方法利用了深度神经网络中不确定性建模的好处,用于可靠的光度立体声(PS)和多视图立体声(MVS)网络预测。然而,与最近提出的最新作品相反,我们引入了神经量渲染方法,以供MVS和PS测量值可信赖。引入神经音量渲染的优点是,它有助于对具有不同材料类型的对象的可靠建模,其中现有MVS方法,PS方法或两者都可能失败。此外,它使我们能够从事神经3D形状表示,最近显示了许多几何处理任务的出色结果。我们建议的新损失函数旨在使用最某些MVS和PS网络预测以及加权神经量渲染成本符合隐式神经函数的零级集。当在几个基准数据集上进行广泛测试时,提出的方法显示了最先进的结果。
Multi-view photometric stereo (MVPS) is a preferred method for detailed and precise 3D acquisition of an object from images. Although popular methods for MVPS can provide outstanding results, they are often complex to execute and limited to isotropic material objects. To address such limitations, we present a simple, practical approach to MVPS, which works well for isotropic as well as other object material types such as anisotropic and glossy. The proposed approach in this paper exploits the benefit of uncertainty modeling in a deep neural network for a reliable fusion of photometric stereo (PS) and multi-view stereo (MVS) network predictions. Yet, contrary to the recently proposed state-of-the-art, we introduce neural volume rendering methodology for a trustworthy fusion of MVS and PS measurements. The advantage of introducing neural volume rendering is that it helps in the reliable modeling of objects with diverse material types, where existing MVS methods, PS methods, or both may fail. Furthermore, it allows us to work on neural 3D shape representation, which has recently shown outstanding results for many geometric processing tasks. Our suggested new loss function aims to fits the zero level set of the implicit neural function using the most certain MVS and PS network predictions coupled with weighted neural volume rendering cost. The proposed approach shows state-of-the-art results when tested extensively on several benchmark datasets.