论文标题

Dipe:更深入地陷入光度错误,以无监督的深度学习和自我感动。

DiPE: Deeper into Photometric Errors for Unsupervised Learning of Depth and Ego-motion from Monocular Videos

论文作者

Jiang, Hualie, Ding, Laiyan, Sun, Zhenglong, Huang, Rui

论文摘要

无标记的单眼视频对深度和自我运动的无监督学习最近引起了人们的极大关注,这避免了在监督的人中使用昂贵的地面真相。它通过使用目标视图与综合视图之间的光度误差作为损失来实现这一目标。尽管取得了重大进展,但学习仍然遭受阻塞和场景动态的影响。本文表明,仔细操纵光度错误可以更好地解决这些困难。主要的改进是通过统计技术来实现的,该技术可以在光度误差图中掩盖看不见的或非平稳的像素,从而防止误导网络。通过这种离群掩蔽方法,可以更准确地估计对象向相反方向移动的物体的深度。据我们所知,即使在自主驾驶等应用中构成更高的风险,这种情况并未在以前的作品中得到认真考虑。我们还提出了一个有效的加权多尺度方案,以减少预测深度图中的伪影。 Kitti数据集的广泛实验显示了所提出的方法的有效性。总体系统在深度和自我估计上都达到了最先进的表现。

Unsupervised learning of depth and ego-motion from unlabelled monocular videos has recently drawn great attention, which avoids the use of expensive ground truth in the supervised one. It achieves this by using the photometric errors between the target view and the synthesized views from its adjacent source views as the loss. Despite significant progress, the learning still suffers from occlusion and scene dynamics. This paper shows that carefully manipulating photometric errors can tackle these difficulties better. The primary improvement is achieved by a statistical technique that can mask out the invisible or nonstationary pixels in the photometric error map and thus prevents misleading the networks. With this outlier masking approach, the depth of objects moving in the opposite direction to the camera can be estimated more accurately. To the best of our knowledge, such scenarios have not been seriously considered in the previous works, even though they pose a higher risk in applications like autonomous driving. We also propose an efficient weighted multi-scale scheme to reduce the artifacts in the predicted depth maps. Extensive experiments on the KITTI dataset show the effectiveness of the proposed approaches. The overall system achieves state-of-theart performance on both depth and ego-motion estimation.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源