论文标题

排量不变的匹配成本学习以进行准确的光流估计

Displacement-Invariant Matching Cost Learning for Accurate Optical Flow Estimation

论文作者

Wang, Jianyuan, Zhong, Yiran, Dai, Yuchao, Zhang, Kaihao, Ji, Pan, Li, Hongdong

论文摘要

学习匹配成本已被证明对于最先进的深入立体声匹配方法的成功至关重要,在该方法中,在4D功能量上应用3D卷积以学习3D成本量。但是,这种机制从未用于光流任务。这主要是由于在光流计算的情况下显着增加了搜索维度,即直接扩展将需要密集的4D卷积才能处理5D特征量,这是计算上的较高的。本文提出了一种新颖的解决方案,该解决方案能够绕开构建5D功能量的要求,同时仍允许网络从数据中学习合适的匹配成本。我们的关键创新是将2D位移之间的连接解除,并在每个2D位移假设中独立学习匹配成本,即,流离失所的成本学习。具体而言,我们在每个2D位移假设上独立应用相同的基于2D卷积的匹配网,以学习4D成本量。此外,我们提出一个位移感知投影层来扩展学习的成本量,这重新构建了不同位移候选者之间的相关性,并减轻了学习成本量中的多模式问题。然后,将成本量投影到通过2D软弧线层的光流估计。广泛的实验表明,我们的方法在各种数据集上实现了最新的精度,并且在Sintel基准测试上都超越了所有发布的光流方法。

Learning matching costs has been shown to be critical to the success of the state-of-the-art deep stereo matching methods, in which 3D convolutions are applied on a 4D feature volume to learn a 3D cost volume. However, this mechanism has never been employed for the optical flow task. This is mainly due to the significantly increased search dimension in the case of optical flow computation, ie, a straightforward extension would require dense 4D convolutions in order to process a 5D feature volume, which is computationally prohibitive. This paper proposes a novel solution that is able to bypass the requirement of building a 5D feature volume while still allowing the network to learn suitable matching costs from data. Our key innovation is to decouple the connection between 2D displacements and learn the matching costs at each 2D displacement hypothesis independently, ie, displacement-invariant cost learning. Specifically, we apply the same 2D convolution-based matching net independently on each 2D displacement hypothesis to learn a 4D cost volume. Moreover, we propose a displacement-aware projection layer to scale the learned cost volume, which reconsiders the correlation between different displacement candidates and mitigates the multi-modal problem in the learned cost volume. The cost volume is then projected to optical flow estimation through a 2D soft-argmin layer. Extensive experiments show that our approach achieves state-of-the-art accuracy on various datasets, and outperforms all published optical flow methods on the Sintel benchmark.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源