论文标题

通过图卷积网络进行结构的图像检索

Image Retrieval for Structure-from-Motion via Graph Convolutional Network

论文作者

Yan, Shen, Pen, Yang, Lai, Shiming, Liu, Yu, Zhang, Maojun

论文摘要

结构上移动(SFM)的常规图像检索技术受到有效识别重复模式的极限,并且无法保证创建具有高精度和高回忆的足够匹配对。在本文中,我们提出了一种基于图形卷积网络(GCN)的新型检索方法,以生成精确的成对匹配,而无需昂贵的冗余。我们将图像检索任务作为图形数据中的节点二进制分类问题提出:如果该节点与查询图像重叠,则将节点标记为正面。关键的想法是,我们发现查询图像周围特征空间中的本地上下文包含有关此图像与其邻居之间可匹配关系的丰富信息。通过构建围绕查询图像作为输入数据的子图,我们采用可学习的GCN来利用子图中的节点是否与查询照片重叠区域。实验表明,我们的方法在高度模棱两可和重复的场景的挑战性数据集中表现出色。此外,与最新的可匹配检索方法相比,提出的方法大大降低了无用的尝试匹配,而无需牺牲重建的准确性和完整性。

Conventional image retrieval techniques for Structure-from-Motion (SfM) suffer from the limit of effectively recognizing repetitive patterns and cannot guarantee to create just enough match pairs with high precision and high recall. In this paper, we present a novel retrieval method based on Graph Convolutional Network (GCN) to generate accurate pairwise matches without costly redundancy. We formulate image retrieval task as a node binary classification problem in graph data: a node is marked as positive if it shares the scene overlaps with the query image. The key idea is that we find that the local context in feature space around a query image contains rich information about the matchable relation between this image and its neighbors. By constructing a subgraph surrounding the query image as input data, we adopt a learnable GCN to exploit whether nodes in the subgraph have overlapping regions with the query photograph. Experiments demonstrate that our method performs remarkably well on the challenging dataset of highly ambiguous and duplicated scenes. Besides, compared with state-of-the-art matchable retrieval methods, the proposed approach significantly reduces useless attempted matches without sacrificing the accuracy and completeness of reconstruction.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源