Learn to Triangulate Scene Coordinates for Visual Localization

Xiang Guo, Tianrui Chen, Bo Li, Qi Liu, Huarong Jia, Yuchao Dai

科研成果: 期刊稿件文章同行评审

1 引用 (Scopus)

摘要

Visual localization plays a critical role in robotics. The scene coordinate regression-based localization methods have achieved state-of-the-art performance. However, current methods still have a gap between the regressed scene coordinates and the ground truth scene coordinates, which hinders the improvement of localization accuracy. These methods generally use structure-from-motion (SfM) or depth sensors to generate proxy scene coordinate supervision labels for training, but these proxy labels are contaminated with errors and noises, which are sub-optimal for training. To resolve this issue, we introduce a simple yet effective triangulation constraint, which could be easily incorporated into any scene coordinate regression-based framework. Instead of directly regressing the scene coordinates, our constraint reinforces the network, which learns to triangulate the ground truth scene coordinates without any proxy scene coordinate labels for supervision. Extensive experiments across multiple public datasets show that our triangulation constraint establishes significant improvement and even achieves better results without proxy labels for supervision. Furthermore, our method could recover denser and more complete 3D models compared with the SfM and other localization methods.

源语言英语
页(从-至)3339-3346
页数8
期刊IEEE Robotics and Automation Letters
9
4
DOI
出版状态已出版 - 1 4月 2024

指纹

探究 'Learn to Triangulate Scene Coordinates for Visual Localization' 的科研主题。它们共同构成独一无二的指纹。

引用此