@inproceedings{8ac83bd548ea472d9ac9b64a86783183,
title = "Mono3DVG: 3D Visual Grounding in Monocular Images",
abstract = "We introduce a novel task of 3D visual grounding in monocular RGB images using language descriptions with both appearance and geometry information. Specifically, we build a large-scale dataset, Mono3DRefer, which contains 3D object targets with their corresponding geometric text descriptions, generated by ChatGPT and refined manually. To foster this task, we propose Mono3DVG-TR, an end-to-end transformer-based network, which takes advantage of both the appearance and geometry information in text embeddings for multi-modal learning and 3D object localization. Depth predictor is designed to explicitly learn geometry features. The dual text-guided adapter is proposed to refine multiscale visual and geometry features of the referred object. Based on depth-text-visual stacking attention, the decoder fuses object-level geometric cues and visual appearance into a learnable query. Comprehensive benchmarks and some insightful analyses are provided for Mono3DVG. Extensive comparisons and ablation studies show that our method significantly outperforms all baselines. The dataset and code will be released.",
author = "Yang Zhan and Yuan Yuan and Zhitong Xiong",
note = "Publisher Copyright: Copyright {\textcopyright} 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.; 38th AAAI Conference on Artificial Intelligence, AAAI 2024 ; Conference date: 20-02-2024 Through 27-02-2024",
year = "2024",
month = mar,
day = "25",
doi = "10.1609/aaai.v38i7.28525",
language = "英语",
series = "Proceedings of the AAAI Conference on Artificial Intelligence",
publisher = "Association for the Advancement of Artificial Intelligence",
number = "7",
pages = "6988--6996",
editor = "Michael Wooldridge and Jennifer Dy and Sriraam Natarajan",
booktitle = "Technical Tracks 14",
edition = "7",
}