TY - GEN
T1 - MVS2
T2 - 7th International Conference on 3D Vision, 3DV 2019
AU - Dai, Yuchao
AU - Zhu, Zhidong
AU - Rao, Zhibo
AU - Li, Bo
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/9
Y1 - 2019/9
N2 - The success of existing deep-learning based multi-view stereo (MVS) approaches greatly depends on the availability of large-scale supervision in the form of dense depth maps. Such supervision, while not always possible, tends to hinder the generalization ability of the learned models in never-seen-before scenarios. In this paper, we propose the first unsupervised learning based MVS network, which learns the multi-view depth maps from the input multi-view images and does not need ground-truth 3D training data. Our network is symmetric in predicting depth maps for all views simultaneously, where we enforce cross-view consistency of multi-view depth maps during both training and testing stages. Thus, the learned multi-view depth maps naturally comply with the underlying 3D scene geometry. Besides, our network also learns the multi-view occlusion maps, which further improves the robustness of our network in handling real-world occlusions. Experimental results on multiple benchmarking datasets demonstrate the effectiveness of our network and the excellent generalization ability.
AB - The success of existing deep-learning based multi-view stereo (MVS) approaches greatly depends on the availability of large-scale supervision in the form of dense depth maps. Such supervision, while not always possible, tends to hinder the generalization ability of the learned models in never-seen-before scenarios. In this paper, we propose the first unsupervised learning based MVS network, which learns the multi-view depth maps from the input multi-view images and does not need ground-truth 3D training data. Our network is symmetric in predicting depth maps for all views simultaneously, where we enforce cross-view consistency of multi-view depth maps during both training and testing stages. Thus, the learned multi-view depth maps naturally comply with the underlying 3D scene geometry. Besides, our network also learns the multi-view occlusion maps, which further improves the robustness of our network in handling real-world occlusions. Experimental results on multiple benchmarking datasets demonstrate the effectiveness of our network and the excellent generalization ability.
KW - multi view stereo
KW - multi view symmetry
KW - unsupervised learning
UR - http://www.scopus.com/inward/record.url?scp=85075013801&partnerID=8YFLogxK
U2 - 10.1109/3DV.2019.00010
DO - 10.1109/3DV.2019.00010
M3 - 会议稿件
AN - SCOPUS:85075013801
T3 - Proceedings - 2019 International Conference on 3D Vision, 3DV 2019
SP - 1
EP - 8
BT - Proceedings - 2019 International Conference on 3D Vision, 3DV 2019
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 15 September 2019 through 18 September 2019
ER -