TY - JOUR
T1 - Generalizable 3D Gaussian Splatting for novel view synthesis
AU - Zhao, Chuyue
AU - Huang, Xin
AU - Yang, Kun
AU - Wang, Xue
AU - Wang, Qing
N1 - Publisher Copyright:
© 2024
PY - 2025/5
Y1 - 2025/5
N2 - We present a generalizable 3D Gaussian Splatting (3DGS) method that can synthesize novel views of unseen scenes. Existing methods directly input image features into the parameter regression network without establishing a connection to the 3D representation, leading to inaccurate parameter predictions and artifacts in the rendered views. To address this issue, our method integrates spatial information from multiple source views. Specifically, by leveraging multi-view feature mapping to bridge 2D features with 3D representations, our method directly align the Gaussians with image features. The well-aligned features provide guidance for the accurate prediction of Gaussian parameters, thereby enhancing the ability to represent unseen scenes and alleviating artifacts caused by feature sampling ambiguity. The proposed framework is fully differentiable and allows optimizing Gaussian parameters in a feed-forward manner. After training on a large dataset of real-world scenes, our method enables novel view synthesis of unseen scenes without the need for optimization. Experimental results on real-world datasets demonstrate that our method outperforms recent novel view synthesis methods that also seek to generalize to unseen scenes.
AB - We present a generalizable 3D Gaussian Splatting (3DGS) method that can synthesize novel views of unseen scenes. Existing methods directly input image features into the parameter regression network without establishing a connection to the 3D representation, leading to inaccurate parameter predictions and artifacts in the rendered views. To address this issue, our method integrates spatial information from multiple source views. Specifically, by leveraging multi-view feature mapping to bridge 2D features with 3D representations, our method directly align the Gaussians with image features. The well-aligned features provide guidance for the accurate prediction of Gaussian parameters, thereby enhancing the ability to represent unseen scenes and alleviating artifacts caused by feature sampling ambiguity. The proposed framework is fully differentiable and allows optimizing Gaussian parameters in a feed-forward manner. After training on a large dataset of real-world scenes, our method enables novel view synthesis of unseen scenes without the need for optimization. Experimental results on real-world datasets demonstrate that our method outperforms recent novel view synthesis methods that also seek to generalize to unseen scenes.
KW - 3D Gaussian Splatting
KW - Generalizable scene representation
KW - Image-based rendering
KW - Novel view synthesis
UR - http://www.scopus.com/inward/record.url?scp=85212130948&partnerID=8YFLogxK
U2 - 10.1016/j.patcog.2024.111271
DO - 10.1016/j.patcog.2024.111271
M3 - 文章
AN - SCOPUS:85212130948
SN - 0031-3203
VL - 161
JO - Pattern Recognition
JF - Pattern Recognition
M1 - 111271
ER -