TY - JOUR
T1 - Rugularizing generalizable neural radiance field with limited-view images
AU - Sun, Wei
AU - Cui, Ruijia
AU - Wang, Qianzhou
AU - Kong, Xianguang
AU - Zhang, Yanning
N1 - Publisher Copyright:
© The Author(s) 2024.
PY - 2025/1
Y1 - 2025/1
N2 - We present a novel learning model with attention and prior guidance for view synthesis. In contrast to previous works that focus on optimizing for specific scenes with densely captured views, our model explores a generic deep neural framework to reconstruct radiance fields from a limited number of input views. To address challenges arising from under-constrained conditions, our approach employs cost volumes for geometry-aware scene reasoning, and integrates relevant knowledge from the ray-cast space and the surrounding-view space using an attention model. Additionally, a denoising diffusion model learns a prior over scene color, facilitating regularization of the training process and enabling high-quality radiance field reconstruction. Experimental results on diverse benchmark datasets demonstrate that our approach can generalize across scenes and produce realistic view synthesis results using only three input images, surpassing the performance of previous state-of-the-art methods. Moreover, our reconstructed radiance field can be effectively optimized by fine-tuning the target scene to achieve higher quality results with reduced optimization time. The code will be released at https://github.com/dsdefv/nerf.
AB - We present a novel learning model with attention and prior guidance for view synthesis. In contrast to previous works that focus on optimizing for specific scenes with densely captured views, our model explores a generic deep neural framework to reconstruct radiance fields from a limited number of input views. To address challenges arising from under-constrained conditions, our approach employs cost volumes for geometry-aware scene reasoning, and integrates relevant knowledge from the ray-cast space and the surrounding-view space using an attention model. Additionally, a denoising diffusion model learns a prior over scene color, facilitating regularization of the training process and enabling high-quality radiance field reconstruction. Experimental results on diverse benchmark datasets demonstrate that our approach can generalize across scenes and produce realistic view synthesis results using only three input images, surpassing the performance of previous state-of-the-art methods. Moreover, our reconstructed radiance field can be effectively optimized by fine-tuning the target scene to achieve higher quality results with reduced optimization time. The code will be released at https://github.com/dsdefv/nerf.
KW - Diffusion model
KW - Generalizable network
KW - Limited views
KW - Neural radiance field
UR - http://www.scopus.com/inward/record.url?scp=85212776119&partnerID=8YFLogxK
U2 - 10.1007/s40747-024-01696-6
DO - 10.1007/s40747-024-01696-6
M3 - 文章
AN - SCOPUS:85212776119
SN - 2199-4536
VL - 11
JO - Complex and Intelligent Systems
JF - Complex and Intelligent Systems
IS - 1
M1 - 78
ER -