BladeNeRF: Exploiting camera constraints for NeRF in repetitive texture-less 3D reconstruction
N. El Coudi El Amrani (TU Delft - Electrical Engineering, Mathematics and Computer Science)
J.C. van Gemert – Mentor (TU Delft - Pattern Recognition and Bioinformatics)
Y. Lin – Mentor (TU Delft - Intelligent Vehicles)
M. Weinmann – Graduation committee member (TU Delft - Computer Graphics and Visualisation)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Neural Radiance Fields (NeRFs) have demonstrated remarkable capabilities in photo-realistic 3D reconstruction. NeRFs often take as input posed images where the camera poses come from either off-the-shelf S\textit{f}M or online optimization together with NeRFs. However, we find that both strategies yield suboptimal results in recovering camera poses from images when encountering texture-less and repetitive patterns, particularly in aircraft engine inspection. To reconstruct photo-realistic 3D engine blades from images, we propose BladeNeRF, a new variant of NeRF model that incorporates camera constraints into learning and enables accurate pose learning. In addition, we propose to separate the blades in the foreground from the constant background, eliminating background artefacts and enhancing depth estimation accuracy. Experimental evaluations on synthetic data demonstrate the advantage of our model in precise camera pose estimation and high-fidelity 3D scene reconstruction compared to other NeRF variants.