Neural Radiance Field (NeRF) as a Rendering Primitive

StreamNeRF - Adapting a NeRF Model for Progressive Decoding

Bachelor Thesis (2023)
Author(s)

M. Găleşanu (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

E. Eisemann – Mentor (TU Delft - Computer Graphics and Visualisation)

Petr Kellnhofer – Mentor (TU Delft - Computer Graphics and Visualisation)

Michael Weinmann – Mentor (TU Delft - Computer Graphics and Visualisation)

J.C. Gemert – Graduation committee member (TU Delft - Pattern Recognition and Bioinformatics)

Faculty
Electrical Engineering, Mathematics and Computer Science
Copyright
© 2023 Matei Găleşanu
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 Matei Găleşanu
Graduation Date
05-07-2023
Awarding Institution
Delft University of Technology
Project
['CSE3000 Research Project']
Programme
['Computer Science and Engineering']
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Neural Radiance Fields (NeRF) and their adaptations are known to be computationally intensive during both the training and the evaluating stages. Despite being the end goal, directly rendering a full-resolution representation of the scene is not necessary and not very practical for scenarios like streamed applications. Our goal is to design a streamable adaptation for a model that can produce fast, rough estimates of 3D scenes, by only using a shallow part of the network. The quality is subsequently improved as more parts of the network are available, such that it can be used in online applications where the model needs to be transferred. Separate models can be trained at different resolutions, but this approach results in a large space overhead and also increases the evaluation time. This can be mitigated by reducing the depth of low-resolution models, but redundancy will still be high as each new model needs to re-evaluate the input data, rendering previous calculations obsolete. Our method combines key concepts from previous approaches to create a progressively trained model that is able to produce intermediate outputs of increasing quality while attempting to optimize the trade-off between overhead and quality. Our model is able to produce a recognizable representation of the scene with as little as one hidden layer from the original model. It also allows for division into streamable chunks which can be sent individually and, upon reconstruction, provide intermediate outputs that bring consistent improvement in quality. The newly streamed data uses the residual output from previous computations in order to reduce redundancy. We show that the final quality of our adaptation is within 2% of the original in terms of previously used quantitative metrics.

Files

Final_paper_mgalesanu.pdf
(pdf | 14.8 Mb)
License info not available