Indoor 3D Reconstruction from a Single Image

More Info
expand_more

Abstract

3D indoor reconstruction has been an important research area in the field of computer vision and photogrammetry. While the initial techniques developed for this purpose use sensor devices and multiple images for data acquisition and extracting 3D information and representation of the scene, with the advent of deep learning techniques, there has been good progress in extracting 3D information of an indoor scene reconstruction using a single image. This has potential in minimizing user efforts and cost for data acquisition. The current state-of-the-art method involves two main components, the global depth map and plane instances. After investigating the current state-of-the-art methods, it is observed that there is inconsistency in reconstructed surface boundaries and depth estimation over the curvature and edges of the objects present in the scene, despite having good 3D representation in the surrounding regions. We devise a loss function for optimizing depth estimation during supervision of the neural network by providing geometric awareness to the pixels at local level based on its neighborhood properties defined by spatial compatibility and color similarity. A similar function is used during 3D reconstruction for orientation consistency in the point cloud. Based on the quantitative and qualitative analysis, it is observed that the proposed approach helps in improving the 3D reconstruction from a single image in an indoor environment.