dc.description.abstracten |
In recent years, the field of 3D scene reconstruction has witnessed significant ad-
vancements, fueled by growing interest in applications ranging from augmented
reality to autonomous navigation. A key component of this progress has been the
development of Neural Radiance Fields (NeRF), which have revolutionized the way
we render and interact with 3D environments. Despite these advancements, the
process of camera pose estimation remains a bottleneck, often requiring extensive
computational resources and time. This thesis introduces an innovative approach
that leverages 3D Gaussian Splatting, a technique that provides a more explicit rep-
resentation during both the rendering and training phases, enhancing the efficiency
and clarity of 3D reconstructions. Specifically, we focus on a method that utilizes es-
timated monocular depth maps to recover camera poses, which are then used to re-
construct the 3D scene. This methodology not only simplifies the traditional pipeline
by obviating the need for direct pose estimation but also improves the speed of the
reconstruction process. We evaluate our approach using both synthetic and real-
world datasets, in order to see it performance in different scenarios. |
uk |