Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published August 2024 | Published
Journal Article Open

NeRF-VO: Real-Time Sparse Visual Odometry With Neural Radiance Fields

Abstract

We introduce a novel monocular visual odometry (VO) system, NeRF-VO, that integrates learning-based sparse visual odometry for low-latency camera tracking and a neural radiance scene representation for fine-detailed dense reconstruction and novel view synthesis. Our system initializes camera poses using sparse visual odometry and obtains view-dependent dense geometry priors from a monocular prediction network. We harmonize the scale of poses and dense geometry, treating them as supervisory cues to train a neural implicit scene representation. NeRF-VO demonstrates exceptional performance in both photometric and geometric fidelity of the scene representation by jointly optimizing a sliding window of keyframed poses and the underlying dense geometry, which is accomplished through training the radiance field with volume rendering. We surpass SOTA methods in pose estimation accuracy, novel view synthesis fidelity, and dense reconstruction quality across a variety of synthetic and real-world datasets while achieving a higher camera tracking frequency and consuming less GPU memory.

Data Availability

The supplemented video showcases the online dense reconstruction process of NeRF-VO on both public and our self-collected dataset. The dense geometry is iteratively refined with the back-end optimization in NeRF-VO. It also provides an overview of the framework of NeRF-VO.

Files

Files (44.5 MB)
Name Size Download all
md5:85adb54ec8bd9470c414c6358b300409
44.5 MB Download

Additional details

Created:
July 10, 2024
Modified:
July 10, 2024