BEV-SLAM: Building a Globally-Consistent World Map Using Monocular Vision | IEEE Conference Publication | IEEE Xplore

BEV-SLAM: Building a Globally-Consistent World Map Using Monocular Vision


Abstract:

The ability to produce large-scale maps for nav-igation, path planning and other tasks is a crucial step for autonomous agents, but has always been challenging. In this w...Show More

Abstract:

The ability to produce large-scale maps for nav-igation, path planning and other tasks is a crucial step for autonomous agents, but has always been challenging. In this work, we introduce BEV-SLAM, a novel type of graph-based SLAM that aligns semantically-segmented Bird's Eye View (BEV) predictions from monocular cameras. We introduce a novel form of occlusion reasoning into BEV estimation and demonstrate its importance to aid spatial aggregation of BEV predictions. The result is a versatile SLAM system that can operate across arbitrary multi-camera configurations and can be seamlessly integrated with other sensors. We show that the use of multiple cameras significantly increases performance, and achieves lower relative error than high-performance GPS. The resulting system is able to create large, dense, globally-consistent world maps from monocular cameras mounted around an ego vehicle. The maps are metric and correctly-scaled, making them suitable for downstream navigation tasks.
Date of Conference: 23-27 October 2022
Date Added to IEEE Xplore: 26 December 2022
ISBN Information:

ISSN Information:

Conference Location: Kyoto, Japan

I. Introduction

Mobile autonomous agents require an information-rich representation of their environment for navigation, planning and localisation. Typically, a top-down orthographic projection (also known as a Bird's Eye View (BEV) map) is preferred, since planning tasks are simplified in orthographic space.

Contact IEEE to Subscribe

References

References is not available for this document.