AI Navigate

FlatLands: Generative Floormap Completion From a Single Egocentric View

arXiv cs.CV / 3/18/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • FlatLands introduces a dataset and benchmark for single-view BEV floor completion, aggregating 270,575 observations from 17,656 real indoor scenes with aligned observations, visibility, validity, and ground-truth BEV maps.
  • The benchmark provides both in- and out-of-distribution evaluation protocols and benchmarks a range of modeling approaches from training-free methods to deterministic models, ensembles, and stochastic generative models.
  • The work demonstrates an end-to-end monocular RGB-to-floormaps pipeline, enabling uncertainty-aware indoor mapping for embodied navigation.
  • FlatLands establishes a rigorous testbed for uncertainty-aware indoor mapping and generative completion, with potential impact on navigation systems operating under perceptual uncertainty.

Abstract

A single egocentric image typically captures only a small portion of the floor, yet a complete metric traversability map of the surroundings would better serve applications such as indoor navigation. We introduce FlatLands, a dataset and benchmark for single-view bird's-eye view (BEV) floor completion. The dataset contains 270,575 observations from 17,656 real metric indoor scenes drawn from six existing datasets, with aligned observation, visibility, validity, and ground-truth BEV maps, and the benchmark includes both in- and out-of-distribution evaluation protocols. We compare training-free approaches, deterministic models, ensembles, and stochastic generative models. Finally, we instantiate the task as an end-to-end monocular RGB-to-floormaps pipeline. FlatLands provides a rigorous testbed for uncertainty-aware indoor mapping and generative completion for embodied navigation.