COMPASS: COmpact Multi-channel Prior-map And Scene Signature for Floor-Plan-Based Visual Localization
arXiv cs.CV / 4/29/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- The paper introduces COMPASS, a visual localization algorithm that leverages both geometric and semantic priors from architectural floor plans rather than relying mainly on geometry.
- COMPASS builds a multi-channel, radial “scan-context”-inspired descriptor using 360 azimuth bins and encodes five channels: normalized range, structural hit type (wall/window/opening), range gradient, inverse range, and local range variance.
- On the vision side, the method populates the same descriptor structure by detecting structural elements in dual-fisheye images, enabling structural matching between the camera view and the floor-plan-derived descriptor.
- As a first step toward full cross-modal matching, the authors propose a fisheye window detection algorithm based on line segment detection, vertical edge clustering, and brightness verification.
- In a proof-of-concept using the Hilti-Trimble SLAM Challenge 2026 dataset, window/wall patterns from the first camera frames show close correspondence to the floor-plan descriptor, supporting the feasibility of cross-modal structural localization.
Related Articles

How I Use AI Agents to Maintain a Living Knowledge Base for My Team
Dev.to
IK_LLAMA now supports Qwen3.5 MTP Support :O
Reddit r/LocalLLaMA
OpenAI models, Codex, and Managed Agents come to AWS
Dev.to

Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

Automatic Error Recovery in AI Agent Networks
Dev.to