AI Navigate

Pano360: Perspective to Panoramic Vision with Geometric Consistency

arXiv cs.CV / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Existing panorama stitching approaches rely on pairwise feature correspondences and struggle to enforce geometric consistency across multiple views, causing distortion in challenging scenes with weak texture or large parallax.
  • The paper extends 2D alignment to 3D photogrammetric space by leveraging 3D multi-view correspondences and a transformer-based architecture that is aware of camera poses to guide image warping for global 3D alignment.
  • It introduces a multi-feature joint optimization strategy for seam computation and builds a large-scale real-world dataset to train and evaluate the method.
  • Experiments show significant improvements in alignment accuracy and perceptual quality over existing methods, with potential benefits for VR/AR and wide-baseline panorama applications.

Abstract

Prior panorama stitching approaches heavily rely on pairwise feature correspondences and are unable to leverage geometric consistency across multiple views. This leads to severe distortion and misalignment, especially in challenging scenes with weak textures, large parallax, and repetitive patterns. Given that multi-view geometric correspondences can be directly constructed in 3D space, making them more accurate and globally consistent, we extend the 2D alignment task to the 3D photogrammetric space. We adopt a novel transformer-based architecture to achieve 3D awareness and aggregate global information across all views. It directly utilizes camera poses to guide image warping for global alignment in 3D space and employs a multi-feature joint optimization strategy to compute the seams. Additionally, to establish an evaluation benchmark and train our network, we constructed a large-scale dataset of real-world scenes. Extensive experiments show that our method significantly outperforms existing alternatives in alignment accuracy and perceptual quality.