GPA-VGGT:Adapting VGGT to Large Scale Localization by Self-Supervised Learning with Geometry and Physics Aware Loss

arXiv cs.RO / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces GPA-VGGT, a self-supervised training framework for the Visual Geometry Grounded Transformer (VGGT) to improve camera localization in large-scale, unlabeled environments.
  • It replaces hard-label supervision by extending pair-wise geometric relations to sequence-wise geometric constraints, sampling multiple source frames and projecting them onto target frames to enforce temporal feature consistency.
  • The method uses a joint optimization loss that combines physical photometric consistency with geometric constraints, enabling learning of multi-view geometry without ground truth labels.
  • Experiments report fast convergence (within hundreds of iterations) and significant gains in large-scale localization, including improvements to cross-view attention layers as well as camera and depth prediction heads.
  • The authors state they will release the code on GitHub, supporting reproducibility and further research use.

Abstract

Transformer-based general visual geometry frameworks have shown promising performance in camera pose estimation and 3D scene understanding. Recent advancements in Visual Geometry Grounded Transformer (VGGT) models have shown great promise in camera pose estimation and 3D reconstruction. However, these models typically rely on ground truth labels for training, posing challenges when adapting to unlabeled and unseen scenes. In this paper, we propose a self-supervised framework to train VGGT with unlabeled data, thereby enhancing its localization capability in large-scale environments. To achieve this, we extend conventional pair-wise relations to sequence-wise geometric constraints for self-supervised learning. Specifically, in each sequence, we sample multiple source frames and geometrically project them onto different target frames, which improves temporal feature consistency. We formulate physical photometric consistency and geometric constraints as a joint optimization loss to circumvent the requirement for hard labels. By training the model with this proposed method, not only the local and global cross-view attention layers but also the camera and depth heads can effectively capture the underlying multi-view geometry. Experiments demonstrate that the model converges within hundreds of iterations and achieves significant improvements in large-scale localization. Our code will be released at https://github.com/X-yangfan/GPA-VGGT.