UniCorrn: Unified Correspondence Transformer Across 2D and 3D

arXiv cs.CV / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces UniCorrn, a new correspondence-matching model that uses shared weights to unify geometric matching across 2D-2D, 2D-3D, and 3D-3D tasks.
  • It argues that Transformer attention can naturally capture cross-modal feature similarity and uses a dual-stream decoder to separately preserve appearance and positional features.
  • UniCorrn uses modality-specific backbones followed by shared encoder/decoder components, enabling end-to-end training with stackable layers and query-based correspondence estimation across heterogeneous modalities.
  • Trained jointly on diverse data (including pseudo point clouds from depth maps plus real 3D correspondence annotations), the model delivers competitive results for 2D-2D matching.
  • It reports state-of-the-art improvements of 8% on 7Scenes for 2D-3D and 10% on 3DLoMatch for 3D-3D in registration recall, outperforming prior methods.

Abstract

Visual correspondence across image-to-image (2D-2D), image-to-point cloud (2D-3D), and point cloud-to-point cloud (3D-3D) geometric matching forms the foundation for numerous 3D vision tasks. Despite sharing a similar problem structure, current methods use task-specific designs with separate models for each modality combination. We present UniCorrn, the first correspondence model with shared weights that unifies geometric matching across all three tasks. Our key insight is that Transformer attention naturally captures cross-modal feature similarity. We propose a dual-stream decoder that maintains separate appearance and positional feature streams. This design enables end-to-end learning through stack-able layers while supporting flexible query-based correspondence estimation across heterogeneous modalities. Our architecture employs modality-specific backbones followed by shared encoder and decoder components, trained jointly on diverse data combining pseudo point clouds from depth maps with real 3D correspondence annotations. UniCorrn achieves competitive performance on 2D-2D matching and surpasses prior state-of-the-art by 8% on 7Scenes (2D-3D) and 10% on 3DLoMatch (3D-3D) in registration recall. Project website: https://neu-vi.github.io/UniCorrn