Causal Bootstrapped Alignment for Unsupervised Video-Based Visible-Infrared Person Re-Identification

arXiv cs.CV / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper targets unsupervised visible–infrared person re-identification (VVI-ReID) from unlabeled video tracklets for all-day surveillance, aiming to avoid costly cross-modality annotations required by supervised methods.
  • It finds that naively extending image-based unsupervised VI-ReID approaches to video using generic pretrained encoders performs poorly due to weak identity discrimination and strong modality bias.
  • To fix these problems, the proposed Causal Bootstrapped Alignment (CBA) framework uses Causal Intervention Warm-up (CIW) to suppress spurious correlations from modality and motion while preserving identity-relevant semantics.
  • It further introduces Prototype-Guided Uncertainty Refinement (PGUR), a coarse-to-fine cross-modality alignment method that handles visible–infrared granularity mismatch using uncertainty-aware supervision guided by reliable visible prototypes.
  • Experiments on HITSZ-VCM and BUPTCampus show CBA substantially outperforms existing unsupervised VI-ReID methods when adapted to the unsupervised video VVI-ReID setting.

Abstract

VVI-ReID is a critical technique for all-day surveillance, where temporal information provides additional cues beyond static images. However, existing approaches rely heavily on fully supervised learning with expensive cross-modality annotations, limiting scalability. To address this issue, we investigate Unsupervised Learning for VVI-ReID (USL-VVI-ReID), which learns identity-discriminative representations directly from unlabeled video tracklets. Directly extending image-based USL-VI-ReID methods to this setting with generic pretrained encoders leads to suboptimal performance. Such encoders suffer from weak identity discrimination and strong modality bias, resulting in severe intra-modality identity confusion and pronounced clustering granularity imbalance between visible and infrared modalities. These issues jointly degrade pseudo-label reliability and hinder effective cross-modality alignment. To address these challenges, we propose a Causal Bootstrapped Alignment (CBA) framework that explicitly exploits inherent video priors. First, we introduce Causal Intervention Warm-up (CIW), which performs sequence-level causal interventions by leveraging temporal identity consistency and cross-modality identity consistency to suppress modality- and motion-induced spurious correlations while preserving identity-relevant semantics, yielding cleaner representations for unsupervised clustering. Second, we propose Prototype-Guided Uncertainty Refinement (PGUR), which employs a coarse-to-fine alignment strategy to resolve cross-modality granularity mismatch, reorganizing under-clustered infrared representations under the guidance of reliable visible prototypes with uncertainty-aware supervision. Extensive experiments on the HITSZ-VCM and BUPTCampus benchmarks demonstrate that CBA significantly outperforms existing USL-VI-ReID methods when extended to the USL-VVI-ReID setting.