AI Navigate

VirPro: Visual-referred Probabilistic Prompt Learning for Weakly-Supervised Monocular 3D Detection

arXiv cs.CV / 3/19/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • VirPro introduces Visual-referred Probabilistic Prompt Learning, a multi-modal pretraining paradigm for weakly supervised monocular 3D detection that combines visual cues with learnable textual prompts.
  • The method uses an Adaptive Prompt Bank (APB) to store instance-conditioned prompts and Multi-Gaussian Prompt Modeling (MGPM) to inject scene-based visual features into textual embeddings, capturing visual uncertainty.
  • A RoI-level contrastive matching strategy is employed to align vision-language embeddings and tighten semantic coherence among co-occurring objects in the same scene.
  • Experiments on the KITTI benchmark show consistent performance gains, achieving up to 4.8% average precision improvement over baselines.
  • The work proposes a new direction for weakly supervised 3D detection by leveraging probabilistic, scene-aware prompts to better model visual diversity in real-world scenes.

Abstract

Monocular 3D object detection typically relies on pseudo-labeling techniques to reduce dependency on real-world annotations. Recent advances demonstrate that deterministic linguistic cues can serve as effective auxiliary weak supervision signals, providing complementary semantic context. However, hand-crafted textual descriptions struggle to capture the inherent visual diversity of individuals across scenes, limiting the model's ability to learn scene-aware representations. To address this challenge, we propose Visual-referred Probabilistic Prompt Learning (VirPro), an adaptive multi-modal pretraining paradigm that can be seamlessly integrated into diverse weakly supervised monocular 3D detection frameworks. Specifically, we generate a diverse set of learnable, instance-conditioned prompts across scenes and store them in an Adaptive Prompt Bank (APB). Subsequently, we introduce Multi-Gaussian Prompt Modeling (MGPM), which incorporates scene-based visual features into the corresponding textual embeddings, allowing the text prompts to express visual uncertainties. Then, from the fused vision-language embeddings, we decode a prompt-targeted Gaussian, from which we derive a unified object-level prompt embedding for each instance. RoI-level contrastive matching is employed to enforce modality alignment, bringing embeddings of co-occurring objects within the same scene closer in the latent space, thus enhancing semantic coherence. Extensive experiments on the KITTI benchmark demonstrate that integrating our pretraining paradigm consistently yields substantial performance gains, achieving up to a 4.8% average precision improvement than the baseline.