AI Navigate

ABRA: Teleporting Fine-Tuned Knowledge Across Domains for Open-Vocabulary Object Detection

arXiv cs.CV / 3/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • ABRA presents a method that treats domain adaptation as a geometric transport problem in the weight space of a pretrained detector to move class-specific knowledge from a labeled source domain to a target domain lacking labeled examples for those classes.
  • The approach aligns source and target domain experts to transport class-specific knowledge, enabling fine-tuning-free adaptation under domain shifts such as nighttime or foggy conditions.
  • Experimental results across challenging domain shifts show ABRA can teleport class-level specialization for open-vocabulary object detection, improving robustness when data is scarce.
  • The authors plan to release code upon acceptance to facilitate replication and practical use.

Abstract

Although recent Open-Vocabulary Object Detection architectures, such as Grounding DINO, demonstrate strong zero-shot capabilities, their performance degrades significantly under domain shifts. Moreover, many domains of practical interest, such as nighttime or foggy scenes, lack large annotated datasets, preventing direct fine-tuning. In this paper, we introduce Aligned Basis Relocation for Adaptation(ABRA), a method that transfers class-specific detection knowledge from a labeled source domain to a target domain where no training images containing these classes are accessible. ABRA formulates this adaptation as a geometric transport problem in the weight space of a pretrained detector, aligning source and target domain experts to transport class-specific knowledge. Extensive experiments across challenging domain shifts demonstrate that ABRA successfully teleports class-level specialization under multiple adverse conditions. Our code will be made public upon acceptance.