AI Navigate

Continual Learning with Vision-Language Models via Semantic-Geometry Preservation

arXiv cs.CV / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies semantic geometry drift as a key challenge in continual learning for vision-language models and proposes an exemplar-free method to address it.
  • It introduces Semantic Geometry Preservation for Continual Learning (SeGP-CL), which constructs a compact set of adversarial anchors using dual-targeted projected gradient descent to steer new-task seeds toward old-class semantics while staying faithful in raw visual space.
  • Training with SeGP-CL combines anchor-guided cross-modal geometry distillation (ACGD) to preserve cross-modal structure and a lightweight text semantic-geometry regularization (TSGR) to stabilize the textual reference frame.
  • Experiments on five continual learning benchmarks demonstrate improved stability and forward transfer, achieving state-of-the-art results while better preserving the semantic geometry of vision-language models.

Abstract

Continual learning of pretrained vision-language models (VLMs) is prone to catastrophic forgetting, yet current approaches adapt to new tasks without explicitly preserving the cross-modal semantic geometry inherited from pretraining and previous stages, allowing new-task supervision to induce geometric distortion. We observe that the most pronounced drift tends to concentrate in vulnerable neighborhoods near the old-new semantic interface, where shared visual patterns are easily re-explained by new textual semantics. To address this under an exemplar-free constraint, we propose Semantic Geometry Preservation for Continual Learning (SeGP-CL). SeGP-CL first probes the drift-prone region by constructing a compact set of adversarial anchors with dual-targeted projected gradient descent (DPGD), which drives selected new-task seeds toward old-class semantics while remaining faithful in raw visual space. During training, we preserve cross-modal structure by anchor-guided cross-modal geometry distillation (ACGD), and stabilize the textual reference frame across tasks via a lightweight text semantic-geometry regularization (TSGR). After training, we estimate anchor-induced raw-space drift to transfer old visual prototypes and perform dual-path inference by fusing cross-modal and visual cues. Extensive experiments on five continual learning benchmarks demonstrate that SeGP-CL consistently improves stability and forward transfer, achieving state-of-the-art performance while better preserving semantic geometry of VLMs.