The Geometry of Robustness: Optimizing Loss Landscape Curvature and Feature Manifold Alignment for Robust Finetuning of Vision-Language Models

arXiv cs.CV / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that robust fine-tuning of vision-language models fails to balance ID accuracy, OOD generalization, and adversarial robustness because of two geometric issues: sharp/anistropic minima and perturbation-sensitive feature representations.
  • It introduces GRACE (Gram-aligned Robustness via Adaptive Curvature Estimation), a unified fine-tuning framework that regularizes parameter-space curvature to encourage flatter minima while enforcing feature-space invariance across clean, adversarial, and OOD inputs.
  • GRACE uses adaptive weight perturbations scaled by locally estimated curvature and combines this with a feature alignment loss, motivated by Robust PAC-Bayes theory.
  • Experiments on ImageNet fine-tuning of CLIP show simultaneous gains: +10.8% ID accuracy and +13.5% adversarial accuracy, with OOD accuracy staying essentially unchanged (57.0% vs 57.4% zero-shot baseline).
  • Additional geometric analysis claims GRACE converges to flatter minima and avoids feature distortion under distribution shifts, aiming for generalized robustness in foundation VLMs.

Abstract

Fine-tuning approaches for Vision-Language Models (VLMs) face a critical three-way trade-off between In-Distribution (ID) accuracy, Out-of-Distribution (OOD) generalization, and adversarial robustness. Existing robust fine-tuning strategies resolve at most two axes of this trade-off. Generalization-preserving methods retain ID/OOD performance but leave models vulnerable to adversarial attacks, while adversarial training improves robustness to targeted attacks but degrades ID/OOD accuracy. Our key insight is that the robustness trade-off stems from two geometric failures: sharp, anisotropic minima in parameter space and unstable feature representations that deform under perturbation. To address this, we propose GRACE (Gram-aligned Robustness via Adaptive Curvature Estimation), a unified fine-tuning framework that jointly regularizes the parameter-space curvature and feature-space invariance for VLMs. Grounded in Robust PAC-Bayes theory, GRACE employs adaptive weight perturbations scaled by local curvature to promote flatter minima, combined with a feature alignment loss that maintains representation consistency across clean, adversarial, and OOD inputs. On ImageNet fine-tuning of CLIP models, GRACE simultaneously improves ID accuracy by 10.8%, and adversarial accuracy by 13.5% while maintaining 57.0% OOD accuracy (vs. 57.4% zero-shot baseline). Geometric analysis confirms that GRACE converges to flatter minima without feature distortion across distribution shifts, providing a principled step toward generalized robustness in foundation VLMs.