Zero-Shot Quantization via Weight-Space Arithmetic

arXiv cs.CV / 4/7/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces “quantization vectors,” a transferable direction in weight space that can be extracted from a donor task using simple weight-space arithmetic.
  • Applying this vector to a receiver model can significantly improve robustness to post-training quantization (PTQ)-induced noise—reported gains are up to ~60%—without requiring receiver-side QAT.
  • The approach is presented as zero-shot and low-cost because it does not need receiver training data.
  • Experiments are demonstrated on Vision Transformer (ViT) models, supporting the claim that quantization robustness can be a reusable property of weight-space geometry.
  • Overall, the results position quantization robustness as transferable across tasks rather than something that must be relearned via task-specific training.

Abstract

We show that robustness to post-training quantization (PTQ) is a transferable direction in weight space. We call this direction the quantization vector: extracted from a donor task by simple weight-space arithmetic, it can be used to patch a receiver model and improve robustness to PTQ-induced noise by as much as 60%, without receiver-side quantization-aware training (QAT). Because the method requires no receiver training data, it provides a zero-shot, low-cost alternative to QAT for extremely low-bit deployment. We demonstrate this on Vision Transformer (ViT) models. More broadly, our results suggest that quantization robustness is not merely a byproduct of task-specific training, but a reusable feature of weight-space geometry that can be transferred rather than retrained.