Colinearity Decay: Training Quantization-Friendly ViTs with Outlier Decay
arXiv cs.CV / 5/5/2026
📰 NewsTools & Practical UsageModels & Research
Key Points
- The paper targets the challenge of quantizing vision Transformers to low bit-width, where activation outliers can degrade fully quantized deployment accuracy.
- Instead of simply suppressing outliers or using post-training quantization, it proposes Colinearity-Decay (CD), a training-time structural regularizer that penalizes harmful cross-matrix alignment inside Transformer blocks.
- CD is designed to be non-invasive: it does not change the model architecture or task loss, and it adds minimal training overhead when applied as a decoupled update.
- Experiments across ImageNet-1K pre-training, COCO detection, and downstream fine-tuning show consistent improvements in quantized accuracy while preserving (or improving) full-precision performance.
- The authors conclude that structural regularization can effectively “prepare” Vision Transformers for low-bit deployment with zero additional inference-time cost.
Related Articles

Black Hat USA
AI Business

Singapore's Fraud Frontier: Why AI Scam Detection Demands Regulatory Precision
Dev.to

First experience with Building Apps with Google AI Studio: Incredibly simple and intuitive.
Dev.to

How AI is Changing the Way We Code in 2026: The Shift from Syntax to Strategy
Dev.to

13 CLAUDE.md Rules That Make AI Write Modern PHP (Not PHP 5 Resurrected)
Dev.to