SynQ: Accurate Zero-shot Quantization by Synthesis-aware Fine-tuning
arXiv cs.CV / 3/20/2026
📰 NewsModels & Research
Key Points
- SynQ is a synthesis-aware fine-tuning framework for zero-shot quantization that quantizes pre-trained models without access to training data.
- It overcomes three main ZSQ challenges by using a low-pass filter to reduce noise in synthetic samples, aligning the quantized model's class activation maps with the pre-trained model, and using soft labels for hard samples to mitigate misguidance from erroneous labels.
- Extensive experiments show that SynQ achieves state-of-the-art accuracy compared with existing ZSQ methods.
- By enabling accurate quantization without data, SynQ facilitates deploying compressed models on privacy- or security-constrained edge devices.
Related Articles
Data Augmentation Using GANs
Dev.to
ItinBench: Benchmarking Planning Across Multiple Cognitive Dimensions with Large Language Models
arXiv cs.AI
Hyperagents
arXiv cs.AI
Teaching an Agent to Sketch One Part at a Time
arXiv cs.AI
PowerLens: Taming LLM Agents for Safe and Personalized Mobile Power Management
arXiv cs.AI