When Flat Minima Fail: Characterizing INT4 Quantization Collapse After FP32 Convergence
arXiv cs.LG / 4/17/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The study challenges the common PTQ assumption that a well-converged FP32 model is automatically “quantization-ready” by showing INT4 can catastrophically collapse even after FP32 convergence.
- Using a calibration-free per-group INT4 probe across 154 public Pythia-160m checkpoints, the authors identify a three-phase behavior: rapid improvement, a ~70k-step metastable plateau, and an explosive INT4 divergence where the INT4 gap grows from 11% to 517% while FP32 perplexity barely changes.
- The onset of divergence aligns with a finer-grained FP32 perplexity convergence trigger—suggesting post-convergence weight updates are the proximate cause rather than learning-rate decay magnitude alone.
- The phenomenon is specific to the coarseness of the 16-level INT4 grid: INT8 is immune across all phases, and weight outlier accumulation is ruled out via direct kurtosis measurements.
- Controlled fork experiments with different learning-rate schedules show SGDR accelerates divergence in all runs, while the proposed Oscillatory Lock-In (with settled “cool” phases) reduces the INT4 gap by 2.2 percentage points on average, indicating that amplitude calibration—not oscillation itself—determines whether perturbations help or hurt.

![[Patterns] AI Agent Error Handling That Actually Works](/_next/image?url=https%3A%2F%2Fmedia2.dev.to%2Fdynamic%2Fimage%2Fwidth%3D1200%2Cheight%3D627%2Cfit%3Dcover%2Cgravity%3Dauto%2Cformat%3Dauto%2Fhttps%253A%252F%252Fdev-to-uploads.s3.amazonaws.com%252Fuploads%252Farticles%252Frn5czaopq2vzo7cglady.png&w=3840&q=75)


![[2026] OpenTelemetry for LLM Observability — Self-Hosted Setup](/_next/image?url=https%3A%2F%2Fmedia2.dev.to%2Fdynamic%2Fimage%2Fwidth%3D1200%2Cheight%3D627%2Cfit%3Dcover%2Cgravity%3Dauto%2Cformat%3Dauto%2Fhttps%253A%252F%252Fdev-to-uploads.s3.amazonaws.com%252Fuploads%252Farticles%252Flu4b6ttuhur71z5gemm0.png&w=3840&q=75)