CodeQuant: Unified Clustering and Quantization for Enhanced Outlier Smoothing in Low-Precision Mixture-of-Experts

arXiv cs.LG / 4/14/2026

💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper identifies outliers as a key bottleneck for maintaining accuracy when applying low-precision post-training quantization (PTQ) to Mixture-of-Experts (MoE) large models.
  • It proposes CodeQuant, a unified approach that combines learnable rotation to smooth activation outliers and clustering to absorb weight outliers into fine-tuned cluster centroids.
  • By reducing the influence of extreme values while preserving model expressiveness, CodeQuant lowers quantization error compared with prior smoothing/quantization methods.
  • The method includes dedicated kernels for GPU and CPU, achieving up to 4.15× speedups and improved accuracy across multiple MoE model variants.
  • The authors provide an open-source implementation on GitHub, positioning CodeQuant as a practical direction for deploying MoE LLMs under low-precision constraints.

Abstract

Outliers have emerged as a fundamental bottleneck in preserving accuracy for low-precision large models, particularly within Mixture-of-Experts (MoE) architectures that are increasingly central to large-scale language modeling. Under post-training quantization (PTQ), these outliers induce substantial quantization errors, leading to severe accuracy degradation. While recent rotation-based smoothing techniques alleviate the problem by redistributing outlier magnitudes, residual errors remain and continue to impede reliable low-precision deployment. In this work, we tackle this challenge by introducing \textit{CodeQuant}, a unified quantization-and-clustering scheme that contains smoothing activation outliers via learnable rotation and absorbing weight outliers into fine-tuned cluster centroids for MoE. This design reduces the influence of extreme values by fitting them within cluster centroids, thereby lowering quantization error while maintaining expressive capacity. Coupled with a dedicated kernel design for GPU and CPU, CodeQuant achieves up to 4.15\times speedup while delivering significantly higher accuracy than state-of-the-art quantization approaches across diverse MoE models. Our results highlight CodeQuant as a promising direction for efficient and accurate deployment of MoE-based large language models under low-precision constraints. Our code is available at https://github.com/SAI-Lab-NYU/CodeQuant.