AI Navigate

MXFP4による量子化の可能性の解明:量子化誤差削減のための戦略

arXiv cs.AI / 2026/3/11

Ideas & Deep AnalysisTools & Practical UsageModels & Research

要点

  • 本記事は、NVIDIAのNVFP4と比べた場合の大規模言語モデル推論におけるMXFP4 4ビット量子化フォーマットの精度制限に焦点を当てています。
  • ハードウェアの変更を必要とせず、MXFP4の量子化忠実度を向上させる2つのソフトウェアのみの手法、Overflow-Aware Scaling(OAS)とMacro Block Scaling(MBS)を紹介します。
  • これらの技術は有効なダイナミックレンジを拡大し、精度配分を改善することで、MXFP4とNVFP4間の精度差を約10%から平均で1%未満に縮小します。
  • これらの手法はわずかな計算オーバーヘッド(GEMMで6.2%の増加)を伴うのみで、テンサーコアでの12%の相対面積削減などMXFP4のハードウェア効率の利点は維持されます。
  • これらの改善により、MXFP4は精度とハードウェア効率のバランスを実現する大規模言語モデル推論向けの実用的かつ効率的な代替手段として位置づけられます。

Computer Science > Hardware Architecture

arXiv:2603.08713 (cs)
[Submitted on 30 Jan 2026]

Title:Unveiling the Potential of Quantization with MXFP4: Strategies for Quantization Error Reduction

View a PDF of the paper titled Unveiling the Potential of Quantization with MXFP4: Strategies for Quantization Error Reduction, by Jatin Chhugani and 10 other authors
View PDF HTML (experimental)
Abstract:Large Language Models (LLMs) have intensified the need for low-precision formats that enable efficient, large-scale inference. The Open Compute Project (OCP) Microscaling (MX) standard is attractive due to its favorable hardware efficiency, but its 4-bit variant (MXFP4) lags behind NVIDIA's NVFP4 in accuracy, limiting adoption. We introduce two software-only techniques, Overflow-Aware Scaling (OAS) and Macro Block Scaling (MBS), that improve MXFP4 quantization fidelity without requiring hardware changes. OAS reduces overall errors by increasing effective dynamic range under power-of-two block scaling, while MBS allocates higher-precision scaling at a coarser granularity to better preserve outliers. Across multiple LLMs and standard downstream benchmarks, OAS and MBS reduce the end-to-end accuracy gap between MXFP4 and NVFP4 from about 10% to below 1% on average, while incurring modest GEMM overhead (6.2% on average). These results re-establish MXFP4 as a practical alternative to NVFP4, enabling near-NVFP4 accuracy while retaining MX's hardware-efficiency advantages (e.g., 12% relative area savings in tensor cores).
Subjects: Hardware Architecture (cs.AR); Artificial Intelligence (cs.AI); Machine Learning (cs.LG); Performance (cs.PF)
Cite as: arXiv:2603.08713 [cs.AR]
  (or arXiv:2603.08713v1 [cs.AR] for this version)
  https://doi.org/10.48550/arXiv.2603.08713
Focus to learn more
arXiv-issued DOI via DataCite

Submission history

From: Geonhwa Jeong [view email]
[v1] Fri, 30 Jan 2026 23:24:17 UTC (977 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled Unveiling the Potential of Quantization with MXFP4: Strategies for Quantization Error Reduction, by Jatin Chhugani and 10 other authors
  • View PDF
  • HTML (experimental)
  • TeX Source
Current browse context:
cs.AR
< prev   |   next >
Change to browse by:

References & Citations

export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo
Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle
Bibliographic Explorer (What is the Explorer?)
Connected Papers Toggle
Connected Papers (What is Connected Papers?)
Litmaps Toggle
Litmaps (What is Litmaps?)
scite.ai Toggle
scite Smart Citations (What are Smart Citations?)
Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle
alphaXiv (What is alphaXiv?)
Links to Code Toggle
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub Toggle
DagsHub (What is DagsHub?)
GotitPub Toggle
Gotit.pub (What is GotitPub?)
Huggingface Toggle
Hugging Face (What is Huggingface?)
Links to Code Toggle
Papers with Code (What is Papers with Code?)
ScienceCast Toggle
ScienceCast (What is ScienceCast?)
Demos

Demos

Replicate Toggle
Replicate (What is Replicate?)
Spaces Toggle
Hugging Face Spaces (What is Spaces?)
Spaces Toggle
TXYZ.AI (What is TXYZ.AI?)
Related Papers

Recommenders and Search Tools

Link to Influence Flower
Influence Flower (What are Influence Flowers?)
Core recommender toggle
CORE Recommender (What is CORE?)
About arXivLabs

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.