Adaptive Block-Scaled Data Types
arXiv cs.CL / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper identifies a key limitation of NVFP4 4-bit quantization: its error distribution can produce disproportionately large quantization errors near the maximum values within each 16-value block.
- It proposes Adaptive Block-Scaled Data Types, notably IF4, which chooses between FP4 and INT4 per 16-value group and uses an E4M3 scale factor (encoded via a sign bit) to better match the input distribution.
- The authors extend the same adaptive concept to other bit-widths, including IF3 and IF6, aiming to improve quantization behavior beyond fixed-format schemes.
- Experiments on language models show that IF4 reduces loss during quantized training and improves accuracy in post-training quantization compared with existing 4-bit block-scaled formats.
- To support deployability, the work designs and evaluates an IF4 Multiply-Accumulate (MAC) unit and provides code via the cited GitHub repository, suggesting efficient implementation in hardware accelerators.
Related Articles

Black Hat Asia
AI Business
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

Claude Code's Entire Source Code Was Just Leaked via npm Source Maps — Here's What's Inside
Dev.to

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to