Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper AI Hardware Engineers,ML Researchers,Deep Learning Engineers,Nvidia Engineers 1 week ago

INT v.s. FP: A Comprehensive Study of Fine-Grained Low-bit Quantization Formats

large-language-models › model-architecture
📄 Abstract

Abstract: Modern AI hardware, such as Nvidia's Blackwell architecture, is increasingly embracing low-precision floating-point (FP) formats to handle the pervasive activation outliers in Large Language Models (LLMs). Despite this industry trend, a unified comparison of FP and integer (INT) quantization across varying granularities has been missing, leaving algorithm and hardware co-design without clear guidance. This paper fills that gap by systematically investigating the trade-offs between FP and INT formats. We reveal a critical performance crossover: while FP excels in coarse-grained quantization, the comparison at fine-grained (block-wise) levels is more nuanced. Our comprehensive comparison demonstrates that for popular 8-bit fine-grained formats (e.g., MX with block size 32), MXINT8 is superior to its FP counterpart in both algorithmic accuracy and hardware efficiency. However, for 4-bit formats, FP (e.g., MXFP4, NVFP4) often holds an accuracy advantage , though we show that NVINT4 can surpass NVFP4 when outlier-mitigation techniques like Hadamard rotation are applied. We also introduce a symmetric clipping method that resolves gradient bias in fine-grained low-bit INT training, enabling nearly lossless performance for MXINT8 training. These findings challenge the current hardware trajectory, demonstrating that a one-size-fits-all FP approach is suboptimal and advocating that fine-grained INT formats, particularly MXINT8, offer a better balance of accuracy, power, and efficiency for future AI accelerators.
Authors (13)
Mengzhao Chen
Meng Wu
Hui Jin
Zhihang Yuan
Jing Liu
Chaoyi Zhang
+7 more
Submitted
October 29, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper provides a comprehensive, unified comparison of fine-grained integer (INT) and floating-point (FP) quantization formats for LLMs. It reveals a performance crossover point and demonstrates that MXINT8 is superior to its FP counterpart for 8-bit fine-grained quantization, while FP formats often hold an advantage for 4-bit quantization.

Business Value

Provides crucial guidance for hardware and software developers to optimize LLM deployment by selecting the most efficient quantization formats, leading to reduced costs and faster inference.