Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
I&S-ViT introduces a novel method for inclusive and stable post-training quantization (PTQ) of Vision Transformers (ViTs). It addresses quantization inefficiency in post-Softmax activations with a Shift-Uniform-Log2 Quantizer (SULQ) and mitigates rugged loss landscapes in post-LayerNorm activations, significantly reducing performance drops in low-bit scenarios.
Enables the deployment of powerful Vision Transformer models on resource-constrained devices like mobile phones and edge hardware, reducing inference costs and latency for AI applications.