Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: With the growing computational demands of large language models (LLMs),
efficient inference has become increasingly critical for practical deployment.
Depth pruning has emerged as a promising approach for reducing the
computational costs of large language models by removing transformer layers.
However, existing methods typically rely on fixed block masks, which can lead
to suboptimal performance across different tasks and inputs. In this paper, we
propose IG-Pruning, a novel input-aware block-wise pruning method that
dynamically selects layer masks at inference time. Our approach consists of two
stages: (1) Discovering diverse mask candidates through semantic clustering and
L0 optimization, and (2) Implementing efficient dynamic pruning without the
need for extensive training. Experimental results demonstrate that our method
consistently outperforms state-of-the-art static depth pruning methods, making
it particularly suitable for resource-constrained deployment scenarios.
Key Contributions
This paper proposes IG-Pruning, a novel input-aware block-wise pruning method for LLMs that dynamically selects layer masks at inference time. Unlike static pruning, IG-Pruning discovers diverse mask candidates through semantic clustering and L0 optimization, enabling efficient dynamic pruning without extensive retraining. This approach consistently outperforms static methods, making it highly suitable for resource-constrained deployment scenarios.
Business Value
Enables the deployment of powerful LLMs on devices with limited computational resources (e.g., smartphones, edge devices), opening up new application possibilities and reducing cloud dependency.