Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Network pruning is a commonly used measure to alleviate the storage and
computational burden of deep neural networks. However, the fundamental limit of
network pruning is still lacking. To close the gap, in this work we'll take a
first-principles approach, i.e. we'll directly impose the sparsity constraint
on the loss function and leverage the framework of statistical dimension in
convex geometry, thus enabling us to characterize the sharp phase transition
point, which can be regarded as the fundamental limit of the pruning ratio.
Through this limit, we're able to identify two key factors that determine the
pruning ratio limit, namely, weight magnitude and network sharpness. Generally
speaking, the flatter the loss landscape or the smaller the weight magnitude,
the smaller pruning ratio. Moreover, we provide efficient countermeasures to
address the challenges in the computation of the pruning limit, which mainly
involves the accurate spectrum estimation of a large-scale and non-positive
Hessian matrix. Moreover, through the lens of the pruning ratio threshold, we
can also provide rigorous interpretations on several heuristics in existing
pruning algorithms. Extensive experiments are performed which demonstrate that
our theoretical pruning ratio threshold coincides very well with the
experiments. All codes are available at:
https://github.com/QiaozheZhang/Global-One-shot-Pruning
Authors (4)
Qiaozhe Zhang
Ruijie Zhang
Jun Sun
Yingzhuang Liu
Key Contributions
This paper investigates the fundamental limits of network pruning in deep neural networks by imposing sparsity constraints directly on the loss function and using statistical dimension from convex geometry. It identifies key factors (weight magnitude, network sharpness) determining the pruning ratio limit and proposes efficient methods for computing this limit, which is crucial for understanding the theoretical boundaries of model compression.
Business Value
Enables more aggressive and theoretically grounded model compression, leading to smaller, faster, and more deployable deep learning models, especially in resource-constrained environments.