Abstract
Network pruning is a widely studied technique of obtaining compact representations from over-parameterized deep convolutional neural networks. Existing pruning methods are based on finding an optimal combination of pruned filters in the fixed search space. However, the optimality of those methods is often questionable due to limited search space and pruning choices - e.g., the difficulty with removing the entire layer and the risk of unexpected performance degradation. Inspired by the exploration vs. exploitation trade-off in reinforcement learning, we propose to reconstruct the filter space without increasing the model capacity and prune them by exploiting group sparsity. Our approach challenges the conventional wisdom by advocating the strategy of Growing-before-Pruning (GbP), which allows us to explore more space before exploiting the power of architecture search. Meanwhile, to achieve more efficient pruning, we propose to measure the importance of filters by global group sparsity, which extends the existing Gaussian scale mixture model. Such global characterization of sparsity in the filter space leads to a novel deterministic annealing strategy for progressively pruning the filters. We have evaluated our method on several popular datasets and network architectures. Our extensive experiment results have shown that the proposed method advances the current state-of-the-art.
| Original language | English |
|---|---|
| Article number | 111697 |
| Journal | Pattern Recognition |
| Volume | 166 |
| DOIs | |
| State | Published - Oct 2025 |
Keywords
- Deterministic annealing
- Group sparsity
- Model compression
- Neural architecture search
Fingerprint
Dive into the research topics of 'Growing-before-pruning: A progressive neural architecture search strategy via group sparsity and deterministic annealing'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver