Skip to main navigation Skip to search Skip to main content

Growing-before-pruning: A progressive neural architecture search strategy via group sparsity and deterministic annealing

  • Xiaotong Lu
  • , Weisheng Dong
  • , Zhenxuan Fang
  • , Jie Lin
  • , Xin Li
  • , Guangming Shi

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

Network pruning is a widely studied technique of obtaining compact representations from over-parameterized deep convolutional neural networks. Existing pruning methods are based on finding an optimal combination of pruned filters in the fixed search space. However, the optimality of those methods is often questionable due to limited search space and pruning choices - e.g., the difficulty with removing the entire layer and the risk of unexpected performance degradation. Inspired by the exploration vs. exploitation trade-off in reinforcement learning, we propose to reconstruct the filter space without increasing the model capacity and prune them by exploiting group sparsity. Our approach challenges the conventional wisdom by advocating the strategy of Growing-before-Pruning (GbP), which allows us to explore more space before exploiting the power of architecture search. Meanwhile, to achieve more efficient pruning, we propose to measure the importance of filters by global group sparsity, which extends the existing Gaussian scale mixture model. Such global characterization of sparsity in the filter space leads to a novel deterministic annealing strategy for progressively pruning the filters. We have evaluated our method on several popular datasets and network architectures. Our extensive experiment results have shown that the proposed method advances the current state-of-the-art.

Original languageEnglish
Article number111697
JournalPattern Recognition
Volume166
DOIs
StatePublished - Oct 2025

Keywords

  • Deterministic annealing
  • Group sparsity
  • Model compression
  • Neural architecture search

Fingerprint

Dive into the research topics of 'Growing-before-pruning: A progressive neural architecture search strategy via group sparsity and deterministic annealing'. Together they form a unique fingerprint.

Cite this