Exploring Structural Sparsity in CNN via Selective Penalty
Author(s): Zhao, MX (Zhao, Mingxin); Peng, JB (Peng, Junbo); Yu, SM (Yu, Shuangming); Liu, LY (Liu, Liyuan); Wu, NJ (Wu, Nanjian)
Source: IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY Volume: 32 Issue: 3 Pages: 1658-1666 DOI: 10.1109/TCSVT.2021.3071532 Published: MAR 2022
Abstract: Although massive pruning methods are emerging for achieving structural sparsity in convolutional neural networks (CNN), most of them target structures such as ResNet. Meanwhile, previous works take more interest in pruning filters inside a residual block and keep the shortcut connection intact, leading to an imbalanced network structure. In this paper, we focus on the penalty-based method to prune already compact networks. In contrast to the broadly used L-1 constraint, which shrinks the parameters uniformly, we propose a novel penalty term that is similar in shape to an upside-down Laplace distribution. The penalty allows us to impose more pressure on potential weak channels but protects others during training to avoid damaging crucial channels, especially for compact architectures. We also design a candidate selection strategy to cooperate with the penalty-based training procedure. Besides, we address the residual block pruning problem by a scaling factor elimination skill, which is often ignored in other research. Our method reduces 50% parameters of MobileNet v1/v2 with a tolerable accuracy degradation. We further conduct pruning on MobileNetv1-SSDLite to compress parameters by 60%, manifesting the ability to generalize to different visual tasks. The experiment results demonstrate that our method outperforms pruning frameworks based on channel importance without a complicated tuning for hyper-parameters like search-based methods.
Accession Number: WOS:000766700400060
ISSN: 1051-8215
eISSN: 1558-2205
Full Text: https://ieeexplore.ieee.org/document/9398648