Improving Network Slimming With Nonconvex Regularization

IEEE ACCESS(2021)

引用 3|浏览11
暂无评分
摘要
Convolutl(1) regularization on the channel-associated scaling factors via the batch normalization layers during training. Network slimming thereby identifies insignificant channels that can be pruned for inference. In this paper, we propose replacing the `1 penalty with an alternative nonconvex, sparsity-inducing penalty in order to yield a more compressed and/or accurate CNN architecture. We investigate l(p)(0 < p < 1), transformed l(1) (T l(1)), minimax concave penalty (MCP), and smoothly clipped absolute deviation (SCAD) due to their recent successes and popularity in solving sparse optimization problems, such as compressed sensing and variable selection. We demonstrate the effectiveness of network slimming with nonconvex penalties on three neural network architectures - VGG-19, DenseNet-40, and ResNet-164 = on standard image classification datasets. Based on the numerical experiments, T `1 preserves model accuracy against channel pruning, l(1) =2;3 =4 yield better compressed models with similar accuracies after retraining as l(1), and MCP and SCAD provide more accurate models after retraining with similar compression as l(1). Network slimming with T l(1) regularization also outperforms the latest Bayesian modification of network slimming in compressing a CNN architecture in terms of memory storage while preserving its model accuracy after channel pruning.
更多
查看译文
关键词
Convolutional neural networks (CNN), machine learning, deep learning, network pruning, nonconvex optimization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要