Compressing Convolutional Neural Networks By Pruning Density Peak Filters

IEEE ACCESS(2021)

引用 8|浏览9
暂无评分
摘要
With the recent development of GPUs, the depth of convolutional neural networks (CNNs) has increased, and its structure has become complex. Hence, it is challenging to deploy them into a hardware device owing to its immense computational cost and memory for storage parameters. We propose a method of pruning a filter located near the density peak, which grasps the density of the filter space for each layer to overcome this problem. The density is calculated in the filter space based on the number of neighboring filters within a certain distance around the filter and the distance to a denser space. Moreover, we do not remove all filters at once, but use a method of pruning a certain number iteratively, so that filters can be evenly pruned in multiple locations with high density inside the filter space. After that, we fine-tune the pruned network to restore their performance. The experimental results show the effectiveness of the proposed method with respect to the other methods using CIFAR-10, and ImageNet dataset on VGGNet and ResNet architecture. Notably, on CIFAR-10, our method reduces 60.8% of FLOPs on ResNet56 with 0.31% validation accuracy improvement. Moreover, we achieve up to 51.9% FLOPs reduction with a little accuracy drop on ImageNet for ResNet34.
更多
查看译文
关键词
Convolutional neural networks, compressing CNNs, filter pruning, density peak
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要