Estimation-Based Strategy Generation for Deep Neural Network Model Compression

Hongkai Wang, Jun Feng, Shuai Zhao, Yidan Wang,Dong Mao, Zuge Chen, Gongwu Ke,Gaoli Wang, Youqun Long

2023 IEEE 6th International Conference on Pattern Recognition and Artificial Intelligence (PRAI)(2023)

引用 0|浏览2
暂无评分
摘要
Compressing the neural network can significantly reduce its computational complexity, save resources and speed up inference time. However, current compression methods, whether used individually or in combination, often neglect the issue of compression strategy generation, making it challenging to obtain compressed models with the smallest accuracy degradation that meet the user's deployment requirements. This paper proposes a method for automatically generating compression strategy, aiming to achieve high-performance models that meet deployment requirements with minimal accuracy degradation. Firstly, we design a predictor to estimate the compression performance of the model if it is compressed by different compression methods such as distillation, pruning and quantization. This includes estimating the model size, the number of parameters, computational complexity, and memory access of the model after compression. Then a computational method for estimating the inference time of the model after compression is discussed. Based on the estimated results, user requirements and hardware parameters, a method for automatically generating compression strategy is designed, which outputs suitable combinations of compression methods and compression parameter settings. Experimental results on commonly used convolutional neural networks and Jetson Nano development board validated the effectiveness of the proposed method.
更多
查看译文
关键词
Model Compression,Compression Strategy Generation,Computational Time Estimation,Deploy Requirement
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要