Dynamic Programming Assisted Quantization Approaches for Compressing Normal and Robust DNN Models

ASPDAC(2021)

引用 5|浏览8
暂无评分
摘要
ABSTRACTIn this work, we present effective quantization approaches for compressing the deep neural networks (DNNs). A key ingredient is a novel dynamic programming (DP) based algorithm to obtain the optimal solution of scalar K-means clustering. Based on the approaches with regularization and quantization function, two weight quantization approaches called DPR and DPQ for compressing normal DNNs are proposed respectively. Experiments show that they produce models with higher inference accuracy than recently proposed counterparts while achieving same or larger compression. They are also extended for compressing robust DNNs, and the relevant experiments show 16X compression of the robust ResNet-18 model with less than 3% accuracy drop on both natural and adversarial examples.
更多
查看译文
关键词
Dynamic Programing, Neural Network Compression, Quantization, Robust Model, Weight Sharing
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要