Running sparse and low-precision neural network: When algorithm meets hardware.

ASP-DAC(2018)

引用 19|浏览60
暂无评分
摘要
Deep Neural Networks (DNNs) are pervasively applied in many artificial intelligence (AI) applications. The high performance of DNNs comes at the cost of larger size and higher compute complexity. Recent studies show that DNNs have much redundancy, such as the zero-value parameters and excessive numerical precision. To reduce computing complexity, many redundancy reduction techniques have been proposed, including pruning and data quantization. In this paper, we demonstrate our cooptimization of the DNN algorithm and hardware which exploits the model redundancy to accelerate DNNs.
更多
查看译文
关键词
low-precision neural network,artificial intelligence applications,zero-value parameters,computing complexity,redundancy reduction techniques,deep neural networks,numerical precision,sparse neural network,DNN algorithm co-optimization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要