Block-Based Compression for Reducing Indexing Cost of DNN Accelerators

ICCE-TW(2021)

引用 0|浏览2
暂无评分
摘要
Sparse compression is often used in DNN (deep neural network) accelerators to reduce the data traffic of convolutional computation. As the quantization technique greatly reduces the data bit-width, the indexing cost may account for a large proportion of data traffic after sparse compression. In this paper, we point out the similarity of feature maps in different channels. Based on this observation, we propose a block-based compression approach to reduce the indexing cost. Benchmark data show that the proposed approach can reduce 21.1% indexing cost.
更多
查看译文
关键词
Data Compression,Digital Design,Feature Maps,Neural Networks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要