Block-Based Compression for Reducing Indexing Cost of DNN Accelerators
ICCE-TW(2021)
摘要
Sparse compression is often used in DNN (deep neural network) accelerators to reduce the data traffic of convolutional computation. As the quantization technique greatly reduces the data bit-width, the indexing cost may account for a large proportion of data traffic after sparse compression. In this paper, we point out the similarity of feature maps in different channels. Based on this observation, we propose a block-based compression approach to reduce the indexing cost. Benchmark data show that the proposed approach can reduce 21.1% indexing cost.
更多查看译文
关键词
Data Compression,Digital Design,Feature Maps,Neural Networks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要