An Efficient Hardware Accelerator for Block Sparse Convolutional Neural Networks on FPGA

Xiaodi Yin, Zhipeng Wu,Dejian Li,Chongfei Shen,Yu Liu

IEEE Embedded Systems Letters(2023)

引用 0|浏览1
暂无评分
摘要
Field Programmable Gate Array (FPGA) has become an excellent hardware accelerator solution for convolutional neural networks (CNN). Meanwhile, optimizing methods such as model compression have been proposed. As most CNN accelerators focus on dense neural networks, to solve the problem of difficult hardware deployment due to irregular networks, we propose a method for sparse neural networks in our work. The storage and coding format of sparse data obtained by the block pruning method is designed to make it friendly to implement on FPGA. Besides, we also propose an efficient and simple data flow by the planarization of the whole convolution calculation process. The experimental result demonstrates that our implementation can achieve clock frequency of 190MHz, power consumption of 13.32W and inferencing speed of 16.37ms. Compared with some typical Mobilenet implementation schemes, our method has been proven to achieve a better balance between frequency, accuracy, power consumption and speed.
更多
查看译文
关键词
CNN accelerator,FPGA,sparse CNN,convolutional neural networks,block prunning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要