A Comprehensive Evaluation of Convolutional Hardware Accelerators

IEEE Transactions on Circuits and Systems II: Express Briefs(2023)

引用 0|浏览5
暂无评分
摘要
Industry increasingly adopts Convolutional Neural Networks (CNNs) in applications ranging from IoT to autonomous driving. Convolutional hardware accelerators for the inference phase are an option for CPUs and GPUs due to the smaller power consumption and improved performance. The literature presents hardware accelerators using different 2D architectures, including weight stationary (WS), input stationary (IS), and output stationary (OS). The main differentiation between these architectures is how accelerators access data (input feature map and weight tensors) and compute the output (output feature map tensor). There is a gap in the literature related to a comprehensive evaluation of such architectures. This brief aims to answer the following question: "which accelerator type should I use according to my design constraints and memory type (SRAM or DRAM)". Experiments show that when using SRAM as external memory to the accelerator, WS presents the smallest area and energy consumption, while IS presents the best performance. On the other hand, the IS accelerator stands out when using DRAM because it has a reduced performance sensitivity to memory latency.
更多
查看译文
关键词
Convolution,Random access memory,Hardware acceleration,Memory management,Tensors,Arithmetic,Energy consumption,CNN,convolutional hardware accelerator,PPA,SRAM,DRAM
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要