An Energy-Efficient Ring-Based CIM Accelerator using High-Linearity eNVM for Deep Neural Networks.

Po-Tsang Huang, Ting-Wei Liu, Wei Lu, Yu-Hsien Lin,Wei Hwang

ISOCC(2021)

引用 0|浏览4
暂无评分
摘要
Computation-in-memory (CIM) accelerators reduce the energy consumption of weight accesses from off-chip memory by storing synaptic weights into on-chip embedded NVM (eNVM) devices, such as RRAM, charge-trap transistor and FeFET. However, for mapping a deep neural network (DNN) more than 20 layers into an eNVM-based accelerator, the throughput, energy-efficiency and accuracy are limited due to the non-linearity of weights and energyconsuming weight updating. In this work, the proposed CIM accelerator exploits low-voltage and high-linearity eNVM to reach both the power efficiency of weight updating and the high accuracy. By adopting the layer-level weight stationary, mini-array clusters and a ring-based architecture, the resource utilization of eNVM devices is increased. In addition, channel-wise weight mapping schemes for standard convolution and pointwise convolution can support the structure pruning technique of DNNs. The proposed accelerator achieves 1.814 TOPS/W with only 4.7% accuracy loss on YOLOv3 by Ni-crystal RRAM.
更多
查看译文
关键词
energy-consuming weight updating,synaptic weights,off-chip memory,weight accesses,energy consumption,computation-in-memory accelerators,deep neural networks,energy-efficient ring-based CIM accelerator,channel-wise weight mapping schemes,eNVM devices,ring-based architecture,mini-array clusters,weight updating,power efficiency,high-linearity eNVM,energy-efficiency,charge-trap transistor,on-chip embedded NVM devices
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要