Sot-Mram Based Analog In-Memory Computing For Dnn Inference

2020 IEEE SYMPOSIUM ON VLSI TECHNOLOGY(2020)

引用 20|浏览7
暂无评分
摘要
Deep neural network (DNN) inference requires a massive amount of matrix-vector multiplications which can be computed efficiently on memory arrays in an analog fashion. This approach requires highly resistive memory devices (>MS 2) with low resistance variability to implement DNN weight memories. We propose an optimized Spin-Orbit Torque MRAM (SOTMRAM) as weight memory in Analog in-Memory Computing (AiMC) systems for DNN inference. In SOT-MRAM the write and read path are decoupled. This allows changing the MTJ resistance to the high levels required for AiMC by tuning the tunnel barrier thickness without affecting the writing.The target resistance level and variation are derived from an algorithm driven design-technology-co-optimization (DTCO) study. Resistance levels are obtained from IR-drop simulations of a convolutional neural network (CNN). Variation limits are obtained by testing two noise-resilient CNNs with conductance variability. Finally, we demonstrate experimentally that the requirements for analog DNN inference are met by SOT-MRAM stack optimization.
更多
查看译文
关键词
highly resistive memory devices,weight memory,MTJ resistance,analog DNN inference,SOT-MRAM stack optimization,deep neural network inference,matrix-vector multiplications,memory arrays,algorithm driven design-technology cooptimization,optimized spin-orbit torque MRAM,analog in-memory computing,DNN inference,deep neural network inference,AiMC,tunnel barrier thickness,IR-drop simulation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要