Inner Product Accelerating Scheme Based on RRAM Array for Attention-Mechanism Neural Network

IEEE ELECTRON DEVICE LETTERS(2024)

引用 0|浏览15
暂无评分
摘要
The inner product between two activation vectors is crucial for implementing neural networks with the attention mechanism. In this work, we propose and experimentally validate a novel inner product accelerating scheme based on the resistive random access memory (RRAM) array. By utilizing the RRAM conductance to represent the significance of each bit, the inner product can be conducted with a high energy efficiency of analog computing. Furthermore, an attention-mechanism neural network is demonstrated for solving the Q&A task in the Facebook bAbI dataset. We verify software-comparable accuracy (98.5%) while achieving a 42-fold reduction in latency and a 71-fold reduction in energy consumption compared with the conventional digital platform.
更多
查看译文
关键词
Neural networks,Task analysis,Energy efficiency,Virtual machine monitors,Resistive RAM,Performance evaluation,Energy consumption,RRAM,computation in-memory,inner product,attention-mechanism neural network
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要