29.3 An 8.09TOPS/W Neural Engine Leveraging Bit-Sparsified Sign-Magnitude Multiplications and Dual Adder Trees

2023 IEEE International Solid- State Circuits Conference (ISSCC)(2023)

引用 0|浏览2
暂无评分
摘要
The computational complexity of neural networks (NNs) continues to increase, spurring the development of high-efficiency neural accelerator engines. Previous neural engines have relied on two's-complement (2C) arithmetic for their central MAC units (Fig. 29.3.1 top, left). However, gate-level simulations show that sign-magnitude (SM) multiplication is significantly more energy efficient; ranging from 35% (with uniformly distributed operands) to 67% (with normally distributed operands $(\mu={0}, {\sigma=25})$ ). The drawback of sign-magnitude number representation is that SM addition incurs significant overhead in terms of energy consumption and area, requiring upfront comparison of the sign bits and muxing/control to appropriately select between addition and subtraction (Fig. 29.3.1 center, left). This SM addition overhead substantially offsets the gains from SM multiplication in general purpose computing. One recent effort [1] to employ SM representation in neural computation achieved modest energy improvement at the cost of $2.5\times$ area increase due to full duplication of the MAC units, which would typically be unacceptable for area-/cost-sensitive IoT applications.
更多
查看译文
关键词
2.5× area increase,bit-sparsified sign-magnitude multiplications,central MAC units,computational complexity,energy consumption,Fig. 29.3.1 center,gate-level simulations,general purpose computing,high-efficiency neural accelerator engines,modest energy improvement,neural computation,neural networks,normally distributed operands,previous neural engines,sign bits,sign-magnitude multiplication,sign-magnitude number representation,significant overhead,SM addition,SM multiplication,SM representation,subtraction,uniformly distributed operands
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要