FPGA Demonstrator of a Programmable Ultra-Efficient Memristor-Based Machine Learning Inference Accelerator

Martin Foltin,Glaucimar Aguiar,Rodrigo Antunes,Plinio Silveira,Gustavo Knuppe,Joao Ambrosi,Soumitra Chatterjee, Jitendra Onkar Kolhe, Sunil Lakshiminarashimha,Dejan S. Milojicic,John Paul Strachan,Craig Warner,Amit Sharma, Eddie Lee,Sai Rahul Chalamalasetti, Chris Brueggen, Charles Williams, Nathaniel Jansen, Felipe Saenz, Luis Federico Li

2019 IEEE International Conference on Rebooting Computing (ICRC)(2019)

引用 1|浏览77
暂无评分
摘要
Hybrid analog-digital neuromorphic accelerators show promise for significant increase in performance per watt of deep learning inference and training as compared with conventional technologies. In this work we present an FPGA demonstrator of a programmable hybrid inferencing accelerator, with memristor analog dot product engines emulated by digital matrix-vector multiplication units employing FPGA SRAM memory for in-situ weight storage. The full-chip demonstrator interfaced to a host by PCIe interface serves as a software development platform and a vehicle for further hardware microarchitecture improvements. Implementation of compute cores, tiles, network on a chip, and the host interface is discussed. New pipelining scheme is introduced to achieve high utilization of matrix-vector multiplication units while reducing tile data memory size requirements for neural network layer activations. The data flow orchestration between the tiles is described, controlled by a RISC-V core. Inferencing accuracy analysis is presented for an example RNN and CNN models. The demonstrator is instrumented with hardware monitors to enable performance measurements and tuning. Performance projections for future memristor-based ASIC are also discussed.
更多
查看译文
关键词
Deep Neural Network Inference,Neural Network Acceleration,Memristor
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要