Accelerating CNN computation: quantisation tuning and network resizing

Proceedings of the 2nd Workshop on AutotuniNg and aDaptivity AppRoaches for Energy efficient HPC Systems(2018)

引用 0|浏览11
暂无评分
摘要
The interest in developing cognitive aware systems, specially for vision applications based on artificial neural networks, has grown exponentially in the last years. While high performance systems are key for the success of current Convolutional Neural Network (CNN) implementations, there is a trend to bring these capabilities to embedded real-time systems. This work contributes to tackle this challenge by exploring CNNs design space. Namely, it combines parameter quantisation techniques with a proposed set of CNN architectural transformations to reduce resource and execution time costs on Field Programmable Gate Array (FPGA) devices while maintaining high classification accuracy. An hardware mapping methodology is also proposed for deploying resource constrained CNNs into a reconfigurable platform for efficient algorithm acceleration. The proposed transformations reduce accuracy loss due to quantization by 44% in average. Also, analysis of the performance results obtained in a Central Processing Unit (CPU)+FPGA platform show up to 50% execution time reduction when compared with a state-of-the-art implementation.
更多
查看译文
关键词
FPGA, OpenCL, convolutional neural network, driver assistance, object recognition
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要