Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration

2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)(2017)

引用 34|浏览86
暂无评分
摘要
State-of-the-art convolutional neural networks are enormously costly in both compute and memory, demanding massively parallel GPUs for execution. Such networks strain the computational capabilities and energy available to embedded and mobile processing platforms, restricting their use in many important applications. In this paper, we propose BCNN with Separable Filters (BCNNw/SF), which applies Singular Value Decomposition (SVD) on BCNN kernels to further reduce computational and storage complexity. We provide a closed form of the gradient over SVD to calculate the exact gradient with respect to every binarized weight in backward propagation. We verify BCNNw/SF on the MNIST, CIFAR-10, and SVHN datasets, and implement an accelerator for CIFAR10 on FPGA hardware. Our BCNNw/SF accelerator realizes memory savings of 17% and execution time reduction of 31.3% compared to BCNN with only minor accuracy sacrifices.
更多
查看译文
关键词
binarized convolutional neural network with separable filters,storage complexity,execution time reduction,memory savings,BCNNw/SF accelerator,FPGA hardware,CIFAR-10 datasets,SVHN datasets,MNIST datasets,backward propagation,binarized weight,computational complexity,BCNN kernels,singular value decomposition,BCNN with separable filters,hardware-effective CNN design,mobile processing platforms,embedded processing platforms,computational capabilities
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要