Exploiting Randomness In Sketching For Efficient Hardware Implementation Of Machine Learning Applications

2016 IEEE/ACM International Conference on Computer-Aided Design (ICCAD)(2016)

引用 0|浏览36
暂无评分
摘要
Energy-efficient processing of large matrices for big-data applications using hardware acceleration is an intense area of research. Sketching of large matrices into their lower-dimensional representations is an effective strategy. For the first time, this paper develops a highly energy-efficient hardware implementation of a class of sketching methods based on random projections, known as Johnson Lindenstrauss (JL) transform. Crucially, we show how the randomness inherent in the projection matrix can be exploited to create highly efficient fixed-point arithmetic realizations of several machine-learning applications. Specifically, the transform's random matrices have two key properties that allow for significant energy gains in hardware implementations. The first is the smoothing property that allows us to drastically reduce operand bit-width in computation of the JL transform itself. The second is the randomizing property that allows bit-width reduction in subsequent machine-learning applications. Further, we identify a random matrix construction method that exploits the special sparsity structure to result in the most hardware-efficient realization and implement the highly optimized transform on an FPGA.Experimental results on (1) the k-nearest neighbor (KNN) classification and (2) the principal component analysis (PCA) show that with the same bit-width the proposed flow utilizing random projection achieves an up to 7X improvement in both latency and energy. Furthermore, by exploiting the smoothing and randomizing properties we are able to use a 1-bit instead of a 4-bit multiplier within KNN, which results in additional 50% and 6% improvement in area and energy respectively. The proposed I/O streaming strategy along with the hardware-efficient JL algorithm identified by us is able to achieve a 50% runtime reduction, a 17% area reduction in the stage of random projection compared to a standard design.
更多
查看译文
关键词
Big-Data applications,hardware acceleration,energy-efficient hardware implementation,sketching,random projections,Johnson Lindenstrauss transform,projection matrix,highly efficient fixed-point arithmetic,machine-learning applications,smoothing property,JL transform,randomizing property,operand bit-width reduction,random matrix construction,sparsity structure,FPGA,k-nearest neighbor classification,principal component analysis,KNN,I/O streaming,hardware-efficient JL algorithm,PCA
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要