FPGA-based acceleration of neural network for ranking in web search engine with a streaming architecture

FPL(2009)

引用 7|浏览83
暂无评分
摘要
Web search engine companies are intensively running learning to rank algorithms to improve the search relevance. Neural network (NN)-based approaches, such as LambdaRank, can significantly increase the ranking quality. While, their training is very slow on a single computer and inherent coarse-grained parallelism could be hardly utilized by computer clusters. Thus an efficient implementation is necessary to timely generate acceptable NN models on frequently updated training datasets. This paper presents our work in accelerator. A SIMD streaming architecture is proposed to i) efficiently map the query-level NN computation and data structure to FPGA, ii) fully exploit the inherent fine-grained parallelism, and iii) provide scalability to large scale datasets. The accelerator shows up to 17.9X speedup over the software implementation on datasets from a commercial search engine.
更多
查看译文
关键词
computer cluster,coarse-grained parallelism,software implementation,web search engine ranking,learning (artificial intelligence),data structures,lambdarank algorithm,neural network-based approach,data structure,query-level nn computation,web search relevance,parallel algorithms,internet,simd streaming architecture,large-scale dataset scalability,fpga-based acceleration,learning method,field programmable gate arrays,search engines,training method,neural nets,query processing,search engine,learning artificial intelligence,learning to rank,web search engine,neural network,data mining
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要