Learning From Weights: Cost-Sensitive Approach For Retrieval

PROCEEDINGS OF THE 7TH ACM IKDD CODS AND 25TH COMAD (CODS-COMAD 2020)(2020)

引用 0|浏览10
暂无评分
摘要
In sponsored search, we need to retrieve the relevant top few ads from millions of ad copies within milliseconds. Running deep learning models or interaction models online becomes computationally expensive and hence not feasible for retrieval. There is not much discussion on improving cost-effective online retrieval models which work on representation based learning. In this paper we discuss one such improvement by incorporating cost-sensitive training. Online retrieval models which are trained on click-through data treats each clicked query-document pair as equivalent. While training on click-through data is reasonable, this paper argues that it is sub-optimal because of its noisy and long-tail nature (especially for sponsored search). In this paper, we discuss the impact of incorporating or disregarding the long tail pairs in the training set. Also, we propose a weighing based strategy using which we can learn semantic representations for tail pairs without compromising the quality of retrieval.Online A/B testing on live search engine traffic showed improvements in clicks (11.8% higher CTR) and as well as improvement in quality (8.2% lower bounce rate) when compared to the unweighted model. To prove the efficacy of the model we did offline experimentation with Bi-LSTM based representation model as well.
更多
查看译文
关键词
Neural Networks, Cost-sensitive Learning, Sponsored Ads
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要