Efficient Training of LDA on a GPU by Mean-for-Mode Estimation.

ICML'15: Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37(2015)

引用 16|浏览98
暂无评分
摘要
We introduce Mean-for-Mode estimation, a variant of an uncollapsed Gibbs sampler that we use to train LDA on a GPU. The algorithm combines benefits of both uncollapsed and collapsed Gibbs samplers. Like a collapsed Gibbs sampler--and unlike an uncollapsed Gibbs sampler--it has good statistical performance, and can use sampling complexity reduction techniques such as sparsity. Meanwhile, like an uncollapsed Gibbs sampler--and unlike a collapsed Gibbs sampler --it is embarrassingly parallel, and can use approximate counters.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要