Random Clusterings for Language Modeling

ICASSP '05). IEEE International Conference(2005)

引用 33|浏览46
暂无评分
摘要
In this paper we present an application of randomization techniques to class-based n-gram language models. The idea is to derive a language model from the combination of a set of random class-based models. Each of the constituent random class-based models is built using a separate clustering obtained via a different run of a randomized clustering algorithm. The random class-based model can compensate for some of the shortcomings of conventional class-based models by combining the different solutions obtained through random clusterings. Experimental results show that the combined random class-based model improves considerably in perplexity (PPL) and word error rate (WER) over both the n-gram and baseline class-based models.
更多
查看译文
关键词
error statistics,natural languages,random processes,speech recognition,class-based n-gram language models,perplexity,random class-based models,random clustering,speech recognizer,word error rate
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要