Token-Level Interpolation For Class-Based Language Models

2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2015)

引用 5|浏览75
暂无评分
摘要
We describe a method for interpolation of class-based n-gram language models. Our algorithm is an extension of the traditional EMbased approach that optimizes perplexity of the training set with respect to a collection of n-gram language models linearly combined in the probability space. However, unlike prior work, it naturally supports context-dependent interpolation for class-based LMs. In addition, the method works naturally with the recently introduced word-phrase-entity (WPE) language models that unify words, phrases and entities into a single statistical framework. Applied to the Calendar scenario of the Personal Assistant domain, our method achieved significant perplexity reduction and improved word error rates.
更多
查看译文
关键词
language model interpolation,class-based language models,context-dependent interpolation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要