On integrating a language model into neural machine translation

Computer Speech and Language(2017)

引用 123|浏览276
暂无评分
摘要
Recent advances in end-to-end neural machine translation models have achieved promising results on high-resource language pairs such as En Fr and En De. One of the major factor behind these successes is the availability of high quality parallel corpora. We explore two strategies on leveraging abundant amount of monolingual data for neural machine translation. We observe improvements by both combining scores from neural language model trained only on target monolingual data with neural machine translation model and fusing hidden-states of these two models. We obtain up to 2 BLEU improvement over hierarchical and phrase-based baseline on low-resource language pair, Turkish English. Our method was initially motivated towards tasks with less parallel data, but we also show that it extends to high resource languages such as Cs En and De En translation tasks, where we obtain 0.39 and 0.47 BLEU improvements over the neural machine translation baselines, respectively.
更多
查看译文
关键词
Neural machine translation,Monolingual data,Language models,Low resource machine translation,Deep learning,Neural network
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要