Massively Parallel Cross-Lingual Learning in Low-Resource Target Language Translation.

WMT(2018)

引用 11|浏览15
暂无评分
摘要
We work on translation from rich-resource languages to low-resource languages. The main challenges we identify are the lack of low-resource language data, effective methods for cross-lingual transfer, and the variable-binding problem that is common in neural systems. We build a translation system that addresses these challenges using eight European language families as our test ground. Firstly, we add source and target family labels and study intra- family and inter-family influences for effective cross-lingual transfer. We achieve improvement of +8.4 BLEU score compared to single-family multi-source multi- target NMT baseline. We find that training two neighboring families closest to the low-resource language is often enough. Secondly, we construct an ablation study and find that reasonably good results can be achieved even with considerably less target data. Thirdly, we address the variable-binding problem by building an order-preserving named entity translation model. We obtain 60.6% accuracy in qualitative evaluation where our translations are akin to human translations in a preliminary study.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要