Building a Bilingual QA-system with ruGPT-3.

International Joint Conference on the Analysis of Images, Social Networks and Texts (AIST)(2021)

引用 0|浏览6
暂无评分
摘要
In this work, we present an approach of cross-lingual transfer learning for English and Russian languages within the QA task. Our approach implies using a generative transformer model that has seen Wikipedia texts in both languages during the pretraining phase and then fine-tuning it with a special token of language, forcing the model to generate texts in a particular language. We are focusing on SQuAD data (English) and updated SberQuAD data (Russian) plus their translations for training and testing, and use ruGPT-3 XL model, which is forced to answer questions in English based on Russian paragraphs and reverse: can answer in Russian when provided information in English. Monolingual QA-abilities of the model are also preserved. Our results show that the fine-tuned model demonstrates bilingual ability and can generate answers that are close to correct in fuzzy metrics: model generates answers in Russian when based on English texts: 75% named entities ratio, 28% Levenshtein Distance string matching, 28% ROUGE-L; model generates answers in English when based on Russian data: 51% named entities ratio, 27% Levenshtein Distance string matching, 27% ROUGE-L; monolingual Russian quality: 83% named entities ratio, 59% Levenshtein Distance string matching, 57% ROUGE-L; monolingual English quality: 52% named entities ratio, 24% Levenshtein Distance string matching, 25% ROUGE-L.
更多
查看译文
关键词
qa-system
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要