Language Adaptation for Speaker Recognition Systems Using Contrastive Learning.

SPECOM(2021)

引用 1|浏览16
暂无评分
摘要
In this article we propose to study several approaches to adapt a system between two languages. To train the state of the art x-vector Speaker Verification system, we need a huge amount of labeled speech data. If this constraint is satisfied in English (due to Voxceleb), it is not in our target domain: French. We use a supervised Contrastive Learning to transfer knowledge between source and target domain. Among the two other proposed adaptation approaches (Multilingual Learning and Transfert Learning) we show that the one based on Contrastive Learning gives the best performance: about 30% relative gain in term of Equal Error Rate with respect to the baseline system. We also show the robustness of the Contrastive Learning with respect to the duration (from very short to short) as well as to distortion presence (noise, reverberation).
更多
查看译文
关键词
Speaker recognition, Domain adaptation, Contrastive learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要