Transfer learning from English to Slovak in speech recognition applications

2023 33rd International Conference Radioelektronika (RADIOELEKTRONIKA)(2023)

引用 1|浏览7
暂无评分
摘要
The speed of research can be slowed down by limited access to SOTA computational resources. This paper compares two different speech recognition approaches, training from scratch and transfer learning. It is focused on the transfer learning technique known as fine-tuning in cross-language adaptation speech recognition tasks. The main idea of this paper is to prove the feasibility of adapting the foreign speech recognition model to the Slovak ASR model. We analyze the benefits of BPE tokenization in speech recognition, which we have not previously tested, and a particular pre-trained model in terms of training speed and model quality. In practice, the publicly available pre-trained English Librispeech100 model is used to create a transferred English-to-Slovak model on TUKE-BNews-SK corpus subset. In addition, the paper provides details on the ESPnet framework workflows involved in the creation of the proposed end-to-end fine-tuned model.
更多
查看译文
关键词
byte-pair encoding,cross-language,English,ESPnet,fine-tuning,Librispeech100,low-resource,Slovak,speech recognition
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要