Robustness of end-to-end Automatic Speech Recognition Models -- A Case Study using Mozilla DeepSpeech

arxiv(2021)

引用 0|浏览5
暂无评分
摘要
When evaluating the performance of automatic speech recognition models, usually word error rate within a certain dataset is used. Special care must be taken in understanding the dataset in order to report realistic performance numbers. We argue that many performance numbers reported probably underestimate the expected error rate. We conduct experiments controlling for selection bias, gender as well as overlap (between training and test data) in content, voices, and recording conditions. We find that content overlap has the biggest impact, but other factors like gender also play a role.
更多
查看译文
关键词
automatic speech recognition models,speech recognition,end-to-end
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要