CycleGAN-based speech enhancement for the unpaired training data

Asia-Pacific Signal and Information Processing Association Annual Summit and Conference(2019)

引用 3|浏览4
暂无评分
摘要
Speech enhancement is an important task of improving speech quality in noise scenario. Many speech enhancement methods have achieved remarkable success based on the paired data. However, for many tasks, the paired training data is not available. In this paper, we present a speech enhancement method for the unpaired data based on cycle-consistent generative adversarial network (CycleGAN) that can minimize the reconstruction loss as much as possible. The proposed model employs two discriminators and two generators to preserve speech components and reduce noise so that the network could map features better for the unseen noise. In this method, the generators are used to generate the enhanced speech, and two discriminators are employed to discriminate real inputs and the outputs of the generators. The experimental results showed that the proposed method effectively improved the performance compared to traditional deep neural network (DNN) and the recent GAN-based speech enhancement methods.
更多
查看译文
关键词
cycle-consistent generative adversarial network,speech components,CycleGAN-based speech enhancement,unpaired training data,speech quality,noise scenario,speech enhancement method,paired training data,GAN-based speech enhancement methods,deep neural network
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要