StemGen: A music generation model that listens

Julian D. Parker, Janne Spijkervet,Katerina Kosta, Furkan Yesiler, Boris Kuznetsov,Ju-Chiang Wang, Matt Avent,Jitong Chen, Duc Le

CoRR(2023)

引用 0|浏览4
暂无评分
摘要
End-to-end generation of musical audio using deep learning techniques has seen an explosion of activity recently. However, most models concentrate on generating fully mixed music in response to abstract conditioning information. In this work, we present an alternative paradigm for producing music generation models that can listen and respond to musical context. We describe how such a model can be constructed using a non-autoregressive, transformer-based model architecture and present a number of novel architectural and sampling improvements. We train the described architecture on both an open-source and a proprietary dataset. We evaluate the produced models using standard quality metrics and a new approach based on music information retrieval descriptors. The resulting model reaches the audio quality of state-of-the-art text-conditioned models, as well as exhibiting strong musical coherence with its context.
更多
查看译文
关键词
Music Generation,Deep Learning,LLMs,Generative Models
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要