Text to Image Synthesis based on Multi-Perspective Fusion

2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN)(2021)

引用 2|浏览14
暂无评分
摘要
In this paper, we propose a multi-perspective fusion method to improve the performance of text-to-image synthesis. From the perspective of the generator, we introduce a dynamic selection method to make the text feature match the corresponding image feature better, while the multi-class discriminant method with mask segmentation image as the extra type is introduced from the perspective of the discriminator to improve its discrimination ability. Through the effective integration of these two aspects of improvement, more excellent results by our method are obtained. Experiments on the Caltech-UCSD Birds 200 (CUB) and Microsoft Common Objects in Context (MS COCO) datasets demonstrate our method's effectiveness and superiority. The qualitative and quantitative experiments validate that our method is superior to the existing state-of-theart methods.
更多
查看译文
关键词
dynamic selection method,text feature,corresponding image feature,multiclass discriminant method,mask segmentation image,extra type,discrimination ability,Caltech-UCSD Birds 200,multi- perspective fusion,multiperspective fusion method,text-to-image synthesis
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要