Mix and match networks: encoder-decoder alignment for zero-pair image translation

2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition(2018)

引用 34|浏览66
暂无评分
摘要
We address the problem of image translation between domains or modalities for which no direct paired data is available (i.e. zero-pair translation). We propose mix and match networks, based on multiple encoders and decoders aligned in such a way that other encoder-decoder pairs can be composed at test time to perform unseen image translation tasks between domains or modalities for which explicit paired samples were not seen during training. We study the impact of autoencoders, side information and losses in improving the alignment and transferability of trained pairwise translation models to unseen translations. We show our approach is scalable and can perform colorization and style transfer between unseen combinations of domains. We evaluate our system in a challenging cross-modal setting where semantic segmentation is estimated from depth images, without explicit access to any depth-semantic segmentation training pairs. Our model outperforms baselines based on pix2pix and CycleGAN models.
更多
查看译文
关键词
depth images,depth-semantic segmentation training pairs,encoder-decoder alignment,zero-pair image translation,modalities,multiple encoders,encoder-decoder pairs,unseen image translation tasks,explicit paired samples,transferability,trained pairwise translation models,unseen translations,colorization,style transfer,challenging cross-modal setting,mix and match networks,CycleGAN models,autoencoders
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要