ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation

Jiazheng Xu,Xiao Liu, Yuchen Wu, Yuxuan Tong, Qinkai Li,Ming Ding,Jie Tang大牛学者,Yuxiao Dong

CoRR(2023)

引用 11|浏览1024
暂无评分
摘要
We present ImageReward -- the first general-purpose text-to-image human preference reward model -- to address various prevalent issues in generative models and align them with human values and preferences. Its training is based on our systematic annotation pipeline that covers both the rating and ranking components, collecting a dataset of 137k expert comparisons to date. In human evaluation, ImageReward outperforms existing scoring methods (e.g., CLIP by 38.6\%), making it a promising automatic metric for evaluating and improving text-to-image synthesis. The reward model is publicly available via the \texttt{image-reward} package at \url{https://github.com/THUDM/ImageReward}.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络