CAMixerSR: Only Details Need More "Attention"
CVPR 2024(2024)
摘要
To satisfy the rapidly increasing demands on the large image (2K-8K)
super-resolution (SR), prevailing methods follow two independent tracks: 1)
accelerate existing networks by content-aware routing, and 2) design better
super-resolution networks via token mixer refining. Despite directness, they
encounter unavoidable defects (e.g., inflexible route or non-discriminative
processing) limiting further improvements of quality-complexity trade-off. To
erase the drawbacks, we integrate these schemes by proposing a content-aware
mixer (CAMixer), which assigns convolution for simple contexts and additional
deformable window-attention for sparse textures. Specifically, the CAMixer uses
a learnable predictor to generate multiple bootstraps, including offsets for
windows warping, a mask for classifying windows, and convolutional attentions
for endowing convolution with the dynamic property, which modulates attention
to include more useful textures self-adaptively and improves the representation
capability of convolution. We further introduce a global classification loss to
improve the accuracy of predictors. By simply stacking CAMixers, we obtain
CAMixerSR which achieves superior performance on large-image SR, lightweight
SR, and omnidirectional-image SR.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要