How Powerful Potential of Attention on Image Restoration?
arxiv(2024)
摘要
Transformers have demonstrated their effectiveness in image restoration
tasks. Existing Transformer architectures typically comprise two essential
components: multi-head self-attention and feed-forward network (FFN). The
former captures long-range pixel dependencies, while the latter enables the
model to learn complex patterns and relationships in the data. Previous studies
have demonstrated that FFNs are key-value memories ,
which are vital in modern Transformer architectures. In this paper, we conduct
an empirical study to explore the potential of attention mechanisms without
using FFN and provide novel structures to demonstrate that removing FFN is
flexible for image restoration. Specifically, we propose Continuous Scaling
Attention (CSAttn), a method that computes attention continuously in
three stages without using FFN. To achieve competitive performance, we propose
a series of key components within the attention. Our designs provide a closer
look at the attention mechanism and reveal that some simple operations can
significantly affect the model performance. We apply our CSAttn to
several image restoration tasks and show that our model can outperform
CNN-based and Transformer-based image restoration approaches.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要