AnyPattern: Towards In-context Image Copy Detection
CoRR(2024)
摘要
This paper explores in-context learning for image copy detection (ICD), i.e.,
prompting an ICD model to identify replicated images with new tampering
patterns without the need for additional training. The prompts (or the
contexts) are from a small set of image-replica pairs that reflect the new
patterns and are used at inference time. Such in-context ICD has good realistic
value, because it requires no fine-tuning and thus facilitates fast reaction
against the emergence of unseen patterns. To accommodate the "seen
→ unseen" generalization scenario, we construct the first
large-scale pattern dataset named AnyPattern, which has the largest number of
tamper patterns (90 for training and 10 for testing) among all the existing
ones. We benchmark AnyPattern with popular ICD methods and reveal that existing
methods barely generalize to novel tamper patterns. We further propose a simple
in-context ICD method named ImageStacker. ImageStacker learns to select the
most representative image-replica pairs and employs them as the pattern prompts
in a stacking manner (rather than the popular concatenation manner).
Experimental results show (1) training with our large-scale dataset
substantially benefits pattern generalization (+26.66 % μ AP), (2) the
proposed ImageStacker facilitates effective in-context ICD (another round of
+16.75 % μ AP), and (3) AnyPattern enables in-context ICD, i.e. without
such a large-scale dataset, in-context learning does not emerge even with our
ImageStacker. The project (including the proposed dataset AnyPattern and the
code for ImageStacker) is publicly available at https://anypattern.github.io
under the MIT Licence.
更多查看译文
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要