Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters
CVPR 2024(2024)
摘要
Continual learning can empower vision-language models to continuously acquire
new knowledge, without the need for access to the entire historical dataset.
However, mitigating the performance degradation in large-scale models is
non-trivial due to (i) parameter shifts throughout lifelong learning and (ii)
significant computational burdens associated with full-model tuning. In this
work, we present a parameter-efficient continual learning framework to
alleviate long-term forgetting in incremental learning with vision-language
models. Our approach involves the dynamic expansion of a pre-trained CLIP
model, through the integration of Mixture-of-Experts (MoE) adapters in response
to new tasks. To preserve the zero-shot recognition capability of
vision-language models, we further introduce a Distribution Discriminative
Auto-Selector (DDAS) that automatically routes in-distribution and
out-of-distribution inputs to the MoE Adapter and the original CLIP,
respectively. Through extensive experiments across various settings, our
proposed method consistently outperforms previous state-of-the-art approaches
while concurrently reducing parameter training burdens by 60
at https://github.com/JiazuoYu/MoE-Adapters4CL
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要