Is Meta-training Really Necessary for Molecular Few-Shot Learning ?
arxiv(2024)
摘要
Few-shot learning has recently attracted significant interest in drug
discovery, with a recent, fast-growing literature mostly involving convoluted
meta-learning strategies. We revisit the more straightforward fine-tuning
approach for molecular data, and propose a regularized quadratic-probe loss
based on the the Mahalanobis distance. We design a dedicated block-coordinate
descent optimizer, which avoid the degenerate solutions of our loss.
Interestingly, our simple fine-tuning approach achieves highly competitive
performances in comparison to state-of-the-art methods, while being applicable
to black-box settings and removing the need for specific episodic pre-training
strategies. Furthermore, we introduce a new benchmark to assess the robustness
of the competing methods to domain shifts. In this setting, our fine-tuning
baseline obtains consistently better results than meta-learning methods.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要