🤖 AI Summary
This work addresses the robustness of molecular few-shot learning under domain shift in drug discovery, challenging mainstream meta-learning paradigms by proposing a lightweight, fine-tuning-based framework. Methodologically, it introduces (1) a regularized quadratic probe loss incorporating Mahalanobis distance into few-shot fine-tuning to enhance intra-class compactness and inter-class discriminability; (2) a dedicated block-coordinate descent optimizer designed to avoid degenerate solutions; and (3) the first few-shot benchmark specifically curated for evaluating domain-shift robustness in molecular representation learning. The method is model-agnostic—compatible with any molecular graph neural network—and supports plug-and-play integration of black-box models without episodic pretraining. Experiments demonstrate competitive performance against state-of-the-art meta-learning methods on standard few-shot tasks, while significantly outperforming them under domain shift.
📝 Abstract
Few-shot learning has recently attracted significant interest in drug discovery, with a recent, fast-growing literature mostly involving convoluted meta-learning strategies. We revisit the more straightforward fine-tuning approach for molecular data, and propose a regularized quadratic-probe loss based on the the Mahalanobis distance. We design a dedicated block-coordinate descent optimizer, which avoid the degenerate solutions of our loss. Interestingly, our simple fine-tuning approach achieves highly competitive performances in comparison to state-of-the-art methods, while being applicable to black-box settings and removing the need for specific episodic pre-training strategies. Furthermore, we introduce a new benchmark to assess the robustness of the competing methods to domain shifts. In this setting, our fine-tuning baseline obtains consistently better results than meta-learning methods.