ConML: A Universal Meta-Learning Framework with Task-Level Contrastive Learning

📅 2024-10-08
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limited cross-task generalization capability of models in meta-learning. We propose ConML, a general framework that—uniquely—extends contrastive learning from representation space to *model space*, using task identity as supervision: it pulls together models trained on different subsets of the same task while pushing apart models trained on distinct tasks, all within the parameter space. ConML is architecture- and paradigm-agnostic, seamlessly integrating optimization-based, metric-based, variational, and context-learning meta-learning approaches. Extensive experiments demonstrate that ConML consistently improves performance across diverse few-shot benchmarks, achieving both broad applicability and empirical effectiveness. By introducing contrastive regularization directly over model parameters, ConML establishes a novel perspective on model-space regularization in meta-learning.

Technology Category

Application Category

📝 Abstract
Meta-learning enables learning systems to adapt quickly to new tasks, similar to humans. To emulate this human-like rapid learning and enhance alignment and discrimination abilities, we propose ConML, a universal meta-learning framework that can be applied to various meta-learning algorithms without relying on specific model architectures nor target models. The core of ConML is task-level contrastive learning, which extends contrastive learning from the representation space in unsupervised learning to the model space in meta-learning. By leveraging task identity as an additional supervision signal during meta-training, we contrast the outputs of the meta-learner in the model space, minimizing inner-task distance (between models trained on different subsets of the same task) and maximizing inter-task distance (between models from different tasks). We demonstrate that ConML integrates seamlessly with optimization-based, metric-based, and amortization-based meta-learning algorithms, as well as in-context learning, resulting in performance improvements across diverse few-shot learning tasks.
Problem

Research questions and friction points this paper is trying to address.

Improving meta-learning generalizability using task identity supervision
Contrasting model representations to enhance discrimination ability
Integrating contrastive meta-objective with existing meta-learners seamlessly
Innovation

Methods, ideas, or system contributions that make the work stand out.

Contrastive meta-objective optimizes task identity supervision
Learner-agnostic framework integrates with existing meta-learners
Model representations contrasted to improve generalization ability
🔎 Similar Papers
No similar papers found.
S
Shiguang Wu
Department of Electronic Engineering, Tsinghua University
Y
Yaqing Wang
Beijing Institute of Mathematical Sciences and Applications
Yatao Bian
Yatao Bian
ETH
Scientific IntelligenceEnergy Based ModelGraph Machine LearningLarge Models
Quanming Yao
Quanming Yao
Associate Professor, EE Department, Tsinghua University
Machine Learning