🤖 AI Summary
Vision models lack context-driven conceptual learning capabilities analogous to those of large language models. Method: This paper introduces FLoWN, the first framework to incorporate flow matching into neural network parameter generation. FLoWN models conditional parameter flows in latent space, enabling dynamic synthesis of task-specific weights during inference. Contribution/Results: The approach achieves context-aware zero-shot and few-shot parameter generation, while preserving differentiability and cross-distribution generalization. Experiments demonstrate that FLoWN matches or surpasses baseline methods on in-distribution tasks; moreover, it significantly outperforms state-of-the-art approaches on out-of-distribution few-shot classification—delivering superior classifier initialization that directly addresses core meta-learning requirements.
📝 Abstract
Foundational language models show a remarkable ability to learn new concepts during inference via context data. However, similar work for images lag behind. To address this challenge, we introduce FLoWN, a flow matching model that learns to generate neural network parameters for different tasks. Our approach models the flow on latent space, while conditioning the process on context data. Experiments verify that FLoWN attains various desiderata for a meta-learning model. In addition, it matches or exceeds baselines on in-distribution tasks, provides better initializations for classifier training, and is performant on out-of-distribution few-shot tasks while having a fine-tuning mechanism to improve performance.