🤖 AI Summary
To address the poor generalization and deployment challenges of multi-task model-based reinforcement learning (MBRL) in resource-constrained settings, this paper proposes a knowledge transfer framework tailored for lightweight agents. We first compress a high-capacity multi-task world model (317M parameters) into an ultra-compact agent (1M parameters) via knowledge distillation, then apply FP16 post-training quantization to further reduce model size by 50%. Evaluated on the MT30 multi-task benchmark, our method achieves a normalized score of 28.45—improving upon the original 1M baseline by 50.5% and establishing a new state-of-the-art (SOTA). The model size is reduced by 317× with zero performance degradation. Our core contribution lies in establishing an efficient, fidelity-preserving knowledge distillation and compression paradigm that transfers capabilities from large-scale multi-task world models to extremely lightweight MBRL agents—significantly enhancing multi-task generalization and deployment feasibility on edge devices.
📝 Abstract
We propose an efficient knowledge transfer approach for model-based reinforcement learning, addressing the challenge of deploying large world models in resource-constrained environments. Our method distills a high-capacity multi-task agent (317M parameters) into a compact 1M parameter model, achieving state-of-the-art performance on the MT30 benchmark with a normalized score of 28.45, a substantial improvement over the original 1M parameter model's score of 18.93. This demonstrates the ability of our distillation technique to consolidate complex multi-task knowledge effectively. Additionally, we apply FP16 post-training quantization, reducing the model size by 50% while maintaining performance. Our work bridges the gap between the power of large models and practical deployment constraints, offering a scalable solution for efficient and accessible multi-task reinforcement learning in robotics and other resource-limited domains.