🤖 AI Summary
To address the limited generalization and interaction accuracy of Vision-Language-Action (VLA) models in high-precision robotic manipulation tasks, this work introduces— for the first time—human motor learning principles to devise a three-dimensional post-training taxonomy encompassing environmental perception, proprioceptive awareness, and task comprehension. Methodologically, we integrate vision-language model enhancement, action-generation fine-tuning, environment-feedback alignment, multimodal instruction optimization, and embodied perception integration, trained iteratively on both simulated and real-world data. Our key contributions are: (1) the first systematic, motor-learning-theory–grounded taxonomy for VLA post-training; (2) a scalable, structured optimization framework; and (3) an open-source repository including datasets, code, and standardized evaluation protocols. Experiments demonstrate substantial improvements in both precision and cross-scenario adaptability for complex embodied manipulation tasks.
📝 Abstract
Vision-language-action (VLA) models extend vision-language models (VLM) by integrating action generation modules for robotic manipulation. Leveraging strengths of VLM in vision perception and instruction understanding, VLA models exhibit promising generalization across diverse manipulation tasks. However, applications demanding high precision and accuracy reveal performance gaps without further adaptation. Evidence from multiple domains highlights the critical role of post-training to align foundational models with downstream applications, spurring extensive research on post-training VLA models. VLA model post-training aims to address the challenge of improving an embodiment's ability to interact with the environment for the given tasks, analogous to the process of humans motor skills acquisition. Accordingly, this paper reviews post-training strategies for VLA models through the lens of human motor learning, focusing on three dimensions: environments, embodiments, and tasks. A structured taxonomy is introduced aligned with human learning mechanisms: (1) enhancing environmental perception, (2) improving embodiment awareness, (3) deepening task comprehension, and (4) multi-component integration. Finally, key challenges and trends in post-training VLA models are identified, establishing a conceptual framework to guide future research. This work delivers both a comprehensive overview of current VLA model post-training methods from a human motor learning perspective and practical insights for VLA model development. (Project website: https://github.com/AoqunJin/Awesome-VLA-Post-Training)