Learning Dexterous Bimanual Catch Skills through Adversarial-Cooperative Heterogeneous-Agent Reinforcement Learning

๐Ÿ“… 2025-02-17
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Bimanual robotic grasping suffers from poor hand coordination and limited generalization across diverse objects. Method: We propose an adversarial-cooperative heterogeneous-agent reinforcement learning framework. It comprises two specialized agents: a โ€œthrowingโ€ agent that dynamically increases task difficulty, and a โ€œcatchingโ€ agent that learns adaptive bimanual coordination policies. We introduce a novel adversarial reward mechanism that drives progressive task difficulty escalation, and train the framework in PyBullet/Mujoco simulation environments for multi-object generalization. Contribution/Results: This work is the first to integrate heterogeneous agents and adversarial rewards into bimanual throwing-and-catching tasks, significantly improving control flexibility and robustness. Our method achieves stable grasping across 15 object classes with varying shapes and sizes. The average catching reward improves by approximately 100% over single-agent baselines, demonstrating substantial advances in both task complexity handling and cross-object generalization capability.

Technology Category

Application Category

๐Ÿ“ Abstract
Robotic catching has traditionally focused on single-handed systems, which are limited in their ability to handle larger or more complex objects. In contrast, bimanual catching offers significant potential for improved dexterity and object handling but introduces new challenges in coordination and control. In this paper, we propose a novel framework for learning dexterous bimanual catching skills using Heterogeneous-Agent Reinforcement Learning (HARL). Our approach introduces an adversarial reward scheme, where a throw agent increases the difficulty of throws-adjusting speed-while a catch agent learns to coordinate both hands to catch objects under these evolving conditions. We evaluate the framework in simulated environments using 15 different objects, demonstrating robustness and versatility in handling diverse objects. Our method achieved approximately a 2x increase in catching reward compared to single-agent baselines across 15 diverse objects.
Problem

Research questions and friction points this paper is trying to address.

Bimanual robotic catching
Coordination and control challenges
Adversarial-cooperative learning framework
Innovation

Methods, ideas, or system contributions that make the work stand out.

Heterogeneous-Agent Reinforcement Learning
Adversarial reward scheme
Bimanual coordination control
๐Ÿ”Ž Similar Papers
No similar papers found.
T
Taewoo Kim
Social Robotics Research Section, Electronics and Telecommunications Research Institute (ETRI), Daejeon, Republic of Korea
Youngwoo Yoon
Youngwoo Yoon
ETRI
Human-Robot InteractionHuman-Computer InteractionMachine Learning
Jaehong Kim
Jaehong Kim
Yale University
Environmental EngineeringWater TreatmentEnvironmental NanotechnologyDisinfectionMembrane