🤖 AI Summary
Efficient and robust trajectory planning and real-time grasp re-planning are critical for robotic manipulation in cluttered environments.
Method: This paper proposes a novel multimodal physics-informed neural network (PINN) framework—the first to integrate PINNs into pre-grasp manipulation—enabling unsupervised, expert-demonstration-free solving of the Eikonal equation. The method jointly models neural time fields, fuses multimodal sensor inputs (RGB-D, point clouds, force-torque), and enforces physical constraints, supporting end-to-end sim-to-real transfer learning.
Contribution/Results: Compared to state-of-the-art approaches, our method achieves 2.3× faster planning, 18% shorter trajectories, and a +12.7% improvement in grasp success rate in complex cluttered scenes. Comprehensive evaluations on both high-fidelity simulation and real-world robotic arms validate its strong generalization capability and practical deployability.
📝 Abstract
Object manipulation skills are necessary for robots operating in various daily-life scenarios, ranging from warehouses to hospitals. They allow the robots to manipulate the given object to their desired arrangement in the cluttered environment. The existing approaches to solving object manipulations are either inefficient sampling based techniques, require expert demonstrations, or learn by trial and error, making them less ideal for practical scenarios. In this paper, we propose a novel, multimodal physics-informed neural network (PINN) for solving object manipulation tasks. Our approach efficiently learns to solve the Eikonal equation without expert data and finds object manipulation trajectories fast in complex, cluttered environments. Our method is multimodal as it also reactively replans the robot's grasps during manipulation to achieve the desired object poses. We demonstrate our approach in both simulation and real-world scenarios and compare it against state-of-the-art baseline methods. The results indicate that our approach is effective across various objects, has efficient training compared to previous learning-based methods, and demonstrates high performance in planning time, trajectory length, and success rates. Our demonstration videos can be found at https://youtu.be/FaQLkTV9knI.