🤖 AI Summary
This work addresses the low efficiency and weak coupling between geometry and motion parameters in digital twin modeling of articulated objects. We propose an end-to-end joint reconstruction method leveraging a 3D multimodal large language model. Our key innovations include: (i) a dedicated [SEG] token mechanism that enables point-cloud-feature-driven fine-grained part segmentation and consistent motion parameter prediction; and (ii) an autoregressive framework integrating point cloud and textual inputs to jointly optimize geometric and kinematic representations. Evaluated on both synthetic and real-world datasets, our method significantly outperforms prior approaches—achieving a 17% improvement in segmentation mIoU, a 29% reduction in motion parameter error, and a 50% increase in physical executability—while demonstrating strong cross-object generalization. The method directly supports robot simulation training and embodied AI world model construction.
📝 Abstract
Constructing accurate digital twins of articulated objects is essential for robotic simulation training and embodied AI world model building, yet historically requires painstaking manual modeling or multi-stage pipelines. In this work, we propose extbf{URDF-Anything}, an end-to-end automatic reconstruction framework based on a 3D multimodal large language model (MLLM). URDF-Anything utilizes an autoregressive prediction framework based on point-cloud and text multimodal input to jointly optimize geometric segmentation and kinematic parameter prediction. It implements a specialized $[SEG]$ token mechanism that interacts directly with point cloud features, enabling fine-grained part-level segmentation while maintaining consistency with the kinematic parameter predictions. Experiments on both simulated and real-world datasets demonstrate that our method significantly outperforms existing approaches regarding geometric segmentation (mIoU 17% improvement), kinematic parameter prediction (average error reduction of 29%), and physical executability (surpassing baselines by 50%). Notably, our method exhibits excellent generalization ability, performing well even on objects outside the training set. This work provides an efficient solution for constructing digital twins for robotic simulation, significantly enhancing the sim-to-real transfer capability.