TACT: Humanoid Whole-body Contact Manipulation through Deep Imitation Learning with Tactile Modality

πŸ“… 2025-06-18
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the high computational cost of motion generation and the difficulty of large-area tactile sensing in whole-body contact manipulation for humanoid robots, this paper proposes a deep imitation learning framework integrating high-resolution tactile modalities. Methodologically, it introduces, for the first time, deep integration of tactile signals into the Action Chunking with Transformers (ACT) architecture, jointly leveraging vision and proprioception; it further incorporates whole-body motion retargeting and real-time bipedal dynamic gait–balance control to enable stable contact manipulation during locomotion. Contributions include: (1) establishing the first imitation learning paradigm supporting high-density, whole-body tactile feedback; and (2) experimental validation on the full-scale RHP7 Kaleido humanoid platform, demonstrating significant improvements in robustness and task success rates for both large-area and fine-contact manipulation tasks.

Technology Category

Application Category

πŸ“ Abstract
Manipulation with whole-body contact by humanoid robots offers distinct advantages, including enhanced stability and reduced load. On the other hand, we need to address challenges such as the increased computational cost of motion generation and the difficulty of measuring broad-area contact. We therefore have developed a humanoid control system that allows a humanoid robot equipped with tactile sensors on its upper body to learn a policy for whole-body manipulation through imitation learning based on human teleoperation data. This policy, named tactile-modality extended ACT (TACT), has a feature to take multiple sensor modalities as input, including joint position, vision, and tactile measurements. Furthermore, by integrating this policy with retargeting and locomotion control based on a biped model, we demonstrate that the life-size humanoid robot RHP7 Kaleido is capable of achieving whole-body contact manipulation while maintaining balance and walking. Through detailed experimental verification, we show that inputting both vision and tactile modalities into the policy contributes to improving the robustness of manipulation involving broad and delicate contact.
Problem

Research questions and friction points this paper is trying to address.

Develop humanoid control for whole-body contact manipulation
Address challenges in motion generation and contact measurement
Integrate vision and tactile sensors for robust manipulation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deep imitation learning with tactile sensors
Multi-sensor input including vision and touch
Integration with retargeting and locomotion control
Masaki Murooka
Masaki Murooka
National Institute of Advanced Industrial Science and Technology
Robotics
T
Takahiro Hoshi
Tokyo University of Science
K
Kensuke Fukumitsu
Tokyo University of Science
S
Shimpei Masuda
CNRS-AIST JRL (Joint Robotics Laboratory), IRL and National Institute of Advanced Industrial Science and Technology (AIST), University of Tsukuba
Marwan Hamze
Marwan Hamze
Researcher in the Control of Robotics
RoboticsControl TheoryReinforcement Learning
Tomoya Sasaki
Tomoya Sasaki
Tokyo University of Science
Human AugmentationVirtual RealityRoboticsHapticsEntertainment Computing
M
M. Morisawa
CNRS-AIST JRL (Joint Robotics Laboratory), IRL and National Institute of Advanced Industrial Science and Technology (AIST)
Eiichi Yoshida
Eiichi Yoshida
Tokyo University of Science
robotics