Adapting World Models with Latent-State Dynamics Residuals

📅 2025-04-03
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address policy transfer failure in simulation-to-reality (Sim2Real) reinforcement learning caused by dynamics mismatch, this paper proposes a latent-space residual calibration method. Instead of modeling state-transition residuals in high-dimensional pixel space—where pixel-level reconstruction is challenging—it learns residuals in a low-dimensional latent space. The approach integrates an autoregressive latent world model, simulation-based pretraining, and fine-tuning on minimal real-world data, enabling imagination-based rollouts under calibrated dynamics for policy optimization. Its core contribution is the first formulation of residual correction in latent space, achieving robust dynamics adaptation with only a small amount of real interaction data. Experiments demonstrate significant improvements in cross-domain generalization across multiple visual MuJoCo benchmarks and a real-world robotic vision-based lane-following task, outperforming state-of-the-art Sim2Real transfer methods.

Technology Category

Application Category

📝 Abstract
Simulation-to-reality reinforcement learning (RL) faces the critical challenge of reconciling discrepancies between simulated and real-world dynamics, which can severely degrade agent performance. A promising approach involves learning corrections to simulator forward dynamics represented as a residual error function, however this operation is impractical with high-dimensional states such as images. To overcome this, we propose ReDRAW, a latent-state autoregressive world model pretrained in simulation and calibrated to target environments through residual corrections of latent-state dynamics rather than of explicit observed states. Using this adapted world model, ReDRAW enables RL agents to be optimized with imagined rollouts under corrected dynamics and then deployed in the real world. In multiple vision-based MuJoCo domains and a physical robot visual lane-following task, ReDRAW effectively models changes to dynamics and avoids overfitting in low data regimes where traditional transfer methods fail.
Problem

Research questions and friction points this paper is trying to address.

Addressing simulation-to-reality dynamics discrepancies in RL
Correcting latent-state dynamics residuals for high-dimensional states
Enabling RL agent adaptation in low-data real-world tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Latent-state dynamics residuals for simulation-to-reality transfer
Autoregressive world model pretrained in simulation
Corrected dynamics for RL optimization and deployment
🔎 Similar Papers
No similar papers found.
JB Lanier
JB Lanier
UC Irvine
Reinforcement LearningMultiagent SystemsGame TheorySim-to-real
K
Kyungmin Kim
Department of Computer Science, University of California Irvine
Armin Karamzade
Armin Karamzade
University of California, Irvine
Reinforcement LearningMachine LearningArtificial Intelligence
Y
Yifei Liu
Department of Computer Science, University of California Irvine
A
Ankita Sinha
Department of Computer Science, University of California Irvine
K
Kat He
Department of Computer Science, University of California Irvine
Davide Corsi
Davide Corsi
University of California, Irvine
Reinforcement LearningRoboticsSafety
Roy Fox
Roy Fox
Assistant Professor, UC Irvine
Reinforcement LearningAlgorithmic Game TheoryInformation TheoryRobot learning