🤖 AI Summary
Reinforcement learning controllers exhibit poor generalization when dynamics differ between training and testing environments. To address this, we propose a model-driven framework grounded in dimensionless Markov decision processes (Π-MDPs), the first to systematically incorporate the Buckingham-Π theorem into MDP modeling. Our approach applies physical dimension normalization to both state and action spaces, endowing learned policies with equivariance and intrinsic robustness to variations in physical parameters—such as gravity, mass, and length. The framework integrates context-aware MDPs, Gaussian process dynamics modeling, and model-based policy search. Evaluated on simulated inverted pendulum and cart-pole tasks, policies trained in a single environment maintain stable performance under large-scale physical parameter shifts—e.g., ±50% variation in gravity or mass—demonstrating significantly improved generalization over baseline methods.
📝 Abstract
Controllers trained with Reinforcement Learning tend to be very specialized and thus generalize poorly when their testing environment differs from their training one. We propose a Model-Based approach to increase generalization where both world model and policy are trained in a dimensionless state-action space. To do so, we introduce the Dimensionless Markov Decision Process ($Pi$-MDP): an extension of Contextual-MDPs in which state and action spaces are non-dimensionalized with the Buckingham-$Pi$ theorem. This procedure induces policies that are equivariant with respect to changes in the context of the underlying dynamics. We provide a generic framework for this approach and apply it to a model-based policy search algorithm using Gaussian Process models. We demonstrate the applicability of our method on simulated actuated pendulum and cartpole systems, where policies trained on a single environment are robust to shifts in the distribution of the context.