🤖 AI Summary
This work proposes an information-theoretic iterative learning model predictive control framework to address the integrated demands of safety, robustness, and high performance in robotic iterative tasks operating under complex and uncertain environments. The approach leverages historical trajectories to learn a value function, employs normalizing flows to model non-Gaussian uncertainties, and incorporates an adaptive safety penalty mechanism to handle infinite-horizon constrained optimization for nonlinear stochastic systems. Efficient real-time solutions are achieved through highly parallelized GPU implementation. Both simulation and hardware experiments demonstrate that the system consistently improves control performance across iterations while rigorously satisfying safety constraints, thereby validating the effectiveness and superiority of the proposed method.
📝 Abstract
Robots executing iterative tasks in complex, uncertain environments require control strategies that balance robustness, safety, and high performance. This paper introduces a safe information-theoretic learning model predictive control (SIT-LMPC) algorithm for iterative tasks. Specifically, we design an iterative control framework based on an information-theoretic model predictive control algorithm to address a constrained infinite-horizon optimal control problem for discrete-time nonlinear stochastic systems. An adaptive penalty method is developed to ensure safety while balancing optimality. Trajectories from previous iterations are utilized to learn a value function using normalizing flows, which enables richer uncertainty modeling compared to Gaussian priors. SIT-LMPC is designed for highly parallel execution on graphics processing units, allowing efficient real-time optimization. Benchmark simulations and hardware experiments demonstrate that SIT-LMPC iteratively improves system performance while robustly satisfying system constraints.