Deep Reinforcement Learning Optimization for Uncertain Nonlinear Systems via Event-Triggered Robust Adaptive Dynamic Programming

📅 2025-12-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the robust optimal control problem for uncertain nonlinear systems. We propose a novel control architecture integrating an event-triggered mechanism (ETM), an extended state observer (ESO), and value-iteration-based adaptive dynamic programming (VI-ADP). To our knowledge, this is the first work embedding ETM directly into the ADP framework; the ESO estimates composite disturbances online, enabling disturbance compensation and closed-loop stability without requiring an exact system model. Lyapunov-based stability analysis rigorously guarantees both learning convergence and system robustness. Compared with conventional time-triggered ADP, the proposed method reduces sampling and computational load by over 60%, significantly improving computational efficiency and learning sparsity while maintaining high disturbance rejection capability and control accuracy. Numerical experiments validate the theoretical stability claims and demonstrate superior performance in terms of robustness, efficiency, and precision.

Technology Category

Application Category

📝 Abstract
This work proposes a unified control architecture that couples a Reinforcement Learning (RL)-driven controller with a disturbance-rejection Extended State Observer (ESO), complemented by an Event-Triggered Mechanism (ETM) to limit unnecessary computations. The ESO is utilized to estimate the system states and the lumped disturbance in real time, forming the foundation for effective disturbance compensation. To obtain near-optimal behavior without an accurate system description, a value-iteration-based Adaptive Dynamic Programming (ADP) method is adopted for policy approximation. The inclusion of the ETM ensures that parameter updates of the learning module are executed only when the state deviation surpasses a predefined bound, thereby preventing excessive learning activity and substantially reducing computational load. A Lyapunov-oriented analysis is used to characterize the stability properties of the resulting closed-loop system. Numerical experiments further confirm that the developed approach maintains strong control performance and disturbance tolerance, while achieving a significant reduction in sampling and processing effort compared with standard time-triggered ADP schemes.
Problem

Research questions and friction points this paper is trying to address.

Optimizes control for uncertain nonlinear systems via reinforcement learning.
Reduces computational load using event-triggered adaptive dynamic programming.
Ensures stability and disturbance tolerance with Lyapunov analysis.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Event-triggered mechanism reduces computational load
Extended state observer estimates disturbances in real-time
Adaptive dynamic programming approximates optimal control policy
🔎 Similar Papers
No similar papers found.
N
Ningwei Bai
Department of Computer Science, University College London, Gower Street, London, WC1E 6BT United Kingdom
C
Chi Pui Chan
Department of Computer Science, University College London, Gower Street, London, WC1E 6BT United Kingdom
Q
Qichen Yin
Department of Computer Science, University College London, Gower Street, London, WC1E 6BT United Kingdom
T
Tengyang Gong
Department of Electrical and Electronic Engineering, University of Manchester, Oxford Rd, Manchester, M13 9PL United Kingdom
Yunda Yan
Yunda Yan
UK Lecturer (Assistant Professor), SMIEEE, FHEA, University College London
Model predictive controlReinforcement learningAerial roboticsAutonomous systems
Z
Zezhi Tang
Department of Computer Science, University College London, Gower Street, London, WC1E 6BT United Kingdom