International Journal of Control, Automation and Systems 2023; 21(4): 1319-1330
Published online March 3, 2023
https://doi.org/10.1007/s12555-021-0674-z
© The International Journal of Control, Automation, and Systems
This paper proposes a robust near-optimal control algorithm for uncertain nonlinear systems with state constraints and input saturation. By incorporating a barrier function and a non-quadratic term, the robust stabilization problem with constraints and uncertainties is converted into an unconstrained optimal control problem of the nominal system, which requires the solution of the Hamilton-Jacobi-Bellman (HJB) equation. The proposed integral reinforcement learning (IRL)-based method can obtain the approximate solution of the HJB equation without requiring any knowledge of system drift dynamics. An online gain-adjustable update law of the actor-critic architecture is developed to relax the persistence of excitation (PE) condition and ensure the closed-loop system stability throughout learning. The uniform ultimate boundedness of the closed-loop system is verified using Lyapunov’s direct method. Simulation results demonstrate the effectiveness and feasibility of the proposed method.
Keywords Constrained nonlinear system, integral reinforcement learning, optimal control, robust control.
International Journal of Control, Automation and Systems 2023; 21(4): 1319-1330
Published online April 1, 2023 https://doi.org/10.1007/s12555-021-0674-z
Copyright © The International Journal of Control, Automation, and Systems.
Yu-Qing Qiu, Yan Li, and Zhong Wang*
Northwestern Polytechnical University
This paper proposes a robust near-optimal control algorithm for uncertain nonlinear systems with state constraints and input saturation. By incorporating a barrier function and a non-quadratic term, the robust stabilization problem with constraints and uncertainties is converted into an unconstrained optimal control problem of the nominal system, which requires the solution of the Hamilton-Jacobi-Bellman (HJB) equation. The proposed integral reinforcement learning (IRL)-based method can obtain the approximate solution of the HJB equation without requiring any knowledge of system drift dynamics. An online gain-adjustable update law of the actor-critic architecture is developed to relax the persistence of excitation (PE) condition and ensure the closed-loop system stability throughout learning. The uniform ultimate boundedness of the closed-loop system is verified using Lyapunov’s direct method. Simulation results demonstrate the effectiveness and feasibility of the proposed method.
Keywords: Constrained nonlinear system, integral reinforcement learning, optimal control, robust control.
Vol. 22, No. 9, pp. 2673~2953
Juhoon Back, Sejin Kang, Hyungbo Shim*, and Jin Heon Seo
International Journal of Control, Automation and Systems 2008; 6(3): 364-377Zhihui Wu*, Guo-Ping Liu, June Hu, Hui Yu, and Dongyan Chen
International Journal of Control, Automation, and Systems 2024; 22(9): 2699-2710Liuliu Zhang*, Hailong Jing, Cheng Qian, and Changchun Hua
International Journal of Control, Automation, and Systems 2024; 22(8): 2472-2482