Optimal control for constrained discrete-time nonlinear systems based on safe reinforcement learning

  • L Zhang
  • , Lei Xie
  • , Y Jiang
  • , Z Li
  • , Xueqin Liu
  • , H. Su

Research output: Contribution to journalArticlepeer-review

13 Citations (Scopus)

Abstract

The state and input constraints of nonlinear systems could greatly impede the realization of their optimal control when using reinforcement learning (RL)-based approaches since the commonly used quadratic utility functions cannot meet the requirements of solving constrained optimization problems. This article develops a novel optimal control approach for constrained discrete-time (DT) nonlinear systems based on safe RL. Specifically, a barrier function (BF) is introduced and incorporated with the value function to help transform a constrained optimization problem into an unconstrained one. Meanwhile, the minimum of such an optimization problem can be guaranteed to occur at the origin. Then a constrained policy iteration (PI) algorithm is developed to realize the optimal control of the nonlinear system and to enable the state and input constraints to be satisfied. The constrained optimal control policy and its corresponding value function are derived through the implementation of two neural networks (NNs). Performance analysis shows that the proposed control approach still retains the convergence and optimality properties of the traditional PI algorithm. Simulation results of three examples reveal its effectiveness.
Original languageEnglish
Number of pages12
JournalIEEE Transactions on Neural Networks and Learning Systems
Early online date31 Oct 2023
DOIs
Publication statusEarly online date - 31 Oct 2023

Fingerprint

Dive into the research topics of 'Optimal control for constrained discrete-time nonlinear systems based on safe reinforcement learning'. Together they form a unique fingerprint.

Cite this