Summary
This paper studies the robust stabilization problem of constrained‐input nonlinear systems with mismatched uncertainties. Initially, the robust stabilization problem is converted into a constrained H2 optimal control problem by providing a proper infinite‐horizon cost function for the auxiliary system. It is demonstrated that the solution of the constrained H2 optimal control problem can force the original system to be stable in the sense of uniform ultimate boundedness. Then, under the framework of reinforcement learning, an off‐policy iteration algorithm is proposed to solve the constrained H2 optimal control problem. The off‐policy iteration algorithm is implemented by using two kinds of approximators. That is, the critic approximator is applied to approximate the optimal cost function and the actor approximator is used to approximate the augmented optimal control. The method of weighted residuals together with the Monte‐Carlo integration technique is employed to determine the weight parameters of critic and actor approximators. Finally, two examples, including a pendulum system, are presented to validate the proposed control algorithm.