摘要:AbstractThere has been an increasing interest in explicit and cheap-to-evaluate control policies that approximate (computationally expensive) control laws such as model predictive control (MPC). However, approximate control policies are subject to approximation errors, leading to asymptotic performance losses. The contribution of this paper is three-fold: (i) a closed-loop training scheme is presented for deep neural network approximation of economic MPC; (ii) an online adaptive correction scheme is presented to account for the performance losses induced by approximation errors; and (iii) an offline performance verification scheme is presented to ensure that the approximate control policy converges to an equilibrium point of the system. The proposed approach is illustrated using a Williams-Otto reactor problem.