论文标题
最佳PID和反翼型控制设计作为增强学习问题
Optimal PID and Antiwindup Control Design as a Reinforcement Learning Problem
论文作者
论文摘要
深度强化学习(DRL)已经看到了几个成功的应用程序来处理控制。常见方法依靠深度神经网络结构来对控制器或过程进行建模。随着越来越复杂的控制结构,这种方法的闭环稳定性变得不太清楚。在这项工作中,我们关注DRL控制方法的解释性。特别是,我们将线性固定结构控制器视为嵌入参与者批评框架中的浅神经网络。 PID控制器由于其在工业实践中的简单性和接受而指导我们的发展。然后,我们考虑输入饱和度,导致简单的非线性控制结构。为了在执行器限制内有效运行,我们将调谐参数合并以进行反打印补偿。最后,控制器的简单性允许直接初始化。这使得我们的方法在训练期间和之后都可以固有地稳定,并且可以接受已知的操作PID增益。
Deep reinforcement learning (DRL) has seen several successful applications to process control. Common methods rely on a deep neural network structure to model the controller or process. With increasingly complicated control structures, the closed-loop stability of such methods becomes less clear. In this work, we focus on the interpretability of DRL control methods. In particular, we view linear fixed-structure controllers as shallow neural networks embedded in the actor-critic framework. PID controllers guide our development due to their simplicity and acceptance in industrial practice. We then consider input saturation, leading to a simple nonlinear control structure. In order to effectively operate within the actuator limits we then incorporate a tuning parameter for anti-windup compensation. Finally, the simplicity of the controller allows for straightforward initialization. This makes our method inherently stabilizing, both during and after training, and amenable to known operational PID gains.