IDEAS home Printed from https://ideas.repec.org/a/gam/jeners/v18y2025i6p1517-d1615560.html
   My bibliography  Save this article

A Multi-Variable Coupled Control Strategy Based on a Deep Deterministic Policy Gradient Reinforcement Learning Algorithm for a Small Pressurized Water Reactor

Author

Listed:
  • Jie Chen

    (National Key Laboratory of Nuclear Reactor Technology, Nuclear Power Institute of China, Chengdu 610213, China)

  • Kai Xiao

    (National Key Laboratory of Nuclear Reactor Technology, Nuclear Power Institute of China, Chengdu 610213, China)

  • Ke Huang

    (National Key Laboratory of Nuclear Reactor Technology, Nuclear Power Institute of China, Chengdu 610213, China)

  • Zhen Yang

    (National Key Laboratory of Nuclear Reactor Technology, Nuclear Power Institute of China, Chengdu 610213, China)

  • Qing Chu

    (National Key Laboratory of Nuclear Reactor Technology, Nuclear Power Institute of China, Chengdu 610213, China)

  • Guanfu Jiang

    (National Key Laboratory of Nuclear Reactor Technology, Nuclear Power Institute of China, Chengdu 610213, China)

Abstract

The reactor system has multivariate, nonlinear, and strongly coupled dynamic characteristics, which puts high demands on the robustness, real-time demand, and accuracy of the control strategy. Conventional control approaches depend on the mathematical model of the system being controlled, making it challenging to handle the reactor system’s dynamic complexity and uncertainties. This paper proposes a multi-variable coupled control strategy for a nuclear reactor steam supply system based on a Deep Deterministic Policy Gradient reinforcement learning algorithm, designs and trains a multi-variable coupled intelligent controller to simultaneously realize the coordinated control of multiple parameters, such as the reactor power, average coolant temperature, steam pressure, etc., and performs a simulation validation of the control strategy under the typical transient variable load working conditions. Simulation results show that the reinforcement learning control effect is better than the PID control effect under a ±10% FP step variable load condition, a linear variable load condition, and a load dumping condition, and that the reactor power overshooting amount and regulation time, the maximum deviation of the coolant average temperature, the steam pressure, the pressure of pressurizer and relative liquid level, and the regulation time are improved by at least 15.5% compared with the traditional control method. Therefore, this study offers a theoretical framework for utilizing reinforcement learning in the field of nuclear reactor control.

Suggested Citation

  • Jie Chen & Kai Xiao & Ke Huang & Zhen Yang & Qing Chu & Guanfu Jiang, 2025. "A Multi-Variable Coupled Control Strategy Based on a Deep Deterministic Policy Gradient Reinforcement Learning Algorithm for a Small Pressurized Water Reactor," Energies, MDPI, vol. 18(6), pages 1-26, March.
  • Handle: RePEc:gam:jeners:v:18:y:2025:i:6:p:1517-:d:1615560
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1996-1073/18/6/1517/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1996-1073/18/6/1517/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jeners:v:18:y:2025:i:6:p:1517-:d:1615560. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.