ANS is committed to advancing, fostering, and promoting the development and application of nuclear sciences and technologies to benefit society.
Explore the many uses for nuclear science and its impact on energy, the environment, healthcare, food, and more.
Explore membership for yourself or for your organization.
Conference Spotlight
2026 ANS Annual Conference
May 31–June 3, 2026
Denver, CO|Sheraton Denver
Latest Magazine Issues
Mar 2026
Jan 2026
Latest Journal Issues
Nuclear Science and Engineering
April 2026
Nuclear Technology
February 2026
Fusion Science and Technology
Latest News
IAEA project aims to develop polymer irradiation model
The International Atomic Energy Agency has launched a new coordinated research project (CRP) aimed at creating a database of polymer-radiation interactions in the next five years with the long-term goal of using the database to enable machine learning–based predictive models.
Radiation-induced modifications are widely applicable across a range of fields including healthcare, agriculture, and environmental applications, and exposure to radiation is a major factor when considering materials used at nuclear power plants.
Aidan Rigby, Mike Wagner, Daniel Mikkelson, Ben Lindley
Nuclear Technology | Volume 212 | Number 2 | February 2026 | Pages 427-445
Regular Research Article | doi.org/10.1080/00295450.2025.2466137
Articles are hosted by Taylor and Francis Online.
The ability of nuclear reactors to operate their power conversion cycles more flexibly will enhance their value to energy grids with variable pricing. Current nuclear control systems are typically classical controllers that are often based on proportional-integral-derivative (PID) control. This paper presents a method of augmenting the existing PID control for difficult transient operations in nuclear power plants using a reinforcement learning–derived feedforward signal applied in real time. The agents, which are trained on a test thermal load-following problem, are designed to improve steam generator outlet temperature control for a range of fast load-following scenarios covering ramp rates from 9%/min to 15%/min.
Several reinforcement learning algorithms were initially investigated for the training of the feedforward agents with deep Q-learning (DQN) and proximal policy optimization (PPO) networks, which were found to be the most promising. The DQN controllers utilize discrete actions, giving them a better disturbance rejection at steady state but inconsistent response to initial temperature deviations. In contrast, PPO-trained agents, which take continuous actions except for a dead zone around zero, were shown to have the best combination of high disturbance rejection at steady state and good tracking of the desired temperature value. The ability of the PPO agent was also examined, with the average time of decision making found to be on the order of 1 ms.
The fault properties of the controller under the loss of the reinforcement learning agent feedforward signal were also examined. The controller showed strong performance in situations of “no-signal” faults. but was less good at handling “stuck-at” faults, where the feedforward signal remains at a set value. In both cases, however, the PID was able to successfully maintain stability, eventually returning the system to a steady state. It is hoped that this work will allow for the proposed control architecture to be examined for more difficult control problems such that it may eventually be used to adapt existing nuclear plants for more aggressive load-following on grids of the future.