ANS is committed to advancing, fostering, and promoting the development and application of nuclear sciences and technologies to benefit society.
Explore the many uses for nuclear science and its impact on energy, the environment, healthcare, food, and more.
Explore membership for yourself or for your organization.
Conference Spotlight
2026 ANS Annual Conference
May 31–June 3, 2026
Denver, CO|Sheraton Denver
Latest Magazine Issues
Mar 2026
Jan 2026
Latest Journal Issues
Nuclear Science and Engineering
April 2026
Nuclear Technology
February 2026
Fusion Science and Technology
Latest News
Going Nuclear: Notes from the officially unofficial book tour
I work in the analytical labs at one of Europe’s oldest and largest nuclear sites: Sellafield, in northwestern England. I spend my days at the fume hood front, pipette in one hand and radiation probe in the other (and dosimeter pinned to my chest, of course). Outside the lab, I have a second job: I moonlight as a writer and public speaker. My new popular science book—Going Nuclear: How the Atom Will Save the World—came out last summer, and it feels like my life has been running at full power ever since.
Majdi I. Radaideh, Leo Tunkle, Dean Price, Kamal Abdulraheem, Linyu Lin, Moutaz Elias
Nuclear Science and Engineering | Volume 200 | Number 1 | March 2026 | Pages S309-S321
Research Article | doi.org/10.1080/00295639.2024.2447012
Articles are hosted by Taylor and Francis Online.
Reducing operation and maintenance costs is a key objective for advanced reactors in general and microreactors in particular. To achieve this reduction, developing robust autonomous control algorithms is essential to ensure safe and autonomous reactor operation. Recently, artificial intelligence and machine learning algorithms, specifically reinforcement learning (RL) algorithms, have seen rapid increased application to control problems, such as plasma control in fusion tokamaks and building energy management. In this work, we introduce the use of RL for intelligent control in nuclear microreactors. The RL agent is trained using Proximal Policy Optimization (PPO) and Advantage Actor-Critic (A2C), cutting-edge deep RL techniques, based on a high-fidelity simulation of a microreactor design inspired by the Westinghouse eVinciTM design. We utilized a Serpent model to generate data on drum positions, core criticality, and core power distribution for training a feedforward neural network surrogate model. This surrogate model was then used to guide a PPO and A2C control policies in determining the optimal drum position across various reactor burnup states, ensuring critical core conditions and symmetrical power distribution across all six core portions. The results demonstrate the excellent performance of PPO in identifying optimal drum positions, achieving a hexant power tilt ratio of approximately 1.002 (within the limit of 1.02), and maintaining criticality within a 10 pcm range. A2C did not provide as competitive of a performance as PPO in terms of performance metrics for all burnup steps considered in the cycle. Additionally, the results highlight the capability of well-trained RL control policies to quickly identify control actions, suggesting a promising approach for enabling real-time autonomous control through digital twins.