ANS is committed to advancing, fostering, and promoting the development and application of nuclear sciences and technologies to benefit society.
Explore the many uses for nuclear science and its impact on energy, the environment, healthcare, food, and more.
Explore membership for yourself or for your organization.
Conference Spotlight
2026 ANS Annual Conference
May 31–June 3, 2026
Denver, CO|Sheraton Denver
Latest Magazine Issues
Apr 2026
Jan 2026
Latest Journal Issues
Nuclear Science and Engineering
June 2026
Nuclear Technology
March 2026
Fusion Science and Technology
May 2026
Latest News
DOE selects first companies for nuclear launch pad
The Department of Energy’s Office of Nuclear Energy and the National Reactor Innovation Center have announced their first selections for the Nuclear Energy Launch Pad: three companies developing microreactors and one developing fuel supply.
The four companies—Deployable Energy, General Matter, NuCube Energy, and Radiant Industries—were selected from the initial pool of Reactor Pilot Program and Fuel Line Pilot Program applicants, the two precursor programs to the launch pad.
Bernadette L. Kirk, Yousry Y. Azmy
Nuclear Science and Engineering | Volume 111 | Number 1 | May 1992 | Pages 57-65
Technical Paper | doi.org/10.13182/NSE92-A23923
Articles are hosted by Taylor and Francis Online.
The one-group, steady-state neutron diffusion equation in two-dimensional Cartesian geometry is solved using the nodal integral method. The discrete variable equations comprise loosely coupled sets of equations representing the nodal balance of neutrons, as well as neutron current continuity along rows or columns of computational cells. An iterative algorithm that is more suitable for solving large problems concurrently is derived based on the decomposition of the spatial domain and is accelerated using successive overrelaxation. This algorithm is very well suited for parallel computers, especially since the spatial domain decomposition occurs naturally, so that the number of iterations required for convergence does not depend on the number of processors participating in the calculation. Implementation of our algorithm on the Intel iPSC/2 hypercube and Sequent Balance 8000 parallel computers is presented, and measured speedup and efficiency for test problems are reported. The results suggest that the efficiency of the hypercube quickly deteriorates when many processors are used, while the Sequent Balance retains very high efficiency for a comparable number of participating processors. This leads to the conjecture that message-passing parallel computers are not as well suited for this algorithm as shared-memory machines.