Show simple item record

dc.contributor.authorSeurin, Paul
dc.contributor.authorShirvan, Koroush
dc.date.accessioned2025-06-24T15:28:25Z
dc.date.available2025-06-24T15:28:25Z
dc.date.issued2024-01-29
dc.identifier.urihttps://hdl.handle.net/1721.1/159669
dc.description.abstractThe nuclear fuel loading pattern optimization problem belongs to the class of large-scale combinatorial optimization. It is also characterized by multiple objectives and constraints, which makes it impossible to solve explicitly. Stochastic optimization methodologies including Genetic Algorithms and Simulated Annealing are used by different nuclear utilities and vendors but hand-designed solutions continue to be the prevalent method in the industry. To improve the state-of-the-art, Deep Reinforcement Learning (RL), in particular, Proximal Policy Optimization is leveraged. This work presents a first-of-a-kind approach to utilize deep RL to solve the loading pattern problem and could be leveraged for any engineering design optimization. This paper is also to our knowledge the first to propose a study of the behavior of several hyper-parameters that influence the RL algorithm. The algorithm is highly dependent on multiple factors such as the shape of the objective function derived for the core design that behaves as a fudge factor that affects the stability of the learning. But also an exploration/exploitation trade-off that manifests through different parameters such as the number of loading patterns seen by the agents per episode, the number of samples collected before a policy update , and an entropy factor that increases the randomness of the policy during training. We found that RL must be applied similarly to a Gaussian Process in which the acquisition function is replaced by a parametrized policy. Then, once an initial set of hyper-parameters is found, reducing and until no more learning is observed will result in the highest sample efficiency robustly and stably. This resulted in an economic benefit of 535,000 - 642,000 $/year/plant. Future work must extend this research to multi-objective settings and comparing them to state-of-the-art implementation of stochastic optimization methods.en_US
dc.publisherSpringer USen_US
dc.relation.isversionofhttps://doi.org/10.1007/s10489-023-05013-5en_US
dc.rightsCreative Commons Attribution-Noncommercial-ShareAlikeen_US
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/4.0/en_US
dc.sourceSpringer USen_US
dc.titleAssessment of reinforcement learning algorithms for nuclear power plant fuel optimizationen_US
dc.typeArticleen_US
dc.identifier.citationSeurin, P., Shirvan, K. Assessment of reinforcement learning algorithms for nuclear power plant fuel optimization. Appl Intell 54, 2100–2135 (2024).en_US
dc.contributor.departmentMassachusetts Institute of Technology. Department of Nuclear Science and Engineeringen_US
dc.relation.journalApplied Intelligenceen_US
dc.eprint.versionAuthor's final manuscripten_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-03-27T13:48:26Z
dc.language.rfc3066en
dc.rights.holderThe Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature
dspace.embargo.termsY
dspace.date.submission2025-03-27T13:48:26Z
mit.journal.volume54en_US
mit.licenseOPEN_ACCESS_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record