Speeding up Policy Simulation in Supply Chain RL
Visualitza/Obre
Data de publicació
2025-11ISSN
2640-3498
Resum
Simulating a single trajectory of a dynamical system under some state-dependent policy is a core bottleneck in policy optimization (PO) algorithms. The many inherently serial policy evaluations that must be performed in a single simulation constitute the bulk of this bottleneck. In applying PO to supply chain optimization (SCO) problems, simulating a single sample path corresponding to one month of a supply chain can take several hours. We present an iterative algorithm to accelerate policy simulation, dubbed Picard Iteration. This scheme carefully assigns policy evaluation tasks to independent processes. Within an iteration, any given process evaluates the policy only on its assigned tasks while assuming a certain "cached" evaluation for other tasks; the cache is updated at the end of the iteration. Implemented on GPUs, this scheme admits batched evaluation of the policy across a single trajectory. We prove that the structure afforded by many SCO problems allows convergence in a small number of iterations independent of the horizon. We demonstrate practical speedups of 400x on large-scale SCO problems even with a single GPU, and also demonstrate practical efficacy in other RL environments.
Tipus de document
Objecte de conferència
Versió del document
Versió publicada
Llengua
Anglès
Pàgines
16 p.
Publicat per
ML Research Press
Publicat a
Proceedings of Machine Learning Research, 42nd International Conference on Machine Learning, ICML 2025 - Vancouver, Canadá
Enllaç al document relacionat
Citació recomanada
Aquesta citació s'ha generat automàticament.
Aquest element apareix en la col·lecció o col·leccions següent(s)
Drets
© L'autor/a
Excepte que s'indiqui una altra cosa, la llicència de l'ítem es descriu com http://creativecommons.org/licenses/by-nc-nd/4.0/


