Abstract
AbstractThis paper studies the applicability of a deep reinforcement learning approach to three different multi-echelon inventory systems, with the objective of minimizing the holding and backorder costs. First, we conduct an extensive literature review to map the current applications of reinforcement learning in multi-echelon inventory systems. Next, we apply our deep reinforcement learning method to three cases with different network structures (linear, divergent, and general structures). The linear and divergent cases are derived from literature, whereas the general case is based on a real-life manufacturer. We apply the proximal policy optimization (PPO) algorithm, with a continuous action space, and show that it consistently outperforms the benchmark solution. It achieves an average improvement of 16.4% for the linear case, 11.3% for the divergent case, and 6.6% for the general case. We explain the limitations of our approach and propose avenues for future research.
Publisher
Springer Science and Business Media LLC
Subject
Management Science and Operations Research
Reference77 articles.
1. Arnold J, Chapman S, Clive L (2008) Introduction to materials management. Pearson Prentice Hall, Hoboken
2. Arulkumaran K, Deisenroth MP, Brundage M, Bharath AA (2017) Deep reinforcement learning: a brief survey. IEEE Signal Process Mag 34(6):26–38. https://doi.org/10.1109/MSP.2017.2743240
3. Bellman R (1957) Dynamic programming. Princeton University Press, Princeton
4. Ben-Tal A, Golany B, Shtern S (2009) Robust multi-echelon multi-period inventory control. Eur J Oper Res 199(3):922–935. https://doi.org/10.1016/j.ejor.2009.01.058
5. Bertsimas D, Thiele A (2006) A robust optimization approach to inventory theory. Oper Res 54(1):150–168. https://doi.org/10.1287/opre.1050.0238
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献