Hierarchical Multi-Agent Reinforcement Learning for Dynamic Inventory Allocation with Demand Uncertainty
DOI:
https://doi.org/10.71465/mrcis153Keywords:
hierarchical reinforcement learning, multi-agent systems, inventory allocation, demand uncertainty, supply chain management, decentralized controlAbstract
The complexity of modern supply chain networks requires sophisticated approaches to inventory management that can effectively handle demand uncertainty and coordinate decisions across multiple organizational levels. This paper proposes a novel hierarchical multi-agent reinforcement learning framework for dynamic inventory allocation in multi-echelon supply chains facing stochastic demand patterns. The hierarchical architecture decomposes the inventory control problem into strategic and operational decision layers, where high-level agents coordinate allocation policies across distribution networks while low-level agents optimize local replenishment decisions. The framework integrates Centralized Training with Decentralized Execution paradigm, enabling autonomous agents to learn coordinated policies through shared experience while maintaining operational independence during deployment. Experimental results demonstrate that the proposed approach achieves significant reductions in total system costs compared to traditional base-stock policies and single-agent reinforcement learning methods, while effectively mitigating the bullwhip effect in supply chains with high demand variability.
Downloads
Published
Issue
Section
License

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
All articles published in the Multidisciplinary Research in Computing Information Systems are licensed under an open-access model. Authors retain full copyright and grant the journal the right of first publication. The content can be freely accessed, distributed, and reused for non-commercial purposes, provided proper citation is given to the original work.
