系统仿真学报 ›› 2019, Vol. 31 ›› Issue (10): 2155-2163.doi: 10.16182/j.issn1004731x.joss.18-0820

• 仿真应用工程 • 上一篇    下一篇

基于深度强化学习的动态库存路径优化

周建频1, 张姝柳2   

  1. 1. 集美大学航海学院,厦门 361021;
    2. 国网吉林供电公司,吉林 132000
  • 收稿日期:2018-12-10 修回日期:2019-02-15 出版日期:2019-10-10 发布日期:2019-12-12
  • 作者简介:周建频(1968-),男,福建,博士,副教授,研究方向为人工智能与供应链系统仿真; 张姝柳(1989-),女,吉林,硕士,助理工程师,研究方向为电气工程与项目管理。
  • 基金资助:
    福建省自然科学基金(2017J01797,2017J01796)

Dynamic Inventory Routing Optimization Based on Deep Reinforcement Learning

Zhou Jianpin1, Zhang Shuliu2   

  1. 1. School of Navigation, Jimei University, Xiamen 361021, China;
    2. Jilin Power Supply Company of State Grid, Jilin 132000, China
  • Received:2018-12-10 Revised:2019-02-15 Online:2019-10-10 Published:2019-12-12

摘要: 针对具有周期性波动需求的动态随机库存路径问题,提出了基于深度强化学习进行仿真优化并实现周期平稳策略的新方法。所研究问题构建动态组合优化模型,通过深度强化学习和设置启发规则来综合决定每个时期的补货节点集合和补货批量分配权重。仿真实验结果表明,与现有文献中的两种方法相比,所提出的方法在较低波动需求情况下可分别提高一个周期的平均利润约2.7%和3.9%,在较高波动需求情况下提高约8.2%和7.1%,而周期服务水平在不同需求波动环境下都可以平稳地保持在一个较小的波动范围内。

关键词: 库存路径问题, 启发规则, 深度Q-学习, 动态, 周期平稳策略

Abstract: Aiming at the dynamic stochastic inventory routing problem with periodic fluctuation of demand, a novel simulation optimization approach based on deep reinforcement learning is proposed to achieving periodic steady strategy. Firstly a dynamic combinatorial optimization model is constructed. Then, by deep reinforcement learning and setting heuristic rules, the replenishment nodes set selection and the replenishment batch allocation weights in each period are determined. The simulation experimental results show that the proposed method can improve the average profit of a cycle by about 2.7% and 3.9% in low fluctuating demand case and by about 8.2% and 7.1% in high fluctuating demand case compared with the two solution methods in the existing literature, and the cycle service level can be stabilized within a small fluctuation range under different demand fluctuation environments.

Key words: inventory routing problem, heuristic rules, deep Q-learning, dynamics, periodic steady strategy

中图分类号: