系统仿真学报 ›› 2021, Vol. 33 ›› Issue (12): 2782-2791.doi: 10.16182/j.issn1004731x.joss.21-FZ0774

• 综述 • 上一篇    下一篇

基于强化学习的车间调度问题研究简述

王霄汉1,2, 张霖1,2, 任磊1,2, 谢堃钰1,2, 王昆玉1,2, 叶飞1,2, 陈真1,2   

  1. 1.北京航空航天大学,北京 100191;
    2.复杂产品先进制造系统教育部工程研究中心,北京 100191
  • 收稿日期:2021-05-10 修回日期:2021-07-29 出版日期:2021-12-18 发布日期:2022-01-13
  • 作者简介:王霄汉(1998-),男,博士生,研究方向为离散仿真、多智能体系统和强化学习。E-mail:by1903042@buaa.edu.cn
  • 基金资助:
    国家重点研发计划(2018YFB1701600); 国家自然科学基金(61873014)

Brief Review on Applying Reinforcement Learning to Job Shop Scheduling Problems

Wang Xiaohan1,2, Zhang Lin1,2, Ren Lei1,2, Xie Kunyu1,2, Wang Kunyu1,2, Ye Fei1,2, Chen Zhen1,2   

  1. 1. Beihang University, Beijing 100191, China;
    2. Engineering Research Center of Complex Product Advanced Manufacturing Systems, Ministry of Education, Beijing 100191, China
  • Received:2021-05-10 Revised:2021-07-29 Online:2021-12-18 Published:2022-01-13

摘要: 强化学习在车间调度上获得了较低的时间响应和较优的模型泛化性。为阐述基于强化学习的车间调度问题整体研究现状,总结当前基于强化学习的调度框架,同时为后续相关研究奠定基础,介绍了车间调度与强化学习的背景,分析了车间调度问题中常用的2种仿真技术,给出了强化学习解决车间调度问题的2种常用架构。此外,针对强化学习在车间调度问题上的应用,指出了现存的一些挑战,并对相关研究进展从直接调度、基于特征表示的调度、以及基于参数搜索的调度3个方面进行了介绍。

关键词: 强化学习应用, 车间调度, 图神经网络, 组合优化, 深度学习, 特征表示

Abstract: Reinforcement Learning (RL) achieves lower time response and better model generalization in Job Shop Scheduling Problem (JSSP). To explain the current overall research status of JSSP based on RL, summarize the current scheduling framework based on RL, and lay the foundation for follow-up research, the backgrounds of JSSP and RL are introduced. Two simulation techniques commonly used in JSSP are analyzed and two commonly used frameworks for RL to solve JSSP are given. In addition, some existing challenges are pointed out, and related research progress is introduced from three aspects: direct scheduling, feature representation-based scheduling, and parameter search-based scheduling.

Key words: reinforcement learning application, job shop scheduling problem, graph neural network, combinatorial optimization, deep learning, feature representation

中图分类号: