作业车间调度
计算机科学
工作车间
动态优先级调度
调度(生产过程)
马尔可夫决策过程
数学优化
概括性
强化学习
运筹学
流水车间调度
分布式计算
人工智能
工业工程
地铁列车时刻表
马尔可夫过程
工程类
心理治疗师
操作系统
统计
数学
心理学
作者
Yuxin Li,Wenjin Gu,Minghai Yuan,Tang Ya-ming
标识
DOI:10.1016/j.rcim.2021.102283
摘要
With the extensive application of automated guided vehicles in manufacturing system, production scheduling considering limited transportation resources becomes a difficult problem. At the same time, the real manufacturing system is prone to various disturbance events, which increase the complexity and uncertainty of shop floor. To this end, this paper addresses the dynamic flexible job shop scheduling problem with insufficient transportation resources (DFJSP-ITR) to minimize the makespan and total energy consumption. As a sequential decision-making problem, DFJSP-ITR can be modeled as a Markov decision process where the agent should determine the scheduling object and allocation of resources at each decision point. So this paper adopts deep reinforcement learning to solve DFJSP-ITR. In this paper, the multiobjective optimization model of DFJSP-ITR is established. Then, in order to make agent learn to choose the appropriate rule based on the production state at each decision point, a hybrid deep Q network (HDQN) is developed for this problem, which combines deep Q network with three extensions. Moreover, the shop floor state model is established at first, and then the decision point, generic state features, genetic-programming-based action space and reward function are designed. Based on these contents, the training method using HDQN and the strategy for facing new job insertions and machine breakdowns are proposed. Finally, comprehensive experiments are conducted, and the results show that HDQN has superiority and generality compared with current optimization-based approaches, and can effectively deal with disturbance events and unseen situations through learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI