基于深度强化学习的柔性作业车间调度方法

A flexible job shop scheduling method based on deep reinforcement learning

  • 摘要: 由于传统的调度方法在求解大规模柔性作业车间调度问题中存在短视性、计算时间过长和算法参数难以确定等问题,因此提出了一种有效求解大规模柔性作业车间调度问题的深度强化学习方法。首先,将柔性作业车间调度问题转化为一个多智能体马尔科夫决策过程。然后,构建一个用于求解柔性作业车间调度问题的演员评论家模型,演员网络根据状态输出调度规则,智能体根据调度规则选择合适的工序,评论家网络根据状态和奖励对演员网络的动作进行评估。最后,采用不同规模的柔性作业车间调度问题实例验证该方法的性能。实验结果表明,该方法的求解质量优于启发式调度规则,求解效率优于元启发式算法。

     

    Abstract: The traditional scheduling methods suffer from short-sightedness, long computational time, and difficulty in determining algorithm parameters in solving large-scale flexible job shop scheduling problems. Therefore, an effective deep reinforcement learning method for solving the large-scale flexible job shop scheduling problem is proposed. Firstly, the flexible job shop scheduling problem is transformed into a multi-agent Markov decision process. Secondly, an actor-critic model for solving the flexible job shop scheduling problem is constructed. The actor network outputs the scheduling rule according to the state, the agent selects the appropriate operation according to the scheduling rule, and the critic network evaluates the action of the actor network according to the state and the reward. Finally, the flexible job shop scheduling problem instances of different sizes are used to verify the performance of the method. The experimental results show that the solution quality of the method is better than the heuristic scheduling rules and the solution efficiency is better than the meta-heuristic algorithms.

     

/

返回文章
返回