Machine Learning (ML) techniques and algorithms, which are emerging technologies in Industry 4.0, present new possibilities for complex scheduling methods. Since different rules can be applied to different circumstances, it can be difficult for the decision-maker to choose the right rule at any given time. The purpose of the paper is to build an “intelligent” tool that adapts its choices in response to changes in the state of the production line. A Deep Q-Network (DQN), a typical Deep Reinforcement Learning (DRL) method, is proposed for creating a self-optimizing scheduling policy. The system has a set of known dispatching rules for each machine’s queue, from which the best one is dynamically chosen, according to the system state. The novelty of the paper is how the reward function, state, and action space are modelled. A series of experiments were conducted to determine the best DQN network size and the most influential hyperparameters for training.

Dynamic Scheduling in a Flow Shop Using Deep Reinforcement Learning / Marchesano, M. G.; Guizzi, G.; Santillo, L. C.; Vespoli, S.. - 630 IFIP:(2021), pp. 152-160. [10.1007/978-3-030-85874-2_16]

Dynamic Scheduling in a Flow Shop Using Deep Reinforcement Learning

Marchesano M. G.;Guizzi G.;Santillo L. C.;Vespoli S.
2021

Abstract

Machine Learning (ML) techniques and algorithms, which are emerging technologies in Industry 4.0, present new possibilities for complex scheduling methods. Since different rules can be applied to different circumstances, it can be difficult for the decision-maker to choose the right rule at any given time. The purpose of the paper is to build an “intelligent” tool that adapts its choices in response to changes in the state of the production line. A Deep Q-Network (DQN), a typical Deep Reinforcement Learning (DRL) method, is proposed for creating a self-optimizing scheduling policy. The system has a set of known dispatching rules for each machine’s queue, from which the best one is dynamically chosen, according to the system state. The novelty of the paper is how the reward function, state, and action space are modelled. A series of experiments were conducted to determine the best DQN network size and the most influential hyperparameters for training.
2021
978-3-030-85873-5
978-3-030-85874-2
Dynamic Scheduling in a Flow Shop Using Deep Reinforcement Learning / Marchesano, M. G.; Guizzi, G.; Santillo, L. C.; Vespoli, S.. - 630 IFIP:(2021), pp. 152-160. [10.1007/978-3-030-85874-2_16]
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11588/895571
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 5
  • ???jsp.display-item.citation.isi??? 3
social impact