2018
DOI: 10.1016/j.cirp.2018.04.041
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement learning for adaptive order dispatching in the semiconductor industry

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
33
0
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
6
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 106 publications
(34 citation statements)
references
References 10 publications
0
33
0
1
Order By: Relevance
“…However, ML algorithms are not favorable for all industrial applications. The following properties are advantageous: (i) applications with a limited scope in terms of the number of states and actions (the learning period is dependent on these dimensions), (ii) responsive real-time decision systems (computing the output of a ML algorithm requires just linear operations), (iii) "cheap" training data (the trial-and-error approach is intensively data-driven) and (iv) complex environments that can hardly be described in detail (ability to generalize) [15]. This work brings the application of ML algorithms and the transition towards autonomous production systems one step closer to reality.…”
Section: Conclusion Discussion and Outlookmentioning
confidence: 99%
See 1 more Smart Citation
“…However, ML algorithms are not favorable for all industrial applications. The following properties are advantageous: (i) applications with a limited scope in terms of the number of states and actions (the learning period is dependent on these dimensions), (ii) responsive real-time decision systems (computing the output of a ML algorithm requires just linear operations), (iii) "cheap" training data (the trial-and-error approach is intensively data-driven) and (iv) complex environments that can hardly be described in detail (ability to generalize) [15]. This work brings the application of ML algorithms and the transition towards autonomous production systems one step closer to reality.…”
Section: Conclusion Discussion and Outlookmentioning
confidence: 99%
“…Moreover, the heuristic results show an almost stable performance that is not able to adapt to changing conditions. [15]…”
Section: Intelligent Order Dispatchingmentioning
confidence: 99%
“…For a detailed overview of the latest RL research in the domain of production planning and control, we refer to the work of [9,10,20,22].…”
Section: Basics Of Rlmentioning
confidence: 99%
“…• Current machine where the action is asked for (10 entries, binary, one-hot coding for 10 work groups) • Loading status of all machines (10 entries, binary, idle or busy) • Product setup per machine (50 entries, binary, one-hot coding for 5 product variants and ten machines) • Product variant in the current machine's buffer slots (100 entries, binary, one-hot coding for up to twenty buffers and 5 product variants) • Order status per buffer slot (20 entries, real-value, 20 buffer slots) • Full or empty status per buffer slot (20 entries, realvalue, 20 buffer slots)…”
Section: State Space Modellingmentioning
confidence: 99%
“…Nonetheless, both papers do not consider the application of RL for the direct generation of job schedules. The first paper we found that describes RL agents being able to directly allocate and sequence jobs is from Stricker et al (2018). The different agent types are responsible for different production control decisions, such as selecting the next job to be processed or assigning the selected job to a machine [18].…”
Section: Related Workmentioning
confidence: 99%