2018 International Symposium on Electronics and Telecommunications (ISETC) 2018
DOI: 10.1109/isetc.2018.8583854
|View full text |Cite
|
Sign up to set email alerts
|

Multiobjective Reinforcement Learning for Reconfigurable Adaptive Optimal Control of Manufacturing Processes

Abstract: In industrial applications of adaptive optimal control often multiple contrary objectives have to be considered. The relative importance (weights) of the objectives are often not known during the design of the control and can change with changing production conditions and requirements. In this work a novel model-free multiobjective reinforcement learning approach for adaptive optimal control of manufacturing processes is proposed. The approach enables sample-efficient learning in sequences of control configura… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

2
5

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…In addition to the motivating examples discussed above, recent years have seen multiobjective learning and planning methods applied across a wide range of problem domains including: distributed computing [27,124], drug and molecule design [62,214], cybersecurity [162], simulation [132], job shop scheduling [98], cognitive radio networks [100,129], satellite communications [45,63], recommender systems [78], power systems [34,35,97,193], building management [213], traffic management [70], manufacturing [36,54,80], bidding and pricing [76,207], education [151], and robotics [64]. The scope and variety of these applications supports our assertion that many important problems involve multiple objectives, and are best addressed using explicitly multi-objective methods.…”
Section: Other Topicsmentioning
confidence: 99%
“…In addition to the motivating examples discussed above, recent years have seen multiobjective learning and planning methods applied across a wide range of problem domains including: distributed computing [27,124], drug and molecule design [62,214], cybersecurity [162], simulation [132], job shop scheduling [98], cognitive radio networks [100,129], satellite communications [45,63], recommender systems [78], power systems [34,35,97,193], building management [213], traffic management [70], manufacturing [36,54,80], bidding and pricing [76,207], education [151], and robotics [64]. The scope and variety of these applications supports our assertion that many important problems involve multiple objectives, and are best addressed using explicitly multi-objective methods.…”
Section: Other Topicsmentioning
confidence: 99%
“…The presented approach includes deep neural networks, which extract features and a trained reinforcement leaning algorithm that uses these features as an input to control the process in real-time. In the realm of metal forming Dornheim et al (Dornheim & Link, 2018;Dornheim et al, 2019) coupled a deep reinforcement learning algorithm with an FE simulation model to enable multi-objective optimization of a deep drawing process. The presented approach controls the blank holder force over the process in such a way that the product was manufactured as material-efficiently as possible and with as little residual stresses as possible while ensuring the desired product geometry.…”
Section: Machine Learning and Its Applications In Metal Formingmentioning
confidence: 99%
“…We state that the Thresholded Lexicographic Ordering action selection from Section II-C is equivalent to the policy 6 πTLO…”
Section: Appendix Equivalence Of Tlo Formulationsmentioning
confidence: 99%
“…This includes the dynamic preferences scenario [5], where the preferences are non-stationary during application. This scenario is given for example in domains in which the optimization criteria depend on dynamic prices of open markets [4], or on changing requirements for process results [6].…”
Section: Introductionmentioning
confidence: 99%