2021 IEEE International Conference on Development and Learning (ICDL) 2021
DOI: 10.1109/icdl49984.2021.9515661
|View full text |Cite
|
Sign up to set email alerts
|

Priority-Objective Reinforcement Learning

Abstract: Intelligent agents often have to cope with situations in which their various needs must be prioritised. Efforts have been made, in the fields of cognitive robotics and machine learning, to model need prioritization. Examples of existing frameworks include normative decision theory, the subsumption architecture and reinforcement learning. Reinforcement learning algorithms oriented towards active goal prioritization include the options framework from hierarchical reinforcement learning and the ranking approach a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 14 publications
0
4
0
Order By: Relevance
“…the total value for a given state is determined on the basis of a weighted sum of calculated values V k for each objective where the weighting Fig. 1: Utility concepts for the multi-objective problem: (i) standard linear scalarization, (ii) the non-linear MORE scalarization that corresponds to a softmin function [28], (iii) C-MORE, a MORE scalarization that is shifted by c 0 = 2 on the first objective, giving it higher priority [29], but also continuously weighting in the second objective when the first one becomes satisfied.…”
Section: A Standard Linear Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…the total value for a given state is determined on the basis of a weighted sum of calculated values V k for each objective where the weighting Fig. 1: Utility concepts for the multi-objective problem: (i) standard linear scalarization, (ii) the non-linear MORE scalarization that corresponds to a softmin function [28], (iii) C-MORE, a MORE scalarization that is shifted by c 0 = 2 on the first objective, giving it higher priority [29], but also continuously weighting in the second objective when the first one becomes satisfied.…”
Section: A Standard Linear Methodsmentioning
confidence: 99%
“…This paper investigates how intrinsic motivation can be integrated into an agent with the non-linear multi-objective reinforcement learning framework C-MORE [29]. We capitalize on C-MORE's ability to dynamically balance the achievement of different needs, while allowing for a gradual prioritisation at the same time.…”
Section: A Contribution and Outlinementioning
confidence: 99%
See 2 more Smart Citations