Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence 2017
DOI: 10.24963/ijcai.2017/608
|View full text |Cite
|
Sign up to set email alerts
|

Equi-Reward Utility Maximizing Design in Stochastic Environments

Abstract: We present the Equi-Reward Utility Maximizing Design (ER-UMD) problem for redesigning stochastic environments to maximize agent performance. ER-UMD fits well contemporary applications that require offline design of environments where robots and humans act and cooperate. To find an optimal modification sequence we present two novel solution techniques: a compilation that embeds design into a planning problem, allowing use of off-the-shelf solvers to find a solution, and a heuristic search in the modifications s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 17 publications
(30 citation statements)
references
References 7 publications
0
30
0
Order By: Relevance
“…We present an actor-designer framework to mitigate the impacts of NSE by shaping the environment. The general idea of environment modification to influence the behaviors of the acting agents has been previously explored in other contexts (Zhang, Chen, and Parkes 2009), such as to accelerate agent learning (Randløv 2000), to quickly infer the goals of the actor (Keren, Gal, and Karpas 2014), and to maximize the agent's reward (Keren et al 2017). We study how environment shaping can mitigate NSE.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…We present an actor-designer framework to mitigate the impacts of NSE by shaping the environment. The general idea of environment modification to influence the behaviors of the acting agents has been previously explored in other contexts (Zhang, Chen, and Parkes 2009), such as to accelerate agent learning (Randløv 2000), to quickly infer the goals of the actor (Keren, Gal, and Karpas 2014), and to maximize the agent's reward (Keren et al 2017). We study how environment shaping can mitigate NSE.…”
Section: Discussion and Future Workmentioning
confidence: 99%
“…Instead, the possibility to strategically act on the environmental dynamics is studied in a limited number of works only. Some approaches belonging to the planning area [12,38], some are constrained to specific forms of environment configurability [8,9,34], and others based on the curriculum learning framework [4,7]. The goal of the dissertation [18] is to provide a uniform treatment of environment configurability in its diverse aspects.…”
Section: Configurable Environmentsmentioning
confidence: 99%
“…SWOPP relates to the environment design problem in AI in which an interested party, the system designer, can make changes to the environment of an agent in order to encourage desired behavior [ 14 ]. Keren et al [ 15 ] model environment design settings in which the system designer and the worker optimize the same utility function. The designer attempts to introduce changes to the environment that would reduce the cost of achieving the joint goals.…”
Section: Related Workmentioning
confidence: 99%