2012 American Control Conference (ACC) 2012
DOI: 10.1109/acc.2012.6314613
|View full text |Cite
|
Sign up to set email alerts
|

Payoff-based Inhomogeneous Partially Irrational Play for potential game theoretic cooperative control: Convergence analysis

Abstract: This paper handles a kind of strategic game called potential games and develops a novel learning algorithm Payoff-based Inhomogeneous Partially Irrational Play (PIPIP). The present algorithm is based on Distributed Inhomogeneous Synchronous Learning (DISL) presented in an existing work but, unlike DISL, PIPIP allows agents to make irrational decisions with a specified probability, i.e. agents can choose an action with a low utility from the past actions stored in the memory. Due to the irrational decisions, we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
23
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 20 publications
(23 citation statements)
references
References 24 publications
0
23
0
Order By: Relevance
“…Another one verified in simulation is the one proposed in [12] where pareto equilibria are reached by the guards. Moreover, the protocols prosed in [6] and in [7] have been verified to converge to Nash equilibria. All those protocols have been designed for static scenarios, where the intruder is not supposed to move, i.e., it is a fixed threat.…”
Section: Nostop Evaluationmentioning
confidence: 96%
“…Another one verified in simulation is the one proposed in [12] where pareto equilibria are reached by the guards. Moreover, the protocols prosed in [6] and in [7] have been verified to converge to Nash equilibria. All those protocols have been designed for static scenarios, where the intruder is not supposed to move, i.e., it is a fixed threat.…”
Section: Nostop Evaluationmentioning
confidence: 96%
“…We restrict the analysis to two distributed learning algorithms: the Distributed Inhomogeneous Synchronous Learning (DISL) algorithm, [5], and the Payoff-based Inhomogeneous Partially Irrational Play (PIPIP) algorithm, [6].…”
Section: Learning Algorithmsmentioning
confidence: 99%
“…In [5] it is shown that any team of robots playing a constrained potential game Γ satisfying Assumption 1, and following DISL rules converges to a CNE. In [6] is shown that any team of robots playing the same game Γ, and following PIPIP rules converges to a potential maximizer which is an efficient CNE, i.e., it is the global maximizer of the potential function.…”
Section: Learning Algorithmsmentioning
confidence: 99%
See 1 more Smart Citation
“…After that one needs to develop a learning procedure, which leads agents to one of these states. Among learning algorithms that are applicable to potential games and have been presented in the literature so far the following ones demonstrate an efficient performance by approaching the system to its optimum as time runs: the log-linear learning [21], [43], its payoff-based and synchronous versions [21], payoff-based inhomogeneous partially irrational play [12], [50], and adaptive Q-learning and ε-greedy decision rule [9]. All these algorithms can be applied only to discrete optimization problems and, thus, assume players to have discrete actions.…”
mentioning
confidence: 99%