2007
DOI: 10.1007/978-3-540-74171-8_42
|View full text |Cite
|
Sign up to set email alerts
|

Cooperation Between Multiple Agents Based on Partially Sharing Policy

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2012
2012
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 4 publications
0
2
0
Order By: Relevance
“…For this reason, classical learning algorithms, such as Q-learning, have been modified to achieve better conversion. The Win of Learn Fast Policy Hill Climbing (WoLF-PHC) algorithm is proposed as an extension of the Q-learning algorithm for more efficient learning of the dynamic target [19]. Its characteristic is the use of different learning rates depending on the game outcome, which increases the convergence in multi-agent non-stationary environment.…”
Section: Multi-agent Reinforcement Learning Algorithmmentioning
confidence: 99%
See 1 more Smart Citation
“…For this reason, classical learning algorithms, such as Q-learning, have been modified to achieve better conversion. The Win of Learn Fast Policy Hill Climbing (WoLF-PHC) algorithm is proposed as an extension of the Q-learning algorithm for more efficient learning of the dynamic target [19]. Its characteristic is the use of different learning rates depending on the game outcome, which increases the convergence in multi-agent non-stationary environment.…”
Section: Multi-agent Reinforcement Learning Algorithmmentioning
confidence: 99%
“…In the course of the learning process, the chance of the fog node selecting a service demand is progressively increased, which can elevate the expected reward, followed by the reduction of the other actions [19]. Hence, the update of the service demand policy of the fog node can be presented as follows,…”
Section: Multi-agent Reinforcement Learning Algorithmmentioning
confidence: 99%