1998
DOI: 10.1023/a:1007518724497
|View full text |Cite
|
Sign up to set email alerts
|

Untitled

Abstract: Abstract. Recent algorithmic and theoretical advances in reinforcement learning (RL) have attracted widespread interest. RL algorithms have appeared that approximate dynamic programming on an incremental basis. They can be trained on the basis of real or simulated experiences, focusing their computation on areas of state space that are actually visited during control, making them computationally tractable on very large problems. If each member of a team of agents employs one of these algorithms, a new collecti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2004
2004
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 193 publications
(7 citation statements)
references
References 29 publications
0
7
0
Order By: Relevance
“…Bottlenecking inter-agent communication in multiagent RL. A wide range of multi-agent applications have benefitted from inter-agent message passing including distributed smart grid control (Pipattanasomporn et al, 2009), consensus in networks (You & Xie, 2011), multi-robot control (Ren & Sorensen, 2008), autonomous vehicle driving (Petrillo et al, 2018), elevators control (Crites & Barto, 1998) and for language learning in two-agent systems (Lazaridou et al, 2017). An important challenge in MARL is how to facilitate communication among interacting agents, especially in tasks requiring synchronization (Scardovi & Sepulchre, 2009;Wen et al, 2012).…”
Section: Related Workmentioning
confidence: 99%
“…Bottlenecking inter-agent communication in multiagent RL. A wide range of multi-agent applications have benefitted from inter-agent message passing including distributed smart grid control (Pipattanasomporn et al, 2009), consensus in networks (You & Xie, 2011), multi-robot control (Ren & Sorensen, 2008), autonomous vehicle driving (Petrillo et al, 2018), elevators control (Crites & Barto, 1998) and for language learning in two-agent systems (Lazaridou et al, 2017). An important challenge in MARL is how to facilitate communication among interacting agents, especially in tasks requiring synchronization (Scardovi & Sepulchre, 2009;Wen et al, 2012).…”
Section: Related Workmentioning
confidence: 99%
“…For example, [6] presents a methodology to improve elevator performance by applying machine learning. Crites et al [7] present a similar approach, focusing on safety and control issues. Similarly, Ho and Fu [8] uses a learning algorithm to optimize elevator performance.…”
Section: Literature Reviewmentioning
confidence: 99%
“…The real-world scenario introduces uncertainties that necessitate adaptive decision-making without complete knowledge of future events. Heuristic algorithms have traditionally tackled this issue [3], with recent exploration of solutions based on artificial intelligence algorithms [4][5][6][7].…”
Section: Introduction: Contribution Of the Papermentioning
confidence: 99%