2020
DOI: 10.1609/aaai.v34i05.6217
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Communicate Implicitly by Actions

Abstract: In situations where explicit communication is limited, human collaborators act by learning to: (i) infer meaning behind their partner's actions, and (ii) convey private information about the state to their partner implicitly through actions. The first component of this learning process has been well-studied in multi-agent systems, whereas the second — which is equally crucial for successful collaboration — has not. To mimic both components mentioned above, thereby completing the learning process, we introduce … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 30 publications
0
8
0
Order By: Relevance
“…We describe two classes of algorithms that we believe are most closely related to this GME. The first regards modeling the private information of other agents [39,1,31]. These approaches target a more general problem setting than the one we investigate here.…”
Section: Related Methodsmentioning
confidence: 99%
“…We describe two classes of algorithms that we believe are most closely related to this GME. The first regards modeling the private information of other agents [39,1,31]. These approaches target a more general problem setting than the one we investigate here.…”
Section: Related Methodsmentioning
confidence: 99%
“…to teach a non-standard bidding system, can be problematic. • Many learning algorithms in bridge bidding use (or attempt to use) partner's hand estimation [27,26,18,7,22]. BridgeHand2V ec provides better cost functions and methods for assessing estimation quality.…”
Section: Introductionmentioning
confidence: 99%
“…Currently, bridge programs use rule sets. However, experiments using reinforcement learning [25,18,11,7,22,21] indicate that it is possible to improve here by creating a custom system, but this poses problems with revealing arrangements.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…MARL often has centralized training and decentralized execution (Foerster et al, 2016;Kraemer and Banerjee, 2016). There are several protocols for MARL training, such as sharing parameters between agents and explicit (Foerster et al, 2016(Foerster et al, , 2018Sukhbaatar et al, 2016;Mordatch and Abbeel, 2018) or implicit (Tian et al, 2020) communication between agents by using an actor-critic policy gradient with a centralized critic for all agents (Foerster et al, 2018). The aim of these protocols is to correctly assign credits so that an agent can deduce its contribution to the team's success.…”
Section: Introductionmentioning
confidence: 99%