2022
DOI: 10.48550/arxiv.2205.09362
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sparse Adversarial Attack in Multi-agent Reinforcement Learning

Abstract: Cooperative multi-agent reinforcement learning (cMARL) has many real applications, but the policy trained by existing cMARL algorithms is not robust enough when deployed. There exist also many methods about adversarial attacks on the RL system, which implies that the RL system can suffer from adversarial attacks, but most of them focused on single agent RL. In this paper, we propose a sparse adversarial attack on cMARL systems. We use (MA)RL with regularization to train the attack policy. Our experiments show … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…RADAR (Phan et al 2021) learns resilient MARL policy via adversarial value decomposition. Hu and Zhang (2022) further design an action regularizer to attack the CMARL system efficiently. Xue et al (2022c) recently consider the multi-agent adversarial communication, learning robust communication policy when some message senders are poisoned.…”
Section: Related Workmentioning
confidence: 99%
“…RADAR (Phan et al 2021) learns resilient MARL policy via adversarial value decomposition. Hu and Zhang (2022) further design an action regularizer to attack the CMARL system efficiently. Xue et al (2022c) recently consider the multi-agent adversarial communication, learning robust communication policy when some message senders are poisoned.…”
Section: Related Workmentioning
confidence: 99%
“…As for the uncertainty caused by the inaccurate knowledge of the MARL dynamic model, R-MADDPG [44] proposes the concept of robust Nash equilibrium, treats the uncertainty of environment as a natural agent, and exhibits superiority when encountering reward uncertainty. Consider the observation perturbation, [16] learns an adversarial observation policy to attack one participant in a cooperative MARL system, demonstrating the high vulnerability of cooperative MARL facing observation perturbation. For the action robustness in cooperative MARL, ARTS [45] and RADAR [46] learn resilient MARL policies via adversarial value decomposition.…”
Section: Related Workmentioning
confidence: 99%