2021
DOI: 10.1109/tcds.2020.2974509
|View full text |Cite
|
Sign up to set email alerts
|

Minimalistic Attacks: How Little It Takes to Fool Deep Reinforcement Learning Policies

Abstract: Recent studies have revealed that neural network based policies can be easily fooled by adversarial examples. However, while most prior works analyze the effects of perturbing every pixel of every frame assuming white-box policy access, in this paper we take a more restrictive view towards adversary generation -with the goal of unveiling the limits of a model's vulnerability. In particular, we explore minimalistic attacks by defining three key settings: (1) black-box policy access: where the attacker only has … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 31 publications
(17 citation statements)
references
References 39 publications
0
17
0
Order By: Relevance
“…Shannon entropy is another metric that can describe the action quality in the aspect of confidence in the decisionmaking process. Thus, we try to adopt the normalized Shannon entropy in [28] to the S 2 ES framework, named as S 2 ES-H. The normalized Shannon entropy can be written as…”
Section: When To Transfermentioning
confidence: 99%
“…Shannon entropy is another metric that can describe the action quality in the aspect of confidence in the decisionmaking process. Thus, we try to adopt the normalized Shannon entropy in [28] to the S 2 ES framework, named as S 2 ES-H. The normalized Shannon entropy can be written as…”
Section: When To Transfermentioning
confidence: 99%
“…does not comprehensively account for the scope of exploiting frame correlations. For this reason, even though [70] marks an advance over prior forms of adversarial attacks in RL, its adversary generation process may still be insufficient for unveiling the realizability of real-time attacks across many time-sensitive RL applications.…”
Section: Adversarial Attacks Transfermentioning
confidence: 99%
“…In Chapter 3, a minimalistic attack with the afore-stated strong assumptions being relaxed [70] has been investigated. In particular, a genetic algorithm (GA) [37] is used for the generation of selective perturbations (on selective pixels and frames), treating the policy as a black-box.…”
Section: Frame-correlation Transfers Trigger Economical Attacks On De...mentioning
confidence: 99%
See 2 more Smart Citations