2021
DOI: 10.48550/arxiv.2106.15860
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Understanding Adversarial Attacks on Observations in Deep Reinforcement Learning

Abstract: Recent works demonstrate that deep reinforcement learning (DRL) models are vulnerable to adversarial attacks which can decrease the victim's total reward by manipulating the observations. Compared with adversarial attacks in supervised learning, it is much more challenging to deceive a DRL model since the adversary has to infer the environmental dynamics. To address this issue, we reformulate the problem of adversarial attacks in function space and separate the previous gradient based attacks into several subs… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
3

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(13 citation statements)
references
References 24 publications
0
13
0
Order By: Relevance
“…Extensive experimental results illustrate that our Consistent Attack can significantly reduce the performance of existing Embodied Vision Navigation policies on Habitat [Savva et al, 2019], which shows the effectiveness of our method and the vulnerability of existing Embodied Vision Navigation policies. In summary, our contributions are…”
Section: Introductionmentioning
confidence: 86%
See 3 more Smart Citations
“…Extensive experimental results illustrate that our Consistent Attack can significantly reduce the performance of existing Embodied Vision Navigation policies on Habitat [Savva et al, 2019], which shows the effectiveness of our method and the vulnerability of existing Embodied Vision Navigation policies. In summary, our contributions are…”
Section: Introductionmentioning
confidence: 86%
“…Since the agent in Embodied Vision Navigation needs to interact with the environment [Beattie et al, 2016;Kolve et al, 2017;Savva et al, 2019] actively, its security are even more important compared with the traditional AI task, e.g., image classification. For example, in the PointGoal navigation task [Anderson et al, 2018] in Habitat, i.e., the agent tries to find a path to reach the destination, the agent might be in danger when having collisions in this scenario [Savva et al, 2019]. In this paper, we mainly focus on the security of the agent [Savva et al, 2019] in the PointGoal task in Habitat, which is a common and popular baseline in Embodied AI.…”
Section: Embodied Vision Navigationmentioning
confidence: 99%
See 2 more Smart Citations
“…formations (Engstrom et al, 2019;Hendrycks & Dietterich, 2019), which can limit their applications in various securitysensitive tasks. For example, a small adversarial patch on the road markings can mislead the autonomous driving system (Jing et al, 2021;Qiaoben et al, 2021), which raises severe safety concerns. Compared with the p -norm bounded adversarial examples, semantic transformations can occur more naturally in real-world scenarios and are often unrestricted, including image rotation, translation, blur, weather, etc., most of which are common corruptions (Hendrycks & Dietterich, 2019).…”
Section: Introductionmentioning
confidence: 99%