2022
DOI: 10.48550/arxiv.2205.10802
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Inverse-Inverse Reinforcement Learning. How to Hide Strategy from an Adversarial Inverse Reinforcement Learner

Abstract: Inverse reinforcement learning (IRL) deals with estimating an agent's utility function from its actions. In this paper, we consider how an agent can hide its strategy and mitigate an adversarial IRL attack; we call this inverse IRL (I-IRL). How should the decision maker choose its response to ensure a poor reconstruction of its strategy by an adversary performing IRL to estimate the agent's strategy? This paper comprises four results: First, we present an adversarial IRL algorithm that estimates the agent's st… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 22 publications
0
3
0
Order By: Relevance
“…This paper builds significantly on our previous work [4] on ECM for identifying cognitive radars, and [53], [54], [55] on ECCM for masking radar cognition. Theorem 6 extends IRL for cognitive radars [4] when the radar faces multiple resource constraints.…”
Section: Conclusion and Extensionsmentioning
confidence: 87%
See 1 more Smart Citation
“…This paper builds significantly on our previous work [4] on ECM for identifying cognitive radars, and [53], [54], [55] on ECCM for masking radar cognition. Theorem 6 extends IRL for cognitive radars [4] when the radar faces multiple resource constraints.…”
Section: Conclusion and Extensionsmentioning
confidence: 87%
“…Theorem 7 generalizes the cognition masking result of [53] to multiple constraints. Our previous works [53], [54], [55] assume optimal adversarial IRL via Afriat's theorem. This paper generalizes cognition masking to sub-optimal adversarial IRL algorithms.…”
Section: Conclusion and Extensionsmentioning
confidence: 99%
“…Similarly, in the reinforcement learning community, "inverse reinforcement learning" methods (Arora & Doshi, 2021;Ng, Russell, et al, 2000;Ramachandran & Amir, 2007) can infer reward functions, which can then be applied to influence observers, e.g. to make a robot's motion "legible" to humans (Dragan, 2015;Dragan, Lee, & Srinivasa, 2013;Hadfield-Menell, Russell, Abbeel, & Dragan, 2016) or oppositely to strategically fool adversarial viewers about its true intentions (Pattanayak, Krishnamurthy, & Berry, 2022).…”
Section: Background: Bayesian Inverse Planningmentioning
confidence: 99%
“…The Rational Communicative Social Actions (RCSA) framework integrates inverse planning and RSA to model communicative aspects of intimacy (Hung, Thomas, Radkani, Tenenbaum, & Saxe, 2022) and punishment . Similarly, in the reinforcement learning community, "inverse reinforcement learning" methods (Arora & Doshi, 2021;Ng, Russell, & others, 2000;Ramachandran & Amir, 2007) can infer reward functions, which can then be applied to influence observers, for example, to make a robot's motion "legible" to humans (Dragan, 2015;Dragan, Lee, & Srinivasa, 2013;Hadfield-Menell, Russell, Abbeel, & Dragan, 2016) or oppositely to strategically fool adversarial viewers about its true intentions (Pattanayak, Krishnamurthy, & Berry, 2022).…”
Section: Background: Bayesian Inverse Planningmentioning
confidence: 99%