2021
DOI: 10.48550/arxiv.2112.05495
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

How Private Is Your RL Policy? An Inverse RL Based Analysis Framework

Abstract: Reinforcement Learning (RL) enables agents to learn how to perform various tasks from scratch. In domains like autonomous driving, recommendation systems and more, optimal RL policies learned could cause a privacy breach if the policies memorize any part of the private reward. We study the set of existing differentially-private RL policies derived from various RL algorithms such as Value Iteration, Deep Q Networks, and Vanilla Proximal Policy Optimization. We propose a new Privacy-Aware Inverse RL (PRIL) analy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 17 publications
0
3
0
Order By: Relevance
“…However, the concept of "inferring a reward function" naturally prompts thoughts of privacy leakage and unreliable models. Consequently, it might be possible to consider the security and privacy problems of inverse reinforcement learning, like the work in [76]. Besides the security and problems in reinforcement learning itself, there are several works about applying reinforcement learning to security and privacy problems.…”
Section: Summary and Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…However, the concept of "inferring a reward function" naturally prompts thoughts of privacy leakage and unreliable models. Consequently, it might be possible to consider the security and privacy problems of inverse reinforcement learning, like the work in [76]. Besides the security and problems in reinforcement learning itself, there are several works about applying reinforcement learning to security and privacy problems.…”
Section: Summary and Discussionmentioning
confidence: 99%
“…As differential privacy is always adopted to establish a mathematical way of guaranteeing data privacy in reinforcement learning, and considering that inverse reinforcement learning is applied to inferring the reward function from demonstrations and providing rewards to the learning system, Prakash et al [76] investigated the existing set of privacy techniques for reinforcement learning and proposed a new Privacy-Aware Inverse reinforcement Learning (PRIL) analysis framework, which is a new form of privacy attack that targets the private reward function. This reward attack attempts to reconstruct the original reward from a privacy-preserving policy (such as differential privacy) using an inverse reinforcement algorithm.…”
Section: Privacy Of Reward Function In Mdpmentioning
confidence: 99%
See 1 more Smart Citation