2019
DOI: 10.48550/arxiv.1911.11972
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adversarial Deep Reinforcement Learning based Adaptive Moving Target Defense

Abstract: Moving target defense (MTD) is a proactive defense approach that aims to thwart attacks by continuously changing the attack surface of a system (e.g., changing host or network configurations), thereby increasing the adversary's uncertainty and attack cost. To maximize the impact of MTD, a defender must strategically choose when and what changes to make, taking into account both the characteristics of its system as well as the adversary's observed activities. Finding an optimal strategy for MTD presents a signi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 2 publications
0
4
0
Order By: Relevance
“…While game-theoretic formalism has been used to model various cyber-security scenarios [30,31,6], it is impractical to expect security experts to provide the parameters of the game upfront [13,15,16]. In the context of Moving Target Defense (MTD) in particular, determining the impact of various attacks, the asymmetric impacts of a particular defense on performance, and the switching cost of a system are better obtained via interaction with an environment.…”
Section: Strong Stackelberg Q-learning In Bsmgsmentioning
confidence: 99%
See 3 more Smart Citations
“…While game-theoretic formalism has been used to model various cyber-security scenarios [30,31,6], it is impractical to expect security experts to provide the parameters of the game upfront [13,15,16]. In the context of Moving Target Defense (MTD) in particular, determining the impact of various attacks, the asymmetric impacts of a particular defense on performance, and the switching cost of a system are better obtained via interaction with an environment.…”
Section: Strong Stackelberg Q-learning In Bsmgsmentioning
confidence: 99%
“…It should be no surprise that defense strategies learned via single-agent RL methods (eg. [13,15,16]) are prone to be exploitable against strategic opponents in cyber-security scenarios. The existence of such a cycle makes the learned policy exploitable, resulting in low rewards consistently.…”
Section: Mtd For Web-applicationsmentioning
confidence: 99%
See 2 more Smart Citations