2020
DOI: 10.1609/aaai.v34i02.5494
|View full text |Cite
|
Sign up to set email alerts
|

End-to-End Game-Focused Learning of Adversary Behavior in Security Games

Abstract: Stackelberg security games are a critical tool for maximizing the utility of limited defense resources to protect important targets from an intelligent adversary. Motivated by green security, where the defender may only observe an adversary's response to defense on a limited set of targets, we study the problem of learning a defense that generalizes well to a new set of targets with novel feature values and combinations. Traditionally, this problem has been addressed via a two-stage approach where an adversary… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
4
4
2

Relationship

3
7

Authors

Journals

citations
Cited by 19 publications
(13 citation statements)
references
References 14 publications
0
13
0
Order By: Relevance
“…This includes specific problems such as Markowitz portfolio optimization (Bengio 1997) or finding physically feasible state transitions (de Avila Belbute-Peres et al 2018), as well as larger classes which exhibit properties like convexity or submodularity. Problem classes used in end-to-end training include polynomial-time solvable frameworks like quadratic programs (Amos and Kolter 2017) and linear programs (Wilder, Dilkina, and Tambe 2019), as well as zero-sum games (Ling, Fang, and Kolter 2018;Perrault et al 2019). In addition, a solution is proposed for problems encoded as submodular optimization problems in (Wilder, Dilkina, and Tambe 2019).…”
Section: Related Workmentioning
confidence: 99%
“…This includes specific problems such as Markowitz portfolio optimization (Bengio 1997) or finding physically feasible state transitions (de Avila Belbute-Peres et al 2018), as well as larger classes which exhibit properties like convexity or submodularity. Problem classes used in end-to-end training include polynomial-time solvable frameworks like quadratic programs (Amos and Kolter 2017) and linear programs (Wilder, Dilkina, and Tambe 2019), as well as zero-sum games (Ling, Fang, and Kolter 2018;Perrault et al 2019). In addition, a solution is proposed for problems encoded as submodular optimization problems in (Wilder, Dilkina, and Tambe 2019).…”
Section: Related Workmentioning
confidence: 99%
“…Differentiable optimization Amos et al [2] propose using a quadratic program as a differentiable layer and embedding it into deep learning pipeline, and Agrawal et al [1] extend their work to convex programs. Decision-focused learning [6,34] focuses on the predict-then-optimize [4,8] framework by embedding an optimization layer into training pipeline, where the optimization layers can be convex [6], linear [21,34], and non-convex [25,32]. Unfortunately, these techniques are of limited utility for sequential decision problems because their formulations grow linearly in the number of states and actions and thus differentiating through them quickly becomes infeasible.…”
Section: Related Workmentioning
confidence: 99%
“…4.2.1 Task-Oriented Objective: Estimated Clustering Loss. Firstly, we borrow the idea of existing works, which mainly focus on using a surrogate loss function L s to guide the learning process, where practitioners can either choose standard machine learning loss functions or other differentiable task-specific surrogate loss functions [3,9,10,30,40]. For unsupervised clustering task, the training sample is not known previously, whereas the supervised classification task have ground truth as labels.…”
Section: "Warm-up" Based Task-oriented Estimatormentioning
confidence: 99%