2021
DOI: 10.51593/20190041
|View full text |Cite
|
Sign up to set email alerts
|

Key Concepts in AI Safety: Robustness and Adversarial Examples

Abstract: This paper is the second installment in a series on “AI safety,” an area of machine learning research that aims to identify causes of unintended behavior in machine learning systems and develop tools to ensure these systems work safely and reliably. The first paper in the series, “Key Concepts in AI Safety: An Overview,” described three categories of AI safety issues: problems of robustness, assurance, and specification. This paper introduces adversarial examples, a major challenge to robustness in modern mach… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…Theorem 3 (Main result) Consider the SLS Σ as given in (1), where the initial state is i.i.d with the uniform distribution over the unit sphere S n−1 and the switching signal is uniform and i.i.d. in M. Given N, T ∈ Z + , the sample set ω N as defined in (4), and k ≤ T − 1, let (γ * k (ω N ), P * k (ω N )) be the unique solution of the scenario program (9).…”
Section: Main Result: Probabilistic Stability Certificatesmentioning
confidence: 99%
See 3 more Smart Citations
“…Theorem 3 (Main result) Consider the SLS Σ as given in (1), where the initial state is i.i.d with the uniform distribution over the unit sphere S n−1 and the switching signal is uniform and i.i.d. in M. Given N, T ∈ Z + , the sample set ω N as defined in (4), and k ≤ T − 1, let (γ * k (ω N ), P * k (ω N )) be the unique solution of the scenario program (9).…”
Section: Main Result: Probabilistic Stability Certificatesmentioning
confidence: 99%
“…holds for any output sequence (y(0), y(1), • • • , y(T − 1)) and generated by (1). Hence, in order to find a bound of the JSR, we have to solve the optimization problem for some sufficiently large…”
Section: Estimating Jsr From Datamentioning
confidence: 99%
See 2 more Smart Citations
“…The process by which many ML systems reach decisions can often be poorly understood and highly sensitive to small changes that a human analyst would view as trivial, which often makes it possible for attackers to find "adversarial examples"-slightly altered inputs that dramatically change a model's response despite being undetectable to a human. 52 The use of ML models also opens up new avenues of attack: the model itself must be kept secure, but defenders must also make sure that their data is not poisoned and that the (typically open source) algorithms and statistical packages they use have not been tampered with. 53 In addition, while machine learning is sometimes presented as an objective process of "learning patterns from data," in reality the design of ML systems is often the result of many judgment calls.…”
Section: Detection Figure 4 Ai Applications For Detectionmentioning
confidence: 99%