2021
DOI: 10.48550/arxiv.2111.04727
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Efficiently Learning Any One Hidden Layer ReLU Network From Queries

Abstract: Model extraction attacks have renewed interest in the classic problem of learning neural networks from queries. This work gives the first polynomial-time algorithm for learning one hidden layer neural networks provided black-box access to the network. Formally, we show that if F is an arbitrary one hidden layer neural network with ReLU activations, there is an algorithm with query complexity and running time that is polynomial in all parameters that outputs a network F ′ achieving low square loss relative to F… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 5 publications
0
1
0
Order By: Relevance
“…Since in a one-hidden-layer ReLU network (without bias term), F (0) = 0, and furthermore F is a 2kR-Lipschitz continuous function, we conclude that |F (x)| 2kR • x 2 for ∀x ∈ R d . Next, we can apply the proof of Lemma A.1 of [CKM21] to show that G(x) := F (x) 2 − µ is a zero-centered, sub-exponential random variable with sub-exponential norm G Ψ 1 = O(µ + 4R 2 k 3 ). Finally, by using the concentration property of sub-exponential random variables, we conclude that:…”
Section: A2 Measuring Distance Between Networkmentioning
confidence: 99%
“…Since in a one-hidden-layer ReLU network (without bias term), F (0) = 0, and furthermore F is a 2kR-Lipschitz continuous function, we conclude that |F (x)| 2kR • x 2 for ∀x ∈ R d . Next, we can apply the proof of Lemma A.1 of [CKM21] to show that G(x) := F (x) 2 − µ is a zero-centered, sub-exponential random variable with sub-exponential norm G Ψ 1 = O(µ + 4R 2 k 3 ). Finally, by using the concentration property of sub-exponential random variables, we conclude that:…”
Section: A2 Measuring Distance Between Networkmentioning
confidence: 99%