2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00509
|View full text |Cite
|
Sign up to set email alerts
|

Knockoff Nets: Stealing Functionality of Black-Box Models

Abstract: Machine Learning (ML) models are increasingly deployed in the wild to perform a wide range of tasks. In this work, we ask to what extent can an adversary steal functionality of such "victim" models based solely on blackbox interactions: image in, predictions out. In contrast to prior work, we present an adversary lacking knowledge of train/test data used by the model, its internals, and semantics over model outputs. We formulate model functionality stealing as a two-step approach: (i) querying a set of input i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
385
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 393 publications
(387 citation statements)
references
References 40 publications
2
385
0
Order By: Relevance
“…For example, imagine a model that some company has developed through many years of research in a specific field. The knowledge synthesized in the model built might be considered to be confidential, and it may be compromised even by providing only input and output access [356]. The latter shows that, under minimal assumptions, data model functionality stealing is possible.…”
Section: Explanations For Ai Security: Xai and Adversarial Machine Lementioning
confidence: 99%
See 2 more Smart Citations
“…For example, imagine a model that some company has developed through many years of research in a specific field. The knowledge synthesized in the model built might be considered to be confidential, and it may be compromised even by providing only input and output access [356]. The latter shows that, under minimal assumptions, data model functionality stealing is possible.…”
Section: Explanations For Ai Security: Xai and Adversarial Machine Lementioning
confidence: 99%
“…Notwithstanding this explicit concern from regulatory bodies, loss of privacy has been compromised by DL methods in scenarios where no data fusion is performed. For instance, a few images are enough to threaten users' privacy even in the presence of image obfuscation [420], and the model parameters of a DNN can be exposed by simply performing input queries on the model [356,357]. An approach to explain loss of privacy is by using privacy loss and intent loss subjective scores.…”
Section: Opportunities and Challenges In Privacy And Data Fusion Undementioning
confidence: 99%
See 1 more Smart Citation
“…In these adversary models, adversary is not assumed to have access to pre-trained models: both the target and the substitute model DNNs are trained from scratch. Since the time of this writing, stealing DNNs for more complicated datasets like CIFAR-10 6 , have been done by assuming both the target model and attacker models are finetuned from pre-trained ImageNet classifiers [37], [40]. These attacks benefit from correlations between different [40] or same [37] pre-trained models.…”
Section: G Takeawaysmentioning
confidence: 99%
“…A similar work [32] shows the simplicity of reverse-engineering black-box neural network weights, architecture, optimization method and the training/data split. In [33], authors reframe the goal from model theft, to arriving at a 'knockoff' model exhibiting the same functionality. In [34], authors ignore model parameters and instead attempt to steal the hyperparameters of a network.…”
Section: Attacks On Deployed Neural Networkmentioning
confidence: 99%