Proceedings 2018 Network and Distributed System Security Symposium 2018
DOI: 10.14722/ndss.2018.23291
|View full text |Cite
|
Sign up to set email alerts
|

Trojaning Attack on Neural Networks

Abstract: With the fast spread of machine learning techniques, sharing and adopting public machine learning models become very popular. This gives attackers many new opportunities. In this paper, we propose a trojaning attack on neural networks. As the models are not intuitive for human to understand, the attack features stealthiness. Deploying trojaned models can cause various severe consequences including endangering human lives (in applications like autonomous driving). We first inverse the neural network to generate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
950
1

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 917 publications
(1,003 citation statements)
references
References 32 publications
4
950
1
Order By: Relevance
“…TTE attacks, which require knowledge of the classifier [21] [19][4] [16], they are a serious practical threat to the integrity of deployed machine learning solutions. Like many existing works, we focus here on DNN image classifiers for convenience, although backdoor attacks are also studied in other domains such as speech recognition [13].…”
Section: Introductionmentioning
confidence: 99%
“…TTE attacks, which require knowledge of the classifier [21] [19][4] [16], they are a serious practical threat to the integrity of deployed machine learning solutions. Like many existing works, we focus here on DNN image classifiers for convenience, although backdoor attacks are also studied in other domains such as speech recognition [13].…”
Section: Introductionmentioning
confidence: 99%
“…where I is the identity matrix and (21) follows from (x ⊤ u y i )x u = (x u x ⊤ u )y i . Lastly, computing (20) and (21) for all i ∈ Γ u yields J w v (x u ) and J w v (y i ). Note that ∇ w vr ut can be computed in exactly the same procedure.…”
Section: Solving Rating Scores For a Fake Usermentioning
confidence: 99%
“…Trojan Attacks. Neural networks, such as for facial recognition systems, can be trained in a way that they output a specific value, when the input has a certain "trojan trigger" embedded in it [45,68]. The trojan trigger can be a fixed input pattern (e.g., a sub-image) or some transformation that can be stamped on to a benign image.…”
Section: Security Applicationsmentioning
confidence: 99%