2020
DOI: 10.48550/arxiv.2006.11078
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Differentiable Language Model Adversarial Attacks on Categorical Sequence Classifiers

I. Fursov,
A. Zaytsev,
N. Kluchnikov
et al.

Abstract: An adversarial attack paradigm explores various scenarios for the vulnerability of deep learning models: minor changes of the input can force a model failure. Most of the state of the art frameworks focus on adversarial attacks for images and other structured model inputs, but not for categorical sequences models. Successful attacks on classifiers of categorical sequences are challenging because the model input is tokens from finite sets, so a classifier score is non-differentiable with respect to inputs, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 23 publications
0
2
0
Order By: Relevance
“…Adversarial attacks and defences from them are of crucial importance in financial applications [28]. The literature of adversarial attacks on transaction records includes [8,9]. However, the authors do not consider the peculiarities of transaction records data and apply general approaches to adversarial attacks on discrete sequence data.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Adversarial attacks and defences from them are of crucial importance in financial applications [28]. The literature of adversarial attacks on transaction records includes [8,9]. However, the authors do not consider the peculiarities of transaction records data and apply general approaches to adversarial attacks on discrete sequence data.…”
Section: Related Workmentioning
confidence: 99%
“…However, FGSM-based attacks provide better performance scores than SamplingFool-based ones due to the random search for adversarial examples in the second case. It can be useful to unite these approaches to create a generative model that can generate sequences of transaction records that are both realistic and adversarial [8]. Also, more realistic concatenation of tokens to the end of a sequence results in lower performance scores.…”
Section: Overall Attack Qualitymentioning
confidence: 99%