2021
DOI: 10.1109/access.2021.3093456
|View full text |Cite
|
Sign up to set email alerts
|

Attention Meets Perturbations: Robust and Interpretable Attention With Adversarial Training

Abstract: Although attention mechanisms have been applied to a variety of deep learning models and have been shown to improve the prediction performance, it has been reported to be vulnerable to perturbations to the mechanism. To overcome the vulnerability to perturbations in the mechanism, we are inspired by adversarial training (AT), which is a powerful regularization technique for enhancing the robustness of the models. In this paper, we propose a general training technique for natural language processing tasks, incl… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…And we select the Binary Cross Entropy loss as D 1 and D 2 in (6). We compare our method with Vanilla attention (Wiegreffe and Pinter 2019), Word AT (Miyato, Dai, and Goodfellow 2016), Word iAT (Sato et al 2018), Attention RP (attention weight is trained with random perturbation), Attention AT and Attention iAT (Kitada and Iyatomi 2021).…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…And we select the Binary Cross Entropy loss as D 1 and D 2 in (6). We compare our method with Vanilla attention (Wiegreffe and Pinter 2019), Word AT (Miyato, Dai, and Goodfellow 2016), Word iAT (Sato et al 2018), Attention RP (attention weight is trained with random perturbation), Attention AT and Attention iAT (Kitada and Iyatomi 2021).…”
Section: Methodsmentioning
confidence: 99%
“…There exists some work studying or improving either the stability or the robustness of attention from the explanation perspective. Recently, Kitada and Iyatomi (2021) propose a method to improve the robustness to perturbation of embedding vector for attention. Specifically, they adopt adversarial training during the training process.…”
Section: Related Workmentioning
confidence: 99%
“…Therefore, Miyato et al [14] applied FGSM to NLP tasks by perturbing word embeddings rather than real text input, which is applicable in both supervised and semi-supervised scenarios, as it uses Virtual Adversarial Training (VAT) [25] in the latter. [26]- [28] proposed different works to add the perturbation to the attention mechanism of transformer-based methods instead of the word embeddings. To generate adversarial examples, Madry et al [24] adopted the multi-step approach in contrast to the single-step FGSM.…”
Section: A Adversarial Trainingmentioning
confidence: 99%
“…[28] perturbed word embeddings instead of original input text using FGSM for NLP task. Some of the recent works [20,21,44] added perturbations to the attention mechanism of transformer methods using FGSM. [26] used multi-step FGSM to generate adversarial examples that proved more effective at the cost of computational overhead.…”
Section: Adversarial Trainingmentioning
confidence: 99%