2021
DOI: 10.48550/arxiv.2104.08763
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Making Attention Mechanisms More Robust and Interpretable with Virtual Adversarial Training for Semi-Supervised Text Classification

Abstract: We propose a new general training technique for attention mechanisms based on virtual adversarial training (VAT). VAT can compute adversarial perturbations from unlabeled data in a semi-supervised setting for the attention mechanisms that have been reported in previous studies to be vulnerable to perturbations. Empirical experiments reveal that our technique (1) provides significantly better prediction performance compared to not only conventional adversarial training-based techniques but also VAT-based techni… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 27 publications
0
2
0
Order By: Relevance
“…[28] perturbed word embeddings instead of original input text using FGSM for NLP task. Some of the recent works [20,21,44] added perturbations to the attention mechanism of transformer methods using FGSM. [26] used multi-step FGSM to generate adversarial examples that proved more effective at the cost of computational overhead.…”
Section: Adversarial Trainingmentioning
confidence: 99%
“…[28] perturbed word embeddings instead of original input text using FGSM for NLP task. Some of the recent works [20,21,44] added perturbations to the attention mechanism of transformer methods using FGSM. [26] used multi-step FGSM to generate adversarial examples that proved more effective at the cost of computational overhead.…”
Section: Adversarial Trainingmentioning
confidence: 99%
“…[43] applies AT to relation extraction. Recent works [21,22,47] propose to apply perturbations to the attention mechanism in Transformer-based encoders. Compared to single-step FGSM, [26] demonstrates the superior effectiveness of the multi-step approach to generate perturbed examples with projected gradient descent, which comes at a greater computational cost due to the inner loop that iteratively calculates the perturbations.…”
Section: Related Work 21 Adversarial Trainingmentioning
confidence: 99%