Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.85
|View full text |Cite
|
Sign up to set email alerts
|

Posterior Differential Regularization with f-divergence for Improving Model Robustness

Abstract: We address the problem of enhancing model robustness through regularization. Specifically, we focus on methods that regularize the model posterior difference between clean and noisy inputs. Theoretically, we provide a connection of two recent methods, Jacobian Regularization and Virtual Adversarial Training, under this framework. Additionally, we generalize the posterior differential regularization to the family of f -divergences and characterize the overall framework in terms of Jacobian matrix. Empirically, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
1

Relationship

2
4

Authors

Journals

citations
Cited by 17 publications
(10 citation statements)
references
References 24 publications
0
10
0
Order By: Relevance
“…In addition to better textual representations from ELECTRA (Clark et al, 2020) (Cheng et al, 2020b) to improve the robustness of the extractive reader. As an extension to recent methods for improving the model local smoothness (Miyato et al, 2018;Sokolić et al, 2017), PDR aims at regularizing the posterior difference between the clean and noisy inputs with regard to the family of f -divergences (Csiszár and Shields, 2004).…”
Section: Improvement Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…In addition to better textual representations from ELECTRA (Clark et al, 2020) (Cheng et al, 2020b) to improve the robustness of the extractive reader. As an extension to recent methods for improving the model local smoothness (Miyato et al, 2018;Sokolić et al, 2017), PDR aims at regularizing the posterior difference between the clean and noisy inputs with regard to the family of f -divergences (Csiszár and Shields, 2004).…”
Section: Improvement Methodsmentioning
confidence: 99%
“…As an extension to recent methods for improving the model local smoothness (Miyato et al, 2018;Sokolić et al, 2017), PDR aims at regularizing the posterior difference between the clean and noisy inputs with regard to the family of f -divergences (Csiszár and Shields, 2004). Different from Cheng et al (2020b) where only clean supervision setting is considered, in this work, we apply PDR to the weakly supervised open-domain QA scenario. Given it is computationally expensive to enumerate all possible spans, we apply two separate regularization terms for the begin and end position probabilities at the multi-passage level, respectively.…”
Section: Improvement Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Previous works show that virtual adversarial training (VAT, Miyato et al ( 2018)) can improve the performance for general NLU and question answering tasks (Jiang et al, 2020;Cheng et al, 2021). In the multiple-choice commonsense reasoning task, the goal is to minimize the cross-entropy loss:…”
Section: Virtual Adversarial Training (Vat)mentioning
confidence: 99%