2020
DOI: 10.48550/arxiv.2004.02709
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Evaluating Models' Local Decision Boundaries via Contrast Sets

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 37 publications
(19 citation statements)
references
References 55 publications
0
19
0
Order By: Relevance
“…As a metric, the attack success rate has been utilized in numerous research studies to determine the effectiveness of adversarial attacks on NLP models [12], [15], [39], [115], [116]. For instance, Alzantot et al [39] measured the effectiveness of their genetic algorithm-based adversarial attacks using the attack success rate as a metric which eventually indicates NLP model robustness to adversarial examples.…”
Section: A Attack Success Ratementioning
confidence: 99%
“…As a metric, the attack success rate has been utilized in numerous research studies to determine the effectiveness of adversarial attacks on NLP models [12], [15], [39], [115], [116]. For instance, Alzantot et al [39] measured the effectiveness of their genetic algorithm-based adversarial attacks using the attack success rate as a metric which eventually indicates NLP model robustness to adversarial examples.…”
Section: A Attack Success Ratementioning
confidence: 99%
“…Firstly, the typical way of structural analysis is to design probe classifiers to analyze model characteristics, such as syntactic structural features (Elazar et al, 2021) and semantic features (Wu et al, 2021). Secondly, the main idea of behavioral studies is that design experiments that allow researchers to make inferences about computed representations based on the model's behavior, such as proposing various challenge sets that aim to cover specific, diverse phenomena, like systematicity exhaustivity (Gardner et al, 2020;Ravichander et al, 2021). Thirdly, for interactive visualization, neuron activation (Durrani et al, 2020), attention mechanisms (Hao et al, 2020) and saliency measures (Janizek et al, 2021) are three main standard visualization methods.…”
Section: Related Workmentioning
confidence: 99%
“…Annotation artifacts (Gururangan et al, 2018) are gaps present in a dataset that can lead to misleading interpretations of a model's performance on that dataset. To mitigate this, we evaluate F on contrast sets (Gardner et al, 2020), which are (mostly) labelchanging small perturbations on instances to understand the true local boundary of the dataset. Essentially, they help us understand if F has learnt any dataset-specific shortcuts.…”
Section: Contrast Set Testsmentioning
confidence: 99%
“…For contrast set tests, we use an OOD contrast set for sentiment analysis released by the authors of the original paper (Gardner et al, 2020), which are created for the Movies dataset. Furthermore, for functional tests, we use an OOD test suite (flight reviews) from the CheckList (Ribeiro et al, 2020) which contains both template-based instances to test linguistic capabilities, as well as real-world data (tweets).…”
Section: Tasks and Datasetsmentioning
confidence: 99%