Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.254
|View full text |Cite
|
Sign up to set email alerts
|

Pareto Probing: Trading Off Accuracy for Complexity

Abstract: The question of how to probe contextual word representations for linguistic structure in a way that is both principled and useful has seen significant attention recently in the NLP literature. In our contribution to this discussion, we argue for a probe metric that reflects the fundamental trade-off between probe complexity and performance: the Pareto hypervolume. To measure complexity, we present a number of parametric and non-parametric metrics. Our experiments using Pareto hypervolume as an evaluation metri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
68
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 46 publications
(69 citation statements)
references
References 32 publications
1
68
0
Order By: Relevance
“…Other works take an informationtheoretic view: Voita and Titov (2020) measure the complexity of the probe in terms of the bits needed to transmit its parameters, while Pimentel et al (2020b) argue that probing should measure mutual information between the representation and the property. Pimentel et al (2020a) propose a Pareto approach where they plot accuracy versus probe complexity, unifying several of these goals. We use these proposed metrics to compare our probing method to standard probing approaches.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Other works take an informationtheoretic view: Voita and Titov (2020) measure the complexity of the probe in terms of the bits needed to transmit its parameters, while Pimentel et al (2020b) argue that probing should measure mutual information between the representation and the property. Pimentel et al (2020a) propose a Pareto approach where they plot accuracy versus probe complexity, unifying several of these goals. We use these proposed metrics to compare our probing method to standard probing approaches.…”
Section: Related Workmentioning
confidence: 99%
“…We first vary the complexity of each probe, where for subnetwork probing we associate multiple encoder weights with a single mask, 4 and for the MLP probe we restrict the rank of the hidden layer. We then plot the resulting accuracy-complexity curve (Pimentel et al, 2020a).…”
Section: Probe Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…Early probing studies in NLP include (Zhang and Bowman, 2018) and (Tenney et al, 2019c), the former being an early example of the importance of comparing with randomized representations or labels. Further discussion has introduced control tasks and the selectivity metric (Hewitt and Liang, 2019), formalized notions of ease of extraction (Voita and Titov, 2020) and described other strategies for taking model complexity into account (Pimentel et al, 2020a).…”
Section: Related Workmentioning
confidence: 99%
“…Contours belong to our probe. Conneau et al, 2018) is one prominent method, which consists of using a lightly parameterized model to predict linguistic phenomena from intermediate representations, albeit recent work has raised concerns on how model parameterization and evaluation metrics may affect the effectiveness of this approach (Hewitt and Liang, 2019;Pimentel et al, 2020b;Maudslay et al, 2020;Pimentel et al, 2020a). Most work in intrinsic probing has focused in the identification of individual neurons that are important for a task (Li et al, 2016;Kádár et al, 2017;Li et al, 2017;Lakretz et al, 2019).…”
Section: Related Workmentioning
confidence: 99%