Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.223
|View full text |Cite
|
Sign up to set email alerts
|

Game-theoretic Vocabulary Selection via the Shapley Value and Banzhaf Index

Abstract: The input vocabulary and their learned representations are crucial to the performance of neural NLP models. Using the full vocabulary results in less explainable and more memory intensive models, with the embedding layer often constituting the majority of model parameters. It is thus common to use a smaller vocabulary to lower memory requirements and construct more interpertable models.We propose a vocabulary selection method that views words as members of a team trying to maximize the model's performance. We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 45 publications
0
5
0
Order By: Relevance
“…Moreover, SHAP has a fast implementation for tree-based models. Although Shapley value computation requires exponential time complexity, machine learning applications employ Shapley value approximation methods, such as Monte Carlo permutation sampling, which approximates Shapley value in linear time [41][42][43].…”
Section: Shapley Valuesmentioning
confidence: 99%
“…Moreover, SHAP has a fast implementation for tree-based models. Although Shapley value computation requires exponential time complexity, machine learning applications employ Shapley value approximation methods, such as Monte Carlo permutation sampling, which approximates Shapley value in linear time [41][42][43].…”
Section: Shapley Valuesmentioning
confidence: 99%
“…The wide class of semivalues applies to many different realworld applications, such as voting [28], interpretable machine learning [26,27], reinforcement learning [24,14], text summarisation [32], etc. The equality of weights over coalition sizes is not a constraint for the application, but a choice of design that allows the semivalues to take into account the influence of a player on groups of different sizes while preserving anonymity and symmetry.…”
Section: A Semivalues As Weighted Average Marginal Contributionsmentioning
confidence: 99%
“…Lastly, in §3.6 we mentioned how we used Shapley values from cooperative game theory to optimally select features, in this case vocabularies, in language tasks [84].…”
Section: Gamificationmentioning
confidence: 99%