Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume 2021
DOI: 10.18653/v1/2021.eacl-main.284
|View full text |Cite
|
Sign up to set email alerts
|

Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models

Abstract: Recently, it has been found that monolingual English language models can be used as knowledge bases. Instead of structural knowledge base queries, masked sentences such as "Paris is the capital of [MASK]" are used as probes. We translate the established benchmarks TREx and GoogleRE into 53 languages. Working with mBERT, we investigate three questions. (i) Can mBERT be used as a multilingual knowledge base? Most prior work only considers English. Extending research to multiple languages is important for diversi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
62
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 54 publications
(62 citation statements)
references
References 18 publications
0
62
0
Order By: Relevance
“…"Dante was born in __X__". Follow up studies have introduced improved prompts for eliciting such knowledge (Jiang et al, 2020b) as well as multilingual versions (Jiang et al, 2020a;Kassner et al, 2021). However, all these benchmarks assume a static view of the knowledge inside an LM, and consider all answers across time to be correct for a given query.…”
Section: Related Workmentioning
confidence: 99%
“…"Dante was born in __X__". Follow up studies have introduced improved prompts for eliciting such knowledge (Jiang et al, 2020b) as well as multilingual versions (Jiang et al, 2020a;Kassner et al, 2021). However, all these benchmarks assume a static view of the knowledge inside an LM, and consider all answers across time to be correct for a given query.…”
Section: Related Workmentioning
confidence: 99%
“…Knowledge Probing and Knowledge in BERT Our model's capability to memorize tracts of Wikipedia connects to past work on extracting knowledge from language models (Petroni et al, 2019); our task is easier because it is also possible to make type predictions based on context, giving our model a "backoff" capability. Recently, knowledge probing settings have been proposed for the multilingual setting as well (Kassner et al, 2021). Another related line of work attempts to add entity information into BERT (Zhang et al, 2019;Peters et al, 2019;Poerner et al, 2020;Févry et al, 2020, inter alia); our model could potentially benefit from this, but given our focus on generalizing to new entities, a model that more easily memorizes common entities may actually work less well.…”
Section: Related Workmentioning
confidence: 99%
“…shown promising results in cross-lingual entitycentric tasks (Vulić et al, 2020;Liu et al, 2021b;Kassner et al, 2021). Note that the "Mirror-BERT" fine-tuning step is always applied, yielding an increase in performance.…”
Section: Models In Comparisonmentioning
confidence: 99%
“…Task Setup. Since we are interested in probing how much knowledge a PLM contains in multiple languages, we use the multilingual LAnguage Model Analysis (mLAMA) benchmark proposed by Kassner et al (2021). To ensure a strictly fair comparison, we only compare XLM-R and Prix-LM.…”
Section: Prompt-based Knowledge Probingmentioning
confidence: 99%
See 1 more Smart Citation