Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.722
|View full text |Cite
|
Sign up to set email alerts
|

Exploiting Structured Knowledge in Text via Graph-Guided Representation Learning

Abstract: In this work, we aim at equipping pre-trained language models with structured knowledge. We present two self-supervised tasks learning over raw text with the guidance from knowledge graphs. Building upon entity-level masked language models, our first contribution is an entity masking scheme that exploits relational knowledge underlying the text. This is fulfilled by using a linked knowledge graph to select informative entities and then masking their mentions. In addition, we use knowledge graphs to obtain dist… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
36
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 37 publications
(36 citation statements)
references
References 38 publications
0
36
0
Order By: Relevance
“…What lexical cues or correlations should be allowed in knowledge probes? Progress in this direction will Masked token prediction Guu et al, 2020) Contrastive learning (Xiong et al, 2020;Shen et al, 2020) Text-to-KB links-late fusion ( § 4.2)…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…What lexical cues or correlations should be allowed in knowledge probes? Progress in this direction will Masked token prediction Guu et al, 2020) Contrastive learning (Xiong et al, 2020;Shen et al, 2020) Text-to-KB links-late fusion ( § 4.2)…”
Section: Discussionmentioning
confidence: 99%
“…Contrastive learning techniques, which have been used for LM supervision at the word and sentence level , have also been devised for supervision on entity mentions (Shen et al, 2020). For example, Xiong et al ( 2020) replace a proportion of entity mentions in the pretraining corpus with the names of negatively-sampled entities of the same type, and train an LM to predict whether the entity in the span has been replaced (Figure 3b).…”
Section: Masking Tokens In Mention-spans and Trainingmentioning
confidence: 99%
See 1 more Smart Citation
“…Bilinear AVG (Li et al, 2016) 91.70 DNN AVG (Li et al, 2016) 92.00 DNN LSTM (Li et al, 2016) 89.20 DNN AVG + CKBG (Saito et al, 2018) 94.70 Factorized (Jastrzębski et al, 2018) 79.40 Prototypical (Jastrzębski et al, 2018) 89.00 Concatenation (Davison et al, 2019) 68.80 Template (Davison et al, 2019) 72.20 Template + Grammar (Davison et al, 2019) 74.40 Coherency Ranking (Davison et al, 2019) 78.80 KG-BERT BERT-BASE (Shen et al, 2020) 93.20 KG-BERT GLM(RoBERTa-LARGE) (Shen et al, 2020) 94.60…”
Section: Datamentioning
confidence: 99%
“…A series of recent work explores incorporating structured knowledge embedded in text into MRC (Shen et al, 2020;Dhingra et al, 2020). However, such kind of linking information for creating triples is not necessarily prominent in documents other than Wikipedia.…”
Section: Sequence Viewmentioning
confidence: 99%