2021
DOI: 10.1162/tacl_a_00360
|View full text |Cite
|
Sign up to set email alerts
|

KEPLER: A Unified Model for Knowledge Embedding and Pre-trained Language Representation

Abstract: Pre-trained language representation models (PLMs) cannot well capture factual knowledge from text. In contrast, knowledge embedding (KE) methods can effectively represent the relational facts in knowledge graphs (KGs) with informative entity embeddings, but conventional KE models cannot take full advantage of the abundant textual information. In this paper, we propose a unified model for Knowledge Embedding and Pre-trained LanguagERepresentation (KEPLER), which can not only better integrate factual knowledge i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

3
248
1

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 398 publications
(317 citation statements)
references
References 34 publications
3
248
1
Order By: Relevance
“…For the choice of the Knowledge Base, we use a subset of Freebase 2 that includes 3 million entities with the most connections, similar to Xu and Barbosa (2019). For all pairs appearing in the test set of NYT10 (both positive and negative), we remove all links in the subset of Freebase to ensure that we will not memorise any relations between them For the Knowledge Base, we use the version of Wikidata 3 provided by Wang et al (2019b) (in particular the transductive split 4 ), containing approximately 5 million entities. Similarly to Freebase, we remove all links between pairs in the test set from the resulting KB, which contains approximately 20 million triples after pruning.…”
Section: Datasetsmentioning
confidence: 99%
“…For the choice of the Knowledge Base, we use a subset of Freebase 2 that includes 3 million entities with the most connections, similar to Xu and Barbosa (2019). For all pairs appearing in the test set of NYT10 (both positive and negative), we remove all links in the subset of Freebase to ensure that we will not memorise any relations between them For the Knowledge Base, we use the version of Wikidata 3 provided by Wang et al (2019b) (in particular the transductive split 4 ), containing approximately 5 million entities. Similarly to Freebase, we remove all links between pairs in the test set from the resulting KB, which contains approximately 20 million triples after pruning.…”
Section: Datasetsmentioning
confidence: 99%
“…5) (e.g. : (Cohen et al, 2020;Wang et al, 2019;Peters et al, 2019) 2 ) including the creation of many annotated data sets (e.g. : (Zhang et al, 2017;Alt et al, 2020;Mesquita et al, 2019;Elsahar et al, 2019)).…”
Section: Related Workmentioning
confidence: 99%
“…Neither other vocabularies in the static text graphs nor real-world facts (available in KGs) affect the final embeddings of tokens. On the other hand, while ENRIE (Zhang et al, 2019b) and KEPLER (Wang et al, 2019c) utilize KGs to reach an improved model, they do not employ other graphs derived from the corpus. Also, ERNIE does not learn graph-based embedding during representation learning and only adopts embeddings trained by TransE (Bordes et al, 2013).…”
Section: Related Workmentioning
confidence: 99%