2021
DOI: 10.1155/2021/7115167
|View full text |Cite
|
Sign up to set email alerts
|

A General Method for Transferring Explicit Knowledge into Language Model Pretraining

Abstract: Recently, pretrained language models, such as Bert and XLNet, have rapidly advanced the state of the art on many NLP tasks. They can model implicit semantic information between words in the text. However, it is solely at the token level without considering the background knowledge. Intuitively, background knowledge influences the efficacy of text understanding. Inspired by this, we focus on improving model pretraining by leveraging external knowledge. Different from recent research that optimizes pretraining m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 23 publications
0
1
0
Order By: Relevance
“…Explicit knowledge integration of knowledge resources into language models can be roughly categorized into fusion based approaches and language modeling based approaches. Fusion based approaches (Peters et al, 2019;Wang et al, 2021;Yan et al, 2021) typically perform knowledge integration by combining language model representations with representations extracted from knowledge bases. Compared to language modeling based approaches, as explored by us, they rely on aligned data and are typically applied during the pre-training stage.…”
Section: Related Workmentioning
confidence: 99%
“…Explicit knowledge integration of knowledge resources into language models can be roughly categorized into fusion based approaches and language modeling based approaches. Fusion based approaches (Peters et al, 2019;Wang et al, 2021;Yan et al, 2021) typically perform knowledge integration by combining language model representations with representations extracted from knowledge bases. Compared to language modeling based approaches, as explored by us, they rely on aligned data and are typically applied during the pre-training stage.…”
Section: Related Workmentioning
confidence: 99%