2020
DOI: 10.48550/arxiv.2002.01808
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters

Abstract: We study the problem of injecting knowledge into large pre-trained models like BERT and RoBERTa. Existing methods typically update the original parameters of pre-trained models when injecting knowledge. However, when multiple kinds of knowledge are injected, they may suffer from catastrophic forgetting. To address this, we propose K-ADAPTER, which remains the original parameters of the pre-trained model fixed and supports continual knowledge infusion. Taking RoBERTa as the pre-trained model, K-ADAPTER has a ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
60
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 105 publications
(60 citation statements)
references
References 33 publications
0
60
0
Order By: Relevance
“…These encyclopedia KG are able to provide abundant knowledge for PLMs to integrate. A majority of exiting work on KE-PLMs [108][26][105] [96][90] [82][66] uses Wikidata 1 as knowledge source. Typically, entities in Wikidata are linked with entity mentions in the text of Wikipedia.…”
Section: Encyclopedia Knowledgementioning
confidence: 99%
See 4 more Smart Citations
“…These encyclopedia KG are able to provide abundant knowledge for PLMs to integrate. A majority of exiting work on KE-PLMs [108][26][105] [96][90] [82][66] uses Wikidata 1 as knowledge source. Typically, entities in Wikidata are linked with entity mentions in the text of Wikipedia.…”
Section: Encyclopedia Knowledgementioning
confidence: 99%
“…Though most existing approaches exploit only one knowledge source, it is worth noting that certain methods attempt to incorporate from more than one knowledge source. For example, K-Adapter [96] incorporate knowledge from multiple sources by learning a different adapter for each knowledge source. It exploits both dependency relation as linguistic knowledge and relation/fact knowledge from Wikidata.…”
Section: Sentiment Knowledgementioning
confidence: 99%
See 3 more Smart Citations