Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.200
|View full text |Cite
|
Sign up to set email alerts
|

GiBERT: Enhancing BERT with Linguistic Information using a Lightweight Gated Injection Method

Abstract: Large pre-trained language models such as BERT have been the driving force behind recent improvements across many NLP tasks. However, BERT is only trained to predict missing words -either through masking or next sentence prediction -and has no knowledge of lexical, syntactic or semantic information beyond what it picks up through unsupervised pre-training. We propose a novel method to explicitly inject linguistic information in the form of word embeddings into any layer of a pre-trained BERT. When injecting co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 39 publications
0
1
0
Order By: Relevance
“…In addition, recently there have been studies [9], [10] showing that transformer-based models struggle or fail to capture rich knowledge. For this reason, there have been proposed methods for enhancing these models with external information or additional modalities [11], [12], [13], [14]. However, existing research initiatives in the tasks of stress and depression detection through social media have not exploited any of these approaches yet.…”
Section: Introductionmentioning
confidence: 99%
“…In addition, recently there have been studies [9], [10] showing that transformer-based models struggle or fail to capture rich knowledge. For this reason, there have been proposed methods for enhancing these models with external information or additional modalities [11], [12], [13], [14]. However, existing research initiatives in the tasks of stress and depression detection through social media have not exploited any of these approaches yet.…”
Section: Introductionmentioning
confidence: 99%