2022
DOI: 10.1007/978-981-16-6723-7_23
|View full text |Cite
|
Sign up to set email alerts
|

A Literature Review on Bidirectional Encoder Representations from Transformers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1
1

Relationship

2
8

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 7 publications
0
6
0
Order By: Relevance
“…As another big trend, techniques that have been successfully used in natural language processing have been introduced. One of them is the use of (multi-head) self-attention mechanism, which was first introduced in Transformer (reviewed in Shreyashree et al, 2022 ). Both Jiang et al and Cong et al report the improvement of prediction performance with the use of the self-attention mechanism ( Jiang, Wang, Yao, et al, 2021 ; Cong et al, 2022 ).…”
Section: Deep Learning and Language Model-based Methodsmentioning
confidence: 99%
“…As another big trend, techniques that have been successfully used in natural language processing have been introduced. One of them is the use of (multi-head) self-attention mechanism, which was first introduced in Transformer (reviewed in Shreyashree et al, 2022 ). Both Jiang et al and Cong et al report the improvement of prediction performance with the use of the self-attention mechanism ( Jiang, Wang, Yao, et al, 2021 ; Cong et al, 2022 ).…”
Section: Deep Learning and Language Model-based Methodsmentioning
confidence: 99%
“…These and many other pre-trained models have one disadvantage in common, which is the unidirectional that restricts the power of those models. This led to the discovery of a specific type of transformer networks [34], BERT [35][36][37]. It is a pre-trained model that considers the left and right context of a word in all layers, while generating embeddings.…”
Section: A Architecture Of Proposed and Related Modelsmentioning
confidence: 99%
“…Shreyashree et al [17] presented "transfer learning" which is a method of creating a model for a specific problem and then utilizing it to create a model for a different problem. It has been proven to be quite successful.…”
Section: Introductionmentioning
confidence: 99%