2022
DOI: 10.20944/preprints202208.0238.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Simple Survey of Pre-trained Language Models

Abstract: Pre-trained Language Models (PTLM) have remarkable and successful performance in solving lots of NLP tasks nowadays. And previous researchers have created many SOTA models and these models are included in many long surveys(Qiu et al., 2020). So, we would like to conduct a simple and short survey on this topic to help researchers understand the sketch of PTLM more quickly and comprehensively. In this short survey, we would provide a simple but comprehensive review of techniques, benchmarks, and methodologies in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 49 publications
0
3
0
Order By: Relevance
“…Currently, many pre‐training tasks can extract a fixed‐dimensional speaker embedding as a representation of speaker identity from the enormous amount of information in speech, such as the large language model [18]. In other words, the speaker encoder can generate a vector from the target audio that can meaningfully represent the speaker's identity.…”
Section: Introductionmentioning
confidence: 99%
“…Currently, many pre‐training tasks can extract a fixed‐dimensional speaker embedding as a representation of speaker identity from the enormous amount of information in speech, such as the large language model [18]. In other words, the speaker encoder can generate a vector from the target audio that can meaningfully represent the speaker's identity.…”
Section: Introductionmentioning
confidence: 99%
“…It has also been reported that ChatGPT could have passed a practice US bar exam with a headline score of 70% (35/50) 4 . If such generalpurpose models are further enhanced by incorporating legal domain knowledge [24], they will be even more proficient at legal tasks.…”
Section: Presentmentioning
confidence: 99%
“…This results in a more compact architecture, reducing computational requirements. • Compression (Zhu et al, 2023c) and Quantization (Tao et al, 2022): reduce the size of the model or reduce the precision of model parameters, using fewer bits to represent them. This reduces memory usage and computational complexity.…”
Section: Efficiency and Costmentioning
confidence: 99%