2020
DOI: 10.48550/arxiv.2006.05744
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MC-BERT: Efficient Language Pre-Training via a Meta Controller

Abstract: Pre-trained contextual representations (e.g., BERT) have become the foundation to achieve state-of-the-art results on many NLP tasks. However, large-scale pretraining is computationally expensive. ELECTRA, an early attempt to accelerate pre-training, trains a discriminative model that predicts whether each input token was replaced by a generator. Our studies reveal that ELECTRA's success is mainly due to its reduced complexity of the pre-training task: the binary classification (replaced token detection) is mo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(13 citation statements)
references
References 17 publications
0
13
0
Order By: Relevance
“…There are inevitable variations in the pretraining corpora used in different work. Our base corpus is obtained from the authors of MC-BERT (Xu et al, 2020) and TUPE (Ke et al, 2020). Our base++ corpus is the most similar with those used in UniLM (Dong et al, 2019;Bao et al, 2020).…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…There are inevitable variations in the pretraining corpora used in different work. Our base corpus is obtained from the authors of MC-BERT (Xu et al, 2020) and TUPE (Ke et al, 2020). Our base++ corpus is the most similar with those used in UniLM (Dong et al, 2019;Bao et al, 2020).…”
Section: Methodsmentioning
confidence: 99%
“…Implementation Details: Our implementation is built upon the open-source release of MC-BERT (Xu et al, 2020) and its ELECTRA reproduction based on fairseq (Ott et al, 2019). Standard hyperparameters in pretraining and finetuning are used.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations