2021
DOI: 10.1007/978-3-030-83527-9_17
|View full text |Cite
|
Sign up to set email alerts
|

RobeCzech: Czech RoBERTa, a Monolingual Contextualized Language Representation Model

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(14 citation statements)
references
References 24 publications
0
8
0
Order By: Relevance
“…The modifications included a larger training process, to remove the next sentence prediction objective, use of longer sequences and a dynamic use of the mask applied over the training data. As in the case of BERT, RoBERTa has been fine‐tuned for specific languages such as dutch (Delobelle et al, 2020) or czech (Straka et al, 2021) and for many specific problems such as metaphor identification (Babieno et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…The modifications included a larger training process, to remove the next sentence prediction objective, use of longer sequences and a dynamic use of the mask applied over the training data. As in the case of BERT, RoBERTa has been fine‐tuned for specific languages such as dutch (Delobelle et al, 2020) or czech (Straka et al, 2021) and for many specific problems such as metaphor identification (Babieno et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…They further propose that, with the only little tuning of hyperparameters, the model outperformed all other tested German and multilingual BERT models. A monolingual RoBERTa language model trained on Czech data has been presented in [21]. The authors show that the model significantly outperforms equally-sized multilingual and Czech language-oriented model variants.…”
Section: Related Workmentioning
confidence: 99%
“…Finally, we have performed a round of experiments with a pair of recently published Czech monolingual models. RobeCzech [48] was pretrained on a set of currated Czech corpora using the RoBERTa-base architecture. FERNET-C5 [49] was pretrained on a large crawled dataset, using Bert-base architecture.…”
Section: Natural Language Inferencementioning
confidence: 99%