Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.645
|View full text |Cite
|
Sign up to set email alerts
|

CamemBERT: a Tasty French Language Model

Abstract: Pretrained language models are now ubiquitous in Natural Language Processing. Despite their success, most available models have either been trained on English data or on the concatenation of data in multiple languages. This makes practical use of such models-in all languages except English-very limited. In this paper, we investigate the feasibility of training monolingual Transformer-based language models for other languages, taking French as an example and evaluating our language models on part-of-speech tagg… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
234
0
5

Year Published

2020
2020
2021
2021

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 491 publications
(242 citation statements)
references
References 36 publications
3
234
0
5
Order By: Relevance
“…In this study, linguistic representation will be used to process French data. Very recently, CamemBERT [26], FlauBERT [27] and GermanBERT 1 models were released for French and German while Ernie models are only available in Chinese and English. As far as the authors know, this is the first time that such models are used for feature extraction in order to perform SER tasks.…”
Section: Linguistic Representationmentioning
confidence: 99%
“…In this study, linguistic representation will be used to process French data. Very recently, CamemBERT [26], FlauBERT [27] and GermanBERT 1 models were released for French and German while Ernie models are only available in Chinese and English. As far as the authors know, this is the first time that such models are used for feature extraction in order to perform SER tasks.…”
Section: Linguistic Representationmentioning
confidence: 99%
“…A benchmark of NER models on French commercial legal cases has been developed (Benesty 2019). The results encourage the use of the NER bi-directional long short term memory (Bi-LSTM) model by using the Flair library (Akbik, Blythe, and Vollgraf 2018) and the NER model of CamemBERT (Martin et al 2020). CamemBERT is a French version of BERT -Bidirectional Encoder Representations from Transformers- (Devlin et al 2018), which is itself based on the encoder part of the Transformer architecture (Vaswani et al 2017).…”
Section: Introductionmentioning
confidence: 89%
“…In order to tackle specific languages problems, different monolingual versions of BERT were trained in different languages. For example BERTje [36] is a Dutch version, AlBERTo [37] is an Italian version, and CamemBERT [38] and FlauBERT [39] are two different models for French. These models outperform vanilla BERT in different NLP tasks specific to these languages.…”
Section: Specific Language Modelsmentioning
confidence: 99%