2020
DOI: 10.48550/arxiv.2009.14304
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Cross-lingual Alignment Methods for Multilingual BERT: A Comparative Study

Abstract: Multilingual BERT (mBERT) has shown reasonable capability for zero-shot cross-lingual transfer when fine-tuned on downstream tasks. Since mBERT is not pre-trained with explicit cross-lingual supervision, transfer performance can further be improved by aligning mBERT with cross-lingual signal. Prior work proposes several approaches to align contextualised embeddings. In this paper we analyse how different forms of cross-lingual supervision and various alignment methods influence the transfer capability of mBERT… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 18 publications
0
4
0
Order By: Relevance
“…Contextual representations can be obtained through multilingual pre-training, which encodes whole sentence and outputs contextual representation for each word (Devlin et al, 2019;Lample and Conneau, 2019). Due to the rich context information contained in the contextual representations, there are endeavors to align them in different languages (Schuster et al, 2019;Aldarmaki and Diab, 2019;Wang et al, 2020;Kulshreshtha et al, 2020;Cao et al, 2020).…”
Section: • Contextual Representation Based Methodsmentioning
confidence: 99%
“…Contextual representations can be obtained through multilingual pre-training, which encodes whole sentence and outputs contextual representation for each word (Devlin et al, 2019;Lample and Conneau, 2019). Due to the rich context information contained in the contextual representations, there are endeavors to align them in different languages (Schuster et al, 2019;Aldarmaki and Diab, 2019;Wang et al, 2020;Kulshreshtha et al, 2020;Cao et al, 2020).…”
Section: • Contextual Representation Based Methodsmentioning
confidence: 99%
“…Cross-lingual Transfer: In an endeavor to enhance the cross-lingual transfer abilities of Multilingual BERT (mBERT), Kulshreshtha et al (2020) managed to improve its performance. The researchers achieved this by aligning mBERT with cross-lingual signals, employing parallel corpora supervision, and fine-tuning the alignment.…”
Section: Evolution Of Language Modeling For Armenian Languagementioning
confidence: 99%
“…In a related effort, Ter-Hovhannisyan and Avetisyan (2022) utilized the transformer-based XLM-RoBERTa model for cross-lingual sentence alignment, highlighting its effectiveness in multilingual contexts. Additionally, Kulshreshtha et al (2020) augmented the cross-lingual transfer abilities of Multilingual BERT (mBERT) to achieve superior performance in language transfer tasks.…”
Section: The Utilization Of Llms In Armenian Nlp Tasksmentioning
confidence: 99%
“…In addition, BERT has been incorporated into NMT models as a pre-training mechanism, leading to improved translation quality in various settings [24]. Moreover, several BERT-based models, such as mBERT (multilingual BERT) and XLM-R (Cross-lingual Language Model-RoBERTa), have been developed to handle multilingual and cross-lingual tasks [25,26].…”
Section: Machine Translation Approaches and Evolutionmentioning
confidence: 99%