Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics 2023
DOI: 10.18653/v1/2023.eacl-main.100
|View full text |Cite
|
Sign up to set email alerts
|

Robustification of Multilingual Language Models to Real-world Noise in Crosslingual Zero-shot Settings with Robust Contrastive Pretraining

Asa Cooper Stickland,
Sailik Sengupta,
Jason Krone
et al.

Abstract: Advances in neural modeling have achieved state-of-the-art (SOTA) results on public natural language processing (NLP) benchmarks, at times surpassing human performance. However, there is a gap between public benchmarks and real-world applications where noise, such as typographical or grammatical mistakes, is abundant and can result in degraded performance. Unfortunately, works which evaluate the robustness of neural models on noisy data and propose improvements, are limited to the English language. Upon analyz… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…The limited research on the robustness of multilingual models has primarily focused on being robust against specific types of noise, e.g., adversarial perturbations for Japanese Natural Language Inference (Yanaka and Mineshima, 2021), a combination of general and task-specific text transformations based on manipulating synonyms, antonyms, syntax, etc. , and introducing errors and noise through Wikipedia edits (Cooper Stickland et al, 2023). Unlike these works, we will evaluate how well zero-shot cross-lingual transfer from English to non-English test samples can generalize in scenarios where there is a shift in domain from train to test data: the domain-specific features of test samples may change, whereas the semantic sentiment features remain invariant.…”
Section: Cross-lingual Transfer Under Distribution Shiftmentioning
confidence: 99%
“…The limited research on the robustness of multilingual models has primarily focused on being robust against specific types of noise, e.g., adversarial perturbations for Japanese Natural Language Inference (Yanaka and Mineshima, 2021), a combination of general and task-specific text transformations based on manipulating synonyms, antonyms, syntax, etc. , and introducing errors and noise through Wikipedia edits (Cooper Stickland et al, 2023). Unlike these works, we will evaluate how well zero-shot cross-lingual transfer from English to non-English test samples can generalize in scenarios where there is a shift in domain from train to test data: the domain-specific features of test samples may change, whereas the semantic sentiment features remain invariant.…”
Section: Cross-lingual Transfer Under Distribution Shiftmentioning
confidence: 99%
“…Zhao et al (2021) study selection of instances in a few-shot scenario and show that methods are sensitive to the quality of the annotated data. Recently, Cooper Stickland et al (2023) proposed an effective pretraining strategy based on modeling typological, grammatical, or morphological noise in the data that boosts the cross-lingual zero-shot performance.…”
Section: Cross-lingual Transfermentioning
confidence: 99%