Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP 2021
DOI: 10.18653/v1/2021.blackboxnlp-1.26
|View full text |Cite
|
Sign up to set email alerts
|

Assessing the Generalization Capacity of Pre-trained Language Models through Japanese Adversarial Natural Language Inference

Abstract: Despite the success of multilingual pre-trained language models, it remains unclear to what extent these models have human-like generalization capacity across languages. The aim of this study is to investigate the out-ofdistribution generalization of pre-trained language models through Natural Language Inference (NLI) in Japanese, the typological properties of which are different from those of English. We introduce a synthetically generated Japanese NLI dataset, called the Japanese Adversarial NLI (JaNLI) data… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 33 publications
0
3
0
Order By: Relevance
“…There are few works for other languages [3,16,23,34], and particularly, for Vietnamese [17]. These works however focus on constructing open-domain datasets and also only for the monolingual problem.…”
Section: Related Workmentioning
confidence: 99%
“…There are few works for other languages [3,16,23,34], and particularly, for Vietnamese [17]. These works however focus on constructing open-domain datasets and also only for the monolingual problem.…”
Section: Related Workmentioning
confidence: 99%
“…The limited research on the robustness of multilingual models has primarily focused on being robust against specific types of noise, e.g., adversarial perturbations for Japanese Natural Language Inference (Yanaka and Mineshima, 2021), a combination of general and task-specific text transformations based on manipulating synonyms, antonyms, syntax, etc. , and introducing errors and noise through Wikipedia edits (Cooper Stickland et al, 2023).…”
Section: Cross-lingual Transfer Under Distribution Shiftmentioning
confidence: 99%
“…Templates We make templates and generate sentences with them using the codebase developed by 9 We provide examples for each condition in Appendix B. Yanaka and Mineshima (2021). 10 Following are examples of templates and sentences.…”
Section: Data Generationmentioning
confidence: 99%