2021
DOI: 10.1007/s12652-021-03297-4
|View full text |Cite
|
Sign up to set email alerts
|

Assessing BERT’s ability to learn Italian syntax: a study on null-subject and agreement phenomena

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 34 publications
(7 citation statements)
references
References 43 publications
0
7
0
Order By: Relevance
“…The previous research has provided us with references regarding the research subjects (pre-trained language models for ancient Chinese) and methods (complex networks). However, most studies [6][7][8][9][10][11][12] on probing language models have not comprehensively elucidated the patterns and principles about the organization of linguistic elements within the models. Therefore, we use complex networks to comprehensively study how these pre-trained language models organize linguistic elements at different levels.…”
Section: Complex Network In Language Systemmentioning
confidence: 99%
See 1 more Smart Citation
“…The previous research has provided us with references regarding the research subjects (pre-trained language models for ancient Chinese) and methods (complex networks). However, most studies [6][7][8][9][10][11][12] on probing language models have not comprehensively elucidated the patterns and principles about the organization of linguistic elements within the models. Therefore, we use complex networks to comprehensively study how these pre-trained language models organize linguistic elements at different levels.…”
Section: Complex Network In Language Systemmentioning
confidence: 99%
“…Thus, it is necessary to investigate the mechanisms by which these models ‘comprehend’ ancient Chinese. In fact, previous work has explored how pre-trained language models ‘learn’ linguistic knowledge, including probing lexical [6,7], syntactic [8,9] and semantic [10–12] knowledge encoded in the models. However, most of these works focus on probing the knowledge at certain aspects, without adopting a holistic perspective to study how the models simulate human language, or to discern potential patterns in how they organize the elements of ancient Chinese.…”
Section: Introductionmentioning
confidence: 99%
“…For instance, [23] uses it for sentiment analysis. [16] evaluates a Transformer's ability to learn Italian syntax. Finally, [6] proposes a chatbot that helps detect and classify fraud in a finance context.…”
Section: Introductionmentioning
confidence: 99%
“…The study by (de Vries, van Cranenburgh, and Nissim 2020) represents an exception in this context: the authors applied the probing task approach to compare the linguistic competence encoded by a Dutch BERT-based model and multilingual BERT (mBERT), showing that earlier layers of mBERT are consistently more informative that earlier layers of the monolingual model. (Guarasci et al 2021) applied instead the structural probe originally defined by (Hewitt and Manning 2019) on the representations of a pre-trained Italian BERT. Testing their approach on different subsets of the Italian Universal Dependency Treebank (IUDT), they showed on the one hand that the model is able to encode properties of syntax especially in its central-upper layers; on the other hand, that such embedded syntactic information can be used to successfully perform two specific syntactic tasks, i.e.…”
Section: Introductionmentioning
confidence: 99%