Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.331
|View full text |Cite
|
Sign up to set email alerts
|

Word Frequency Does Not Predict Grammatical Knowledge in Language Models

Abstract: Neural language models learn, to varying degrees of accuracy, the grammatical properties of natural languages. In this work, we investigate whether there are systematic sources of variation in the language models' accuracy. Focusing on subject-verb agreement and reflexive anaphora, we find that certain nouns are systematically understood better than others, an effect which is robust across grammatical tasks and different language models. Surprisingly, we find that across four orders of magnitude, corpus freque… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(7 citation statements)
references
References 23 publications
1
6
0
Order By: Relevance
“…Noun frequency. We find similar evidence, like Yu et al (2020), that BERT did not perform better on nouns that were more frequent in the training set. Figure 8 shows these results-for each subject (we consider both singular and plural forms as a single subject), we plot that subject's error rate against its frequency in the training data.…”
Section: C2 Comparison With Prior Worksupporting
confidence: 71%
See 1 more Smart Citation
“…Noun frequency. We find similar evidence, like Yu et al (2020), that BERT did not perform better on nouns that were more frequent in the training set. Figure 8 shows these results-for each subject (we consider both singular and plural forms as a single subject), we plot that subject's error rate against its frequency in the training data.…”
Section: C2 Comparison With Prior Worksupporting
confidence: 71%
“…There has been substantial prior work on the ability of language models to perform abstract syntactic processing tasks (Hu et al, 2020) (see Linzen and Baroni (2020) for a review). On SVA specifically, Goldberg (2019) found that BERT achieves high accuracy on both natural sentences (97%) and nonce sentences (83%), and that error rate was independent of the number of "distractor" words between the subject and verb; Yu et al (2020) showed that language models do not exhibit better grammatical knowledge of more frequent nouns. Other work has found that BERT's performance is sensitive to factors that may suggest item-specific learning; Chaves and Richter (2021) found that BERT's performance on number agreement is sensitive to the verb, across seven different verbs, and Newman et al (2021) found that language models performed better on verbs that they predicted were likely in context.…”
Section: Syntactic Reasoning In Lmsmentioning
confidence: 99%
“…Similarly, we observe that models succeed on our MW metric indicating that they correctly inflect verbs with high in-context probability under the model. Relatedly, Yu et al (2020) investigate the nouns used in TSE minimal pairs and find that language model performance at subject-verb number agreement is uncorrelated with unigram probability of the noun. We instead focus on model-estimated in-context probability of the verb in minimal pairs, finding that model performance increases with the model probability.…”
Section: Lexical Choice In Syntactic Evaluationmentioning
confidence: 99%
“…There has been substantial prior work on the ability of language models to perform abstract syntactic processing tasks (Hu et al, 2020) (see Linzen and Baroni (2020) for a review). On SVA specifically, Goldberg (2019) found that BERT achieves high accuracy on both natural sentences (97%) and nonce sentences (83%), and that error rate was independent of the number of "distractor" words between the subject and verb; Yu et al (2020) showed that language models do not exhibit better grammatical knowledge of more frequent nouns. Other work has found that BERT's performance is sensitive to factors that may suggest item-specific learning; Chaves and Richter (2021) found that BERT's performance on number agreement is sensitive to the verb, across seven different verbs, and Newman et al (2021) found that language models performed better on verbs that they predicted were likely in context.…”
Section: Syntactic Reasoning In Lmsmentioning
confidence: 99%