Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.375
|View full text |Cite
|
Sign up to set email alerts
|

Structural Supervision Improves Few-Shot Learning and Syntactic Generalization in Neural Language Models

Abstract: Humans can learn structural properties about a word from minimal experience, and deploy their learned syntactic representations uniformly in different grammatical contexts. We assess the ability of modern neural language models to reproduce this behavior in English and evaluate the effect of structural supervision on learning outcomes. First, we assess few-shot learning capabilities by developing controlled experiments that probe models' syntactic nominal number and verbal argument structure generalizations fo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

1
0
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 34 publications
1
0
0
Order By: Relevance
“…The fact that tokens seen only a few times are generally expected to be able to take direct objects suggests a transitivity learning bias in the model. Such a bias would align with recent work assessing few-shot learning of syntactic categories, specifically Jumelet et al (2019), who hypothesize that models learn default category for number and gender, and Wilcox et al (2020), who provide data from few-shot learning tests that is consistent with the hypotheses in Jumelet et al (2019). Interestingly, the results form Wilcox et al (2020) also suggest that the models tested learn a default transitive category for verbs, although they test Recurrent Neural Network models, not transformers, so more careful cross model comparisons are needed.…”
Section: Psycholinguistic Assessment Resultssupporting
confidence: 52%
“…The fact that tokens seen only a few times are generally expected to be able to take direct objects suggests a transitivity learning bias in the model. Such a bias would align with recent work assessing few-shot learning of syntactic categories, specifically Jumelet et al (2019), who hypothesize that models learn default category for number and gender, and Wilcox et al (2020), who provide data from few-shot learning tests that is consistent with the hypotheses in Jumelet et al (2019). Interestingly, the results form Wilcox et al (2020) also suggest that the models tested learn a default transitive category for verbs, although they test Recurrent Neural Network models, not transformers, so more careful cross model comparisons are needed.…”
Section: Psycholinguistic Assessment Resultssupporting
confidence: 52%