Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP 2021
DOI: 10.18653/v1/2021.blackboxnlp-1.9
|View full text |Cite
|
Sign up to set email alerts
|

Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN

Abstract: Despite their failure to solve the compositional SCAN dataset, seq2seq architectures still achieve astonishing success on more practical tasks. This observation pushes us to question the usefulness of SCAN-style compositional generalization in realistic NLP tasks. In this work, we study the benefit that such compositionality brings about to several machine translation tasks. We present several focused modifications of Transformer that greatly improve generalization capabilities on SCAN and select one that rema… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 14 publications
(16 citation statements)
references
References 21 publications
2
14
0
Order By: Relevance
“…Although we learnt about mechanics involved in idiomatic translations, the vast majority of translations was still word for word, indicating that noncompositional processing does not emerge well (enough) in Transformer. Paradoxically, a recent trend is to encourage more compositional processing in NMT (Chaabouni et al, 2021;Li et al, 2021;Raunak et al, 2019, i.a.). We recommend caution since this inductive bias may harm idiom translations.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Although we learnt about mechanics involved in idiomatic translations, the vast majority of translations was still word for word, indicating that noncompositional processing does not emerge well (enough) in Transformer. Paradoxically, a recent trend is to encourage more compositional processing in NMT (Chaabouni et al, 2021;Li et al, 2021;Raunak et al, 2019, i.a.). We recommend caution since this inductive bias may harm idiom translations.…”
Section: Discussionmentioning
confidence: 99%
“…These patterns are stronger for figurative PIEs that the model paraphrases than for sentences that receive an overly compositional translation and hold across the seven European languages. Considering that a recent trend in NLP is to encourage even more compositional processing in NMT (Raunak et al, 2019;Chaabouni et al, 2021;Li et al, 2021, i.a. ), we recommend caution.…”
Section: Introductionmentioning
confidence: 99%
“…Ruis, Burghouts, & Bucur, 2021), natural language processing Baroni, 2020;Keysers et al, 2020;Kim & Linzen, 2020), and more generally (Nam & McClelland, 2021). Two fundamentally different approaches are taken by the literature; one utilizes additional data while making few changes to the conventional setup and architecture (Furrer, van Zee, Scales, & Schärli, 2020), while the other utilizes additional inductive biases that aim to support systematic generalization (Russin et al, 2019;Lake, 2019;Andreas, 2020;Nye et al, 2020;Gordon et al, 2020;Bogin et al, 2021;Chaabouni, Dessì, & Kharitonov, 2021). In this work we apply both approaches, the former through data augmentation, and the latter through high-level modularity.…”
Section: Related Workmentioning
confidence: 99%
“…The first prominent type of generalisation that can be found in the literature is compositional generalisation, which is often argued to underpin human's ability to quickly generalise to new data, tasks and domains (Fodor and Pylyshyn, 1988;Lake et al, 2017;Marcus, 2018;Schmidhuber, 1990). Because of this strong connection with humans and human language, work on compositional generalisation often has a primarily cognitive motivation, although practical concerns such as sample efficiency, quick adaptation and good generalisation in low-resource scenarios are frequently mentioned as additional or alternative motivations (Chaabouni et al, 2021;Linzen, 2020, to give just a few examples). While it has a strong intuitive appeal and clear mathematical definition (Montague, 1970), compositional generalisation is not easy to pin down empirically.…”
Section: Compositional Generalisationmentioning
confidence: 99%