2018
DOI: 10.48550/arxiv.1805.09657
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning compositionally through attentive guidance

Abstract: While neural network models have been successfully applied to domains that require substantial generalisation skills, recent studies have implied that they struggle when solving the task they are trained on requires inferring its underlying compositional structure. In this paper, we introduce Attentive Guidance, a mechanism to direct a sequence to sequence model equipped with attention to find more compositional solutions. We test it on two tasks, devised precisely to assess the compositional capabilities of n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
9
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2

Relationship

2
4

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 8 publications
1
9
0
Order By: Relevance
“…Our method draws inspiration from the work on compositional learning of Hupkes et al (2018a). The authors introduce the concept of Attentive Guidance, a training signal given to the attention mechanism of a seq2seq model to induce more compositional solutions.…”
Section: Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…Our method draws inspiration from the work on compositional learning of Hupkes et al (2018a). The authors introduce the concept of Attentive Guidance, a training signal given to the attention mechanism of a seq2seq model to induce more compositional solutions.…”
Section: Modelsmentioning
confidence: 99%
“…We borrow the setup presented in Hupkes et al (2018a), which differs slightly from the setup as it was originally presented. In this setup, a typical input output example could be 001 t1 t2 → 001 010 111.…”
Section: Taskmentioning
confidence: 99%
See 1 more Smart Citation
“…Our findings indicate the following. Firstly, as hypothesized before (Dessì and Baroni, 2019;Hupkes et al, 2018), the limited attention span provides a useful inductive bias that allows models to perform better on compositional generalization induction, that SCAN probes for. Further, endowing a model with SCAN-style generalization capabilities can lead to improvements in low-resource and distribution-shifted scenarios as long as we ensure that we do not overfit to SCAN.…”
Section: Discussionmentioning
confidence: 99%
“…Phrase and sentence composition has drawn frequent attention in analysis of neural models, often focusing on analysis of internal representations and downstream task behavior (Ettinger et al, 2018;Conneau et al, 2019;Nandakumar et al, 2019;Yu and Ettinger, 2020;Bhathena et al, 2020;Mu and Andreas, 2020; 1 Datasets and code available at https://github.com/yulang/fine-tuning-and-compositionin-transformers Andreas, 2019). Some work investigates compositionality via constructing linguistic (Keysers et al, 2019) and non-linguistic (Liška et al, 2018;Hupkes et al, 2018;Baan et al, 2019) synthetic datasets.…”
Section: Related Workmentioning
confidence: 99%