Proceedings of the 11th Joint Conference on Lexical and Computational Semantics 2022
DOI: 10.18653/v1/2022.starsem-1.4
|View full text |Cite
|
Sign up to set email alerts
|

Compositional generalization with a broad-coverage semantic parser

Abstract: We show how the AM parser, a compositional semantic parser (Groschwitz et al., 2018), can solve compositional generalization on the COGS dataset. It is the first semantic parser that achieves high accuracy on both naturally occurring language and the synthetic COGS dataset. We discuss implications for corpus and model design for learning human-like generalization. Our results suggest that compositional generalization can be best achieved by building compositionality into semantic parsers.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
15
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 9 publications
(15 citation statements)
references
References 18 publications
0
15
0
Order By: Relevance
“…For small-scale synthetic data, many specialized model architectures proved to be effective on SCAN like tasks Russin et al, 2019;Gordon et al, 2020;Lake, 2019;Liu et al, 2020a;Nye et al, 2020;Chen et al, 2020). To also address natural language variations in non-synthetic tasks, some recent works exploit structure of the source input and its relation to the target side (Herzig and Berant, 2021;Weißenhorn et al, 2022), and employ sourceside parsing that can be computationally demanding for long sentences, and may have coverage challenge and not available in all languages; while we try to exploit target-side structure only for higher efficiency. Some other works leverage source-side structure for data augmentation to overcome distribution divergence (Yang et al, 2022b;Qiu et al, 2022), which can clearly help but is not the focus of this paper.…”
Section: Modelingmentioning
confidence: 99%
“…For small-scale synthetic data, many specialized model architectures proved to be effective on SCAN like tasks Russin et al, 2019;Gordon et al, 2020;Lake, 2019;Liu et al, 2020a;Nye et al, 2020;Chen et al, 2020). To also address natural language variations in non-synthetic tasks, some recent works exploit structure of the source input and its relation to the target side (Herzig and Berant, 2021;Weißenhorn et al, 2022), and employ sourceside parsing that can be computationally demanding for long sentences, and may have coverage challenge and not available in all languages; while we try to exploit target-side structure only for higher efficiency. Some other works leverage source-side structure for data augmentation to overcome distribution divergence (Yang et al, 2022b;Qiu et al, 2022), which can clearly help but is not the focus of this paper.…”
Section: Modelingmentioning
confidence: 99%
“…A bunch of synthetically-generated datasets have been created for assessing compositional generalization (Lake and Baroni, 2017;Bastings et al, 2018;Keysers et al, 2020;Kim and Linzen, 2020), and plain sequence-to-sequence(seq2seq) models exhibit significant out-of-distribution (OOD) compositional generalization performance loss comparing to in-distribution (ID) setting. While effective methods have been proposed to overcome the difficulty in OOD compositional generalization Nye et al, 2020;Weißenhorn et al, 2022), most of them mainly focus on semantic parsing, where some important abilities like summarization Dataset # of samples generalization forms GEOQUERY (Shaw et al, 2021) 880 3 SPIDER-SSP (Shaw et al, 2021) 4,376 3 SMCALFLOW-CS (Yin et al, 2021) 28,054 2 COUNTERFACTUAL (Liu et al, 2022) 2,500 1 DINER (ours) 223,581 4…”
Section: Instructionsmentioning
confidence: 99%
“…Moreover, large pre-trained language models have been shown not to improve compositional generalization (Oren et al, 2020;Qiu et al, 2022b). This has prompted the community to realize that parsers should be designed intentionally with compositionality in mind (Lake, 2019;Gordon et al, 2020;Weißenhorn et al, 2022).…”
Section: Related Workmentioning
confidence: 99%