Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.196
|View full text |Cite
|
Sign up to set email alerts
|

Improving AMR Parsing with Sequence-to-Sequence Pre-training

Abstract: In the literature, the research on abstract meaning representation (AMR) parsing is much restricted by the size of human-curated dataset which is critical to build an AMR parser with good performance. To alleviate such data size restriction, pre-trained models have been drawing more and more attention in AMR parsing. However, previous pre-trained models, like BERT, are implemented for general purpose which may not work as expected for the specific task of AMR parsing. In this paper, we focus on sequence-to-seq… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
44
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 48 publications
(45 citation statements)
references
References 35 publications
(43 reference statements)
1
44
0
Order By: Relevance
“…Continuing this latter direction and arguably closest to our work, Xu et al (2020a) and Bevilacqua et al (2021) show that these models, once paired with adequate pre-training, can behave on par or better than dedicated and more sophisticated graph-based alternatives, surpassing the performances of Cai and Lam (2020). In particular, similarly to our work, Xu et al (2020a) leverage a multilingual framework inspired by Johnson et al (2017) and explore the possibility of pre-training on a range of related tasks, including MT; however, their focus is limited to showing the effectiveness of transfer learning from related tasks to English AMR parsing.…”
Section: Related Worksupporting
confidence: 63%
See 1 more Smart Citation
“…Continuing this latter direction and arguably closest to our work, Xu et al (2020a) and Bevilacqua et al (2021) show that these models, once paired with adequate pre-training, can behave on par or better than dedicated and more sophisticated graph-based alternatives, surpassing the performances of Cai and Lam (2020). In particular, similarly to our work, Xu et al (2020a) leverage a multilingual framework inspired by Johnson et al (2017) and explore the possibility of pre-training on a range of related tasks, including MT; however, their focus is limited to showing the effectiveness of transfer learning from related tasks to English AMR parsing.…”
Section: Related Worksupporting
confidence: 63%
“…We report the Smatch and fine-grained scores that SGL and its current state-of-the-art alternatives attain on AMR-2.0 in Xu et al (2020a). Considering the similarity between the two approaches, this difference is likely caused by the increased number of tasks our model is asked to handle.…”
Section: Amr Parsingmentioning
confidence: 99%
“…Since then, a new state-of-the-art has been established for English AMR, using sequenceto-sequence transduction (Zhang et al, 2019a,b) and iterative inference with graph encoding Lam, 2019, 2020). Xu et al (2020a) improved sequence-to-sequence parsing for AMR by using pre-trained encoders, reaching similar performance to Cai and Lam (2020). introduced a stack-transformer to enhance transitionbased AMR parsing (Ballesteros and Al-Onaizan, 2017), and Lee et al (2020) improved it further, using a trained parser for mining oracle actions and combining it with AMR-to-text generation to outperform the state of the-art.…”
Section: Overview Of Approachesmentioning
confidence: 99%
“…Model Size In Table 3, we compare parameter sizes of recently published models alongside their parsing performances on AMR 2.0. Similar to our approach, most models use large pre-trained models to extract contextualized embeddings as fixed features, with the exception of Xu et al (2020), which is a seq-to-seq pre-training approach on large amount of data, and Bevilacqua et al (2021), which directly fine-tunes a seq-to-seq BART large (Lewis et al, 2019) model. 7 Except the large BART model, our APT small (3 layers) has the least number of trained parameters yet already surpasses all the previous models.…”
Section: Resultsmentioning
confidence: 99%