Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.371
|View full text |Cite
|
Sign up to set email alerts
|

Character-level Representations Improve DRS-based Semantic Parsing Even in the Age of BERT

Abstract: We combine character-level and contextual language model representations to improve performance on Discourse Representation Structure parsing. Character representations can easily be added in a sequence-to-sequence model in either one encoder or as a fully separate encoder, with improvements that are robust to different language models, languages and data sets. For English, these improvements are larger than adding individual sources of linguistic information or adding non-contextual embeddings. A new method o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2

Relationship

2
7

Authors

Journals

citations
Cited by 14 publications
(14 citation statements)
references
References 75 publications
3
11
0
Order By: Relevance
“…CharacterBERT (Boukkouri et al, 2020) ported this technique to BERT (Devlin et al, 2019), augmenting its existing WordPiece-tokenized input. Consistent with previous observations that feeding characters into a transformer stack comes with a huge computational cost while not improving over tokenization-based approaches (Al-Rfou et al, 2019), a BERT model fine-tuned for semantic parsing achieved gains only when characters complemented subwords (van Noord et al, 2020).…”
Section: Character-level Modelssupporting
confidence: 84%
“…CharacterBERT (Boukkouri et al, 2020) ported this technique to BERT (Devlin et al, 2019), augmenting its existing WordPiece-tokenized input. Consistent with previous observations that feeding characters into a transformer stack comes with a huge computational cost while not improving over tokenization-based approaches (Al-Rfou et al, 2019), a BERT model fine-tuned for semantic parsing achieved gains only when characters complemented subwords (van Noord et al, 2020).…”
Section: Character-level Modelssupporting
confidence: 84%
“…Neural Architecture We use a recurrent sequence-to-sequence neural network with two bi-directional LSTM layers (Hochreiter and Schmidhuber, 1997) (Vaswani et al, 2017), as implemented in the same framework. However, similar to van Noord et al (2020), none of our experiments reached the performance of the bi-LSTM model. We will therefore only show results of the bi-LSTM model in this paper.…”
Section: Input Representation Typesmentioning
confidence: 44%
“…Several data-driven methods based on neural networks have been proposed for DRS parsing (van Noord et al, 2018bLiu et al, 2019a;Evang, 2019;Fancellu et al, 2019;Fu et al, 2020;van Noord et al, 2020). These approaches frame semantic parsing as a sequence transformation problem and map the target meaning representation to string format.…”
Section: Introductionmentioning
confidence: 99%
“…We would like to thank Anouck Braggaar, Max Müller-Eberstein and Kristian Nørgaard Jensen for testing development versions. Furthermore, we thank Rik van Noord for his participation in the video, and providing an early use-case for MACHAMP (van Noord et al, 2020). This research was supported by an Amazon Research Award, an STSM in the Multi3Generation COST action (CA18231), a visit supported by COSBI, grant 9063-00077B (Danmarks Frie Forskningsfond), and Nvidia corporation for sponsoring Titan GPUs.…”
Section: Acknowledgmentsmentioning
confidence: 99%