2022
DOI: 10.48550/arxiv.2205.15301
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Can Transformer be Too Compositional? Analysing Idiom Processing in Neural Machine Translation

Abstract: Unlike literal expressions, idioms' meanings do not directly follow from their parts, posing a challenge for neural machine translation (NMT). NMT models are often unable to translate idioms accurately and over-generate compositional, literal translations. In this work, we investigate whether the non-compositionality of idioms is reflected in the mechanics of the dominant NMT model, Transformer, by analysing the hidden states and attention patterns for models with English as source language and one of seven Eu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 30 publications
0
2
0
Order By: Relevance
“…But as shown in Table 3 we observe that BTG-Seq2Seq can translate proper nouns, time expression, and even idioms to an extent. The idiom case is particularly interesting as pure seq2seq systems are known to be too compositional when translating idioms (Dankers et al, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…But as shown in Table 3 we observe that BTG-Seq2Seq can translate proper nouns, time expression, and even idioms to an extent. The idiom case is particularly interesting as pure seq2seq systems are known to be too compositional when translating idioms (Dankers et al, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…Given that figurativeness is commonplace in everyday communication (Lakoff and Johnson, 2008), progress in the field of Natural Language Understanding (NLU) would be incomplete without figurativeness understanding. Consequently, figurative text has been studied in various downstream NLP tasks such as machine translation (Dankers et al, 2022), textual entailment (Agerri, 2008), (Chakrabarty et al, 2021), (Liu et al, 2022) and dialog models (Jhamtani et al, 2021), inter-alia. However, to the best of our knowledge, there has not been a systematic study of figurative language understanding capabilities of question answering models.…”
Section: -Oscar Wildementioning
confidence: 99%