2020
DOI: 10.48550/arxiv.2005.11239
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Character-level Transformer-based Neural Machine Translation

Abstract: Neural machine translation (NMT) is nowadays commonly applied at the subword level, using byte-pair encoding. A promising alternative approach focuses on character-level translation, which simplifies processing pipelines in NMT considerably. This approach, however, must consider relatively longer sequences, rendering the training process prohibitively expensive. In this paper, we discuss a novel, Transformer-based approach, that we compare, both in speed and in quality to the Transformer at subword and charact… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 18 publications
0
3
0
Order By: Relevance
“…However, translating at character level may incur significant computational overhead. Therefore, later works on character-level NMT (Cherry et al, 2018;Banar et al, 2020) mainly focus on reducing computation cost of them. Cherry et al (2018) show that by employing source sequence compression techniques, the quality and efficiency of character-based models can be properly balanced.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…However, translating at character level may incur significant computational overhead. Therefore, later works on character-level NMT (Cherry et al, 2018;Banar et al, 2020) mainly focus on reducing computation cost of them. Cherry et al (2018) show that by employing source sequence compression techniques, the quality and efficiency of character-based models can be properly balanced.…”
Section: Related Workmentioning
confidence: 99%
“…Cherry et al (2018) show that by employing source sequence compression techniques, the quality and efficiency of character-based models can be properly balanced. Banar et al (2020) share the same idea as Cherry et al (2018) but build their models using Transformer architecture. Our work differs from theirs in that we aim to analyze the performance of existing models instead of exploring novel architectures.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation