2019
DOI: 10.1109/access.2019.2902270
|View full text |Cite
|
Sign up to set email alerts
|

Korean-Vietnamese Neural Machine Translation System With Korean Morphological Analysis and Word Sense Disambiguation

Abstract: Although deep neural networks have recently led to great achievements in machine translation (MT), various challenges are still encountered during the development of Korean-Vietnamese MT systems. Because Korean is a morphologically rich language and Vietnamese is an analytic language, neither have clear word boundaries. The high rate of homographs in Korean causes word ambiguities, which causes problems in neural MT (NMT). In addition, as a low-resource language pair, there is no freely available, adequate Kor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
21
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
7

Relationship

2
5

Authors

Journals

citations
Cited by 27 publications
(23 citation statements)
references
References 39 publications
0
21
0
Order By: Relevance
“…However, research papers about Korean-Vietnamese MT remains rare. In this paper, we extend our previous research [1] by applying NER and POS to the Vietnamese corpus as a pre-processing step to improve the performance of Korean-Vietnamese NMT systems.…”
Section: Introductionmentioning
confidence: 90%
See 2 more Smart Citations
“…However, research papers about Korean-Vietnamese MT remains rare. In this paper, we extend our previous research [1] by applying NER and POS to the Vietnamese corpus as a pre-processing step to improve the performance of Korean-Vietnamese NMT systems.…”
Section: Introductionmentioning
confidence: 90%
“…When conducting experiments, we used our Korean-Vietnamese corpus, which has more than 454K sentence pairs [1], to train the MT models. This corpus is the state-ofthe-art and greatest parallel corpus for Vietnamese-Korean translation systems.…”
Section: Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…The Moses MT used statistical models, and the N-gram of the language model was set to four [17,44]. OpenNMT systems utilized an encoder-decoder model with the following typical attributes: recurrent neural networks = 2 × 512, word-embedding dimension = 512, and input feed = 13 epochs [45,46]. We conducted those SMT and NMT systems for bidirectional translation (i.e., Korean-to-English and English-to-Korean).…”
Section: Experimentationmentioning
confidence: 99%
“…mined by the structural features of the sentence. In particular, to commercialize the results of WSD, it will be necessary to address most words and their senses in a wide range of domains; these will range from information retrieval [41], [45], [49] or machine translation [4], [14], [30], [37] to even second language education [7], [50], using various sources such as movies and books.…”
Section: Introductionmentioning
confidence: 99%