Proceedings of the 8th Workshop on Asian Translation (WAT2021) 2021
DOI: 10.18653/v1/2021.wat-1.16
|View full text |Cite
|
Sign up to set email alerts
|

NLPHut’s Participation at WAT2021

Abstract: This paper provides the description of shared tasks to the WAT 2021 by our team "NLPHut". We have participated in the English→Hindi Multimodal translation task, English→Malayalam Multimodal translation task, and Indic Multilingual translation task. We have used the state-of-the-art Transformer model with language tags in different settings for the translation task and proposed a novel "region-specific" caption generation approach using a combination of image CNN and LSTM for the Hindi and Malayalam image capti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 17 publications
0
4
0
Order By: Relevance
“…To generate the Hausa captions, we followed Parida et al (2021b) who proposed a region-specific image captioning method through the fusion of the encoded features of the region and the complete image. The model consists of three modules -an encoder, fusion, and decoder -as shown in Figure 3.…”
Section: Image Caption Generationmentioning
confidence: 99%
“…To generate the Hausa captions, we followed Parida et al (2021b) who proposed a region-specific image captioning method through the fusion of the encoded features of the region and the complete image. The model consists of three modules -an encoder, fusion, and decoder -as shown in Figure 3.…”
Section: Image Caption Generationmentioning
confidence: 99%
“…Thereafter, a similar method was adopted for English-Bengali [54]. Inspired from [23], a sequence-to-sequence model was used for MMT [55]. In this method, the input is the concatenated source and target side, and the output is object tags, which act as dummy image features on unimodal datasets.…”
Section: Multimodal Translation On Indian Languagesmentioning
confidence: 99%
“…However, BLEU decreased slightly by 0.06 points on the challenge subset for English-Hindi. Although [55] used similar methods and observed huge improvements over text-based models, their baseline was much weaker. Our text-based baseline (without fine-tuning on multimodal data) performs better than their fine-tuned model in some subsets in terms of BLEU score.…”
Section: Non-noisy Settingsmentioning
confidence: 99%
See 1 more Smart Citation