2023
DOI: 10.32604/cmc.2023.035814
|View full text |Cite
|
Sign up to set email alerts
|

Neural Machine Translation Models with Attention-Based Dropout Layer

Abstract: In bilingual translation, attention-based Neural Machine Translation (NMT) models are used to achieve synchrony between input and output sequences and the notion of alignment. NMT model has obtained state-of-the-art performance for several language pairs. However, there has been little work exploring useful architectures for Urdu-to-English machine translation. We conducted extensive Urdu-to-English translation experiments using Long short-term memory (LSTM)/Bidirectional recurrent neural networks (Bi-RNN)/Sta… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(1 citation statement)
references
References 34 publications
0
1
0
Order By: Relevance
“…After that, we have used a pre-trained model by PyTorch [17] which had 135,310,918 total parameters out of which 1,050,374 are trainable while 134,260,544 are non-trainable parameters. The batch size was set to 128, Learning rate to 0.005, dropout rate to 0.4 [18] and the output size is 3. The summary of our VGG-16 model can be seen in the figure (Figure 7) and the architecture diagram of the VGG-16 model is displayed in (Figure 8).…”
Section: Vgg-16mentioning
confidence: 99%
“…After that, we have used a pre-trained model by PyTorch [17] which had 135,310,918 total parameters out of which 1,050,374 are trainable while 134,260,544 are non-trainable parameters. The batch size was set to 128, Learning rate to 0.005, dropout rate to 0.4 [18] and the output size is 3. The summary of our VGG-16 model can be seen in the figure (Figure 7) and the architecture diagram of the VGG-16 model is displayed in (Figure 8).…”
Section: Vgg-16mentioning
confidence: 99%