2022
DOI: 10.3390/info13070309
|View full text |Cite
|
Sign up to set email alerts
|

Human Evaluation of English–Irish Transformer-Based NMT

Abstract: In this study, a human evaluation is carried out on how hyperparameter settings impact the quality of Transformer-based Neural Machine Translation (NMT) for the low-resourced English–Irish pair. SentencePiece models using both Byte Pair Encoding (BPE) and unigram approaches were appraised. Variations in model architectures included modifying the number of layers, evaluating the optimal number of heads for attention and testing various regularisation techniques. The greatest performance improvement was recorded… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 11 publications
(10 citation statements)
references
References 48 publications
1
9
0
Order By: Relevance
“…Vision transformers are often pre-trained on extensive datasets like ImageNet, and then subsequently fine-tuned for specialized tasks, proving efficient in areas with limited data. [ 38 ] For large language models, such as GPT and BERT, extensive pre-training on vast text corpora and low-resource languages enables deep language understanding [ 39 ] and adaptability to specialized domains in materials science. [ 40 , 41 ] Notably, large language models and multimodal vision-language models (such as GPT-4 V, LLaVA) exhibit advanced capabilities in two novel approaches, namely zero-shot and few-shot learning.…”
Section: What Is Machine Learning?mentioning
confidence: 99%
“…Vision transformers are often pre-trained on extensive datasets like ImageNet, and then subsequently fine-tuned for specialized tasks, proving efficient in areas with limited data. [ 38 ] For large language models, such as GPT and BERT, extensive pre-training on vast text corpora and low-resource languages enables deep language understanding [ 39 ] and adaptability to specialized domains in materials science. [ 40 , 41 ] Notably, large language models and multimodal vision-language models (such as GPT-4 V, LLaVA) exhibit advanced capabilities in two novel approaches, namely zero-shot and few-shot learning.…”
Section: What Is Machine Learning?mentioning
confidence: 99%
“…Breakthrough performance improvements in the area of MT have been achieved through research efforts focusing on NMT (Bahdanau et al, 2014) but the advent of the Transformer architecture has greatly improved MT performance. Consequently, SOTA performance has been attained on multiple language pairs (Bojar et al, 2017(Bojar et al, , 2018Lankford et al, 2021bLankford et al, , 2022b.…”
Section: Nmtmentioning
confidence: 99%
“…The remaining seven pairs are EN-GA, DE-GA, FR-PL, FR-GA, IT-PL, IT-GA and PL-GA; note that five out of these seven pairs involve the Irish language (GA). As of now, we prioritise corpus combinations for the other two language pairs (FR-PL and IT-PL); given our local expertise and in-house crawled datasets [35,36], all MT models involving Irish already produce high BLEU scores, and so there is less room for improvement. However, we plan to perform corpus combination for the Irish language once all the other languages are covered.…”
Section: Corpus Combinationmentioning
confidence: 99%
“…It is interesting to note that all models containing Irish as either a source or target language outperform Google Translate. While this may be unexpected in general, it was not a surprise to us; as mentioned in Section 5.3.3, we have access to many good-quality datasets for the Irish language (e.g., in the areas of health [35] and the legal domain [36]) from previous projects, and this considerably improves translation performance compared to the other language pairs, despite not being in the specific areas covered by EUComMeet. These datasets are not in the public domain, so are unavailable to Google Translate as additional training data to improve their engines.…”
Section: Baseline Vs Domain-adapted Mt System Performancementioning
confidence: 99%