Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.100
|View full text |Cite
|
Sign up to set email alerts
|

Quality-Aware Decoding for Neural Machine Translation

Abstract: Despite the progress in machine translation quality estimation and evaluation in the last years, decoding in neural machine translation (NMT) is mostly oblivious to this and centers around finding the most probable translation according to the model (MAP decoding), approximated with beam search. In this paper, we bring together these two lines of research and propose quality-aware decoding for NMT, by leveraging recent breakthroughs in reference-free and reference-based MT evaluation through various inference … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(35 citation statements)
references
References 36 publications
0
34
1
Order By: Relevance
“…When using LLM-SCQ, the performance of MAPS is on par with 5-Shot (MAPS LLM-SCQ ≈ 5-Shot); when using COMET-QE, MAPS consistently outperforms 5-Shot (MAPS COMET-QE > 5-Shot). More importantly, MAPS shows higher upper bounds for selection than Rerank (MAPS COMET > Rerank COMET ), implying that superior knowledge selection methods like a better QE model (Rei et al, 2022b), AutoMQM (Fernandes et al, 2023) or ranking strategy (Fernandes et al, 2022) can further improve MAPS.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…When using LLM-SCQ, the performance of MAPS is on par with 5-Shot (MAPS LLM-SCQ ≈ 5-Shot); when using COMET-QE, MAPS consistently outperforms 5-Shot (MAPS COMET-QE > 5-Shot). More importantly, MAPS shows higher upper bounds for selection than Rerank (MAPS COMET > Rerank COMET ), implying that superior knowledge selection methods like a better QE model (Rei et al, 2022b), AutoMQM (Fernandes et al, 2023) or ranking strategy (Fernandes et al, 2022) can further improve MAPS.…”
Section: Resultsmentioning
confidence: 99%
“…The best candidate is selected through QE. It can be considered as a pure reranking method without any guidance from extracted knowledge (Fernandes et al, 2022).…”
Section: Methodsmentioning
confidence: 99%
“…Results Interpretation: The Table enlists quality metrics of each algorithm over the given datasets. The Decision Tree-based Explanation is the most accurate method among all the methods evaluated, and outperforms all other methods in all datasets [12]. This seems to be a case where model tree-based approach for decision trees has more accurate understanding of the model predictions than model agnostic approach like LIME and SHAP.…”
Section: Evaluation Metricsmentioning
confidence: 86%
“…The writers applied Convolutional Neural Networks (CNNs) and Gradient Boosting Machines (GBMs) for image classification, finance, and healthcare processes respectively. • Explanation Algorithms: LIME, SHAP, Anchor, and Decision tree-based explanation is the method we utilized to make explanations for each of the black-box models predictions [12]. The interpretation of each explanation algorithm was determined by its capacity to give interpretative insights into predictions.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…Our proposal RL4F is orthogonal to RLHF; in fact we use an RLHF fine-tuned checkpoint in our experiments. For further discussion, please refer to Fernandes et al (2023) who catalogue different approaches on integrating natural language feedback to textual generations.…”
Section: How Is Feedback Used?mentioning
confidence: 99%