Proceedings of the Human Language Technology Conference of the NAACL, Companion Volume: Short Papers on XX - NAACL '06 2006
DOI: 10.3115/1614049.1614060
|View full text |Cite
|
Sign up to set email alerts
|

Spectral clustering for example based machine translation

Abstract: Prior work has shown that generalization of data in an Example Based Machine Translation (EBMT) system, reduces the amount of pre-translated text required to achieve a certain level of accuracy (Brown, 2000). Several word clustering algorithms have been suggested to perform these generalizations, such as k-Means clustering or Group Average Clustering. The hypothesis is that better contextual clustering can lead to better translation accuracy with limited training data. In this paper, we use a form of spectral … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2010
2010
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 3 publications
0
5
0
Order By: Relevance
“…MWE-aware strategies in EBMT and RBMT. EBMT (Gangadharaiah, Brown, and Carbonell 2006) or RBMT strategies (Anastasiou 2008;Forcada et al 2011;Monti et al 2011) dynamically apply rules to handle MWE translations. Some rules are identified from the syntactic tree alignments (Segura and Prince 2011) and integrated into an EBMT system to handle discontiguous MWEs.…”
Section: Identificationmentioning
confidence: 99%
“…MWE-aware strategies in EBMT and RBMT. EBMT (Gangadharaiah, Brown, and Carbonell 2006) or RBMT strategies (Anastasiou 2008;Forcada et al 2011;Monti et al 2011) dynamically apply rules to handle MWE translations. Some rules are identified from the syntactic tree alignments (Segura and Prince 2011) and integrated into an EBMT system to handle discontiguous MWEs.…”
Section: Identificationmentioning
confidence: 99%
“…While combining different system types has proven beneficial (e.g. Simard et al (2007), where the PB-SMT system Portage corrects the output of the Systran RBMT system), there was no significant improvement in overall translation score when merging these two EBMT systems, although including the MaTrEx system's chunks did improve the CMU clustering method (Gangadharaiah et al 2006).…”
Section: Integration With Other Modelsmentioning
confidence: 97%
“…Random sampling is often surprisingly powerful (Kendall and Smith, 1938;Knuth, 1991;Sennrich et al, 2016a). There is extensive research to beat random sampling by methods based on entropy (Koneru et al, 2022), coverage and uncertainty (Peris and Casacuberta, 2018;Zhao et al, 2020), clustering Gangadharaiah et al, 2009), consensus , syntactic parsing (Miura et al, 2016), density and diversity (Koneru et al, 2022;Ambati et al, 2011), and learning to learn active learning strategies (Liu et al, 2018).…”
Section: Active Learning In Machine Translationmentioning
confidence: 99%