Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2014
DOI: 10.3115/v1/d14-1183
|View full text |Cite
|
Sign up to set email alerts
|

Large-scale Reordering Model for Statistical Machine Translation using Dual Multinomial Logistic Regression

Abstract: Phrase reordering is a challenge for statistical machine translation systems. Posing phrase movements as a prediction problem using contextual features modeled by maximum entropy-based classifier is superior to the commonly used lexicalized reordering model. However, Training this discriminative model using large-scale parallel corpus might be computationally expensive. In this paper, we explore recent advancements in solving large-scale classification problems. Using the dual problem to multinomial logistic r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2018
2018
2018
2018

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 33 publications
0
2
0
Order By: Relevance
“…We numerically showed the effect of three parameters; model size, number of Fourier components and learning rate choices on the stability. Table (1) shows the description of four binary classification datasets used for the analysis. These datasets can be downloaded from UCI machine learning repository website.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…We numerically showed the effect of three parameters; model size, number of Fourier components and learning rate choices on the stability. Table (1) shows the description of four binary classification datasets used for the analysis. These datasets can be downloaded from UCI machine learning repository website.…”
Section: Resultsmentioning
confidence: 99%
“…SGM computes the estimates of the gradient on the basis of a single randomly chosen sample in each iteration. Therefore, applying a stochastic gradient method for large scale machine learning problems can be computationally efficient [22,1,3,11]. In the context of supervised learning, models that are trained by such iterative optimization algorithms are commonly controlled by convergence rate analysis.…”
Section: Introductionmentioning
confidence: 99%