Proceedings of the 16th Meeting on the Mathematics of Language 2019
DOI: 10.18653/v1/w19-5709
|View full text |Cite
|
Sign up to set email alerts
|

Maximum Likelihood Estimation of Factored Regular Deterministic Stochastic Languages

Abstract: This paper proves that for every class C of stochastic languages defined with the coemission product of finitely many probabilistic, deterministic finite-state acceptors (PDFA) and for every data sequence D of finitely many strings drawn i.i.d. from some stochastic language, the Maximum Likelihood Estimate of D with respect to C can be found efficiently by locally optimizing the parameter values. We show that a consequence of the co-emission product is that each PDFA behaves like an independent factor in a joi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(6 citation statements)
references
References 14 publications
0
6
0
Order By: Relevance
“…The learning problem for any subregular languages is the same as SP language (Θ = arg min Θ M (nll(S|Θ M ))). Moreover, Shibata & Heinz (2019) has shown the convexity of this learning problem as long as the production of word likelihood of specific subregular language is defined with respect to coemission probability. One can easily modify each factored WDFA in SP phonotactic model to represent one tier, and model multitier interactions through coemission probability.…”
Section: Discussionmentioning
confidence: 99%
“…The learning problem for any subregular languages is the same as SP language (Θ = arg min Θ M (nll(S|Θ M ))). Moreover, Shibata & Heinz (2019) has shown the convexity of this learning problem as long as the production of word likelihood of specific subregular language is defined with respect to coemission probability. One can easily modify each factored WDFA in SP phonotactic model to represent one tier, and model multitier interactions through coemission probability.…”
Section: Discussionmentioning
confidence: 99%
“…In our framework, it is possible to induce an unrestricted automaton with a given number of states, or an automaton with hard-coded constraints representing various subregular languages. This work fills a gap in the formal linguistics literature, where learners have been developed within certain subregular classes (Shibata and Heinz, 2019;Heinz, 2010;Heinz and Rogers, 2010;Futrell et al, 2017), whereas our learner can in principle induce any (sub)regular language. In addition, we demonstrate how Strictly Local and Strictly Piecewise constraints can be encoded within our framework, and show how informationtheoretic regularization can be applied to produce deterministic automata.…”
Section: Introductionmentioning
confidence: 93%
“…Strictly Piecewise A Strictly k-Piecewise k-SP) language, each symbol depends on the presence of any preceding k − 1 symbols at arbitrary distance (Heinz, 2007(Heinz, , 2018Shibata and Heinz, 2019). For example, in a 2-SP language, in a string abc, c would be conditional on the presence of a and the presence of b, without regard to distance nor the relative order of a and b.…”
Section: Strictly Localmentioning
confidence: 99%
See 1 more Smart Citation
“…Strictly Piecewise A Strictly k-Piecewise k-SP) language, each symbol depends on the presence of any preceding k − 1 symbols at arbitrary distance Shibata and Heinz, 2019). For example, in a 2-SP language, in a string abc, c would be conditional on the presence of a and the presence of b, without regard to distance nor the relative order of a and b.…”
Section: Strictly Localmentioning
confidence: 99%