Proceedings of the Ninth Conference on Computational Natural Language Learning - CONLL '05 2005
DOI: 10.3115/1706543.1706564
|View full text |Cite
|
Sign up to set email alerts
|

Representational bias in unsupervised learning of syllable structure

Abstract: Unsupervised learning algorithms based on Expectation Maximization (EM) are often straightforward to implement and provably converge on a local likelihood maximum. However, these algorithms often do not perform well in practice. Common wisdom holds that they yield poor results because they are overly sensitive to initial parameter values and easily get stuck in local (but not global) maxima. We present a series of experiments indicating that for the task of learning syllable structure, the initial parameter we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2008
2008
2020
2020

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 13 publications
0
8
0
Order By: Relevance
“…Applying a standard smoothing algorithm and fourth-order HMM, Demberg (2006) scores 98.47% word accuracy. A fifth-order joint N -gram model of Schmid et al (2007) Table 4: Word accuracy on the datasets of Goldwater and Johnson (2005).…”
Section: Other Languagesmentioning
confidence: 99%
See 2 more Smart Citations
“…Applying a standard smoothing algorithm and fourth-order HMM, Demberg (2006) scores 98.47% word accuracy. A fifth-order joint N -gram model of Schmid et al (2007) Table 4: Word accuracy on the datasets of Goldwater and Johnson (2005).…”
Section: Other Languagesmentioning
confidence: 99%
“…The results are somewhat lower than in (Müller, 2001), but the approach can be more easily ported across languages. Goldwater and Johnson (2005) also explore using EM to learn the structure of English and German phonemes in an unsupervised setting, following Müller in modeling syllable structure with PCFGs. They initialize their parameters using a deterministic parser implementing the sonority principle and estimate the parameters for their maximum likelihood approach using EM.…”
Section: Previous Computational Approachesmentioning
confidence: 99%
See 1 more Smart Citation
“…3. First, the ARPABET transcription of a phone structure {A, w} for each word is explicitly parsed into syllables by means of a syllabification algorithm based on the English phonotactics principle of the maximization of syllable onsets (Goldwater and Johnson, 2005). Then, from the syllabified ARPABET transcription, the linguistic gestural model generates a synthesized gestural score, fĜ;Bg, specifying the CL and CD targets and stiffnesses for each TV, and the times during which those targets are in force.…”
Section: Architecture: Gestural Annotationmentioning
confidence: 99%
“…For training data we use the BN-VL task of N-Best phonetically segmented with HMM models. For segmenting the vocabulary we use an unsupervised syllabification (Goldwater and Johnson 2005) method we are publishing in a separate paper. The method syllabifies Dutch word lists with a 90% word accuracy with the bulk of errors being in one-phoneme shifts in syllable boundaries of the rarer long words.…”
Section: Training Proceduresmentioning
confidence: 99%