Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2015
DOI: 10.3115/v1/n15-1034
|View full text |Cite
|
Sign up to set email alerts
|

Sign constraints on feature weights improve a joint model of word segmentation and phonology

Abstract: This paper describes a joint model of word segmentation and phonological alternations, which takes unsegmented utterances as input and infers word segmentations and underlying phonological representations. The model is a Maximum Entropy or log-linear model, which can express a probabilistic version of Optimality Theory (OT; Prince and Smolensky (2004)), a standard phonological framework. The features in our model are inspired by OT's Markedness and Faithfulness constraints. Following the OT principle that such… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
6
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 16 publications
0
6
0
Order By: Relevance
“…Like PAIM, but unlike humans, generalization to CONF-UNATT items diminished after multiple exposure sets (in particular for σ = 0.5). A straightforward implementation of MaxEnt is therefore incapable of simulating the human results; better results could potentially be achieved with a regularization method that encouraged sparsity (Goodman, 2004;Johnson et al, 2015).…”
Section: Comparison To Related Modelsmentioning
confidence: 99%
“…Like PAIM, but unlike humans, generalization to CONF-UNATT items diminished after multiple exposure sets (in particular for σ = 0.5). A straightforward implementation of MaxEnt is therefore incapable of simulating the human results; better results could potentially be achieved with a regularization method that encouraged sparsity (Goodman, 2004;Johnson et al, 2015).…”
Section: Comparison To Related Modelsmentioning
confidence: 99%
“…In other words, every underlying form must surface faithfully at least once in order to be considered a possible UR. This assumption is shared by other models of UR acquisition, such as Albright (2002), and of segmentation and UR acquisition (Johnson et al, 2015).…”
Section: Urc Inductionmentioning
confidence: 90%
“…The fact that segmentation is a prerequisite to build the lexicon only precludes lexical information from being used in segmentation if the two processes take place in serial, with learners developing the ability to segment speech before storing any lexical information. Previous models of segmentation either ignore the acquisition of the lexicon (Saffran et al, 1996a;Saffran et al, 1996b;Perruchet and Vinter, 1998) or do not fully utilize the richness of lexical knowledge (Johnson et al, 2015;Goldwater et al, 2009). This paper presents a model of segmentation in which the lexicon, represented by phonological underlying forms which correspond to meanings, is being acquired in parallel with segmentation, and the two processes are mutually informing.…”
Section: Introductionmentioning
confidence: 99%
“…We emphasize that we do not use any additional information besides our algorithms. Hayes (1996); Johnson et al (2015) applied explicit phonological rules or constraints to tasks such as word segmentation. In neural networks, we can implicitly learn from phonetic data and leave the networks to discover hidden phonetic features through end-to-end training opt specific NLP tasks, instead of applying hand-coded constraints.…”
Section: Related Workmentioning
confidence: 99%