Proceedings of Sixth International Conference on Document Analysis and Recognition
DOI: 10.1109/icdar.2001.953780
|View full text |Cite
|
Sign up to set email alerts
|

A maximum-likelihood approach to segmentation-based recognition of unconstrained handwriting text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(4 citation statements)
references
References 1 publication
0
4
0
Order By: Relevance
“…Linguistic based text segmentation approaches are derived from the lexical cohesion theory of Halliday and Hasan (1976). They rely on terms repetition to detect topic changes (Reynar 1998;Hearst 1994;Youmans 1991), n-gram word or phrases (Levow 2004) or word frequency (Senda and Yamada 2001;Reynar 1999;Beeferman et al 1997) as well as lexical chaining to identify topic changes (Stokes 2004;Manning 1998) and prosodic clues to mark shifts to new topics (Levow 2004). To improve the accuracy of a text segmentation algorithm Choi et al (2001) applied Latent Semantic Analysis (LSA), a technique aimed at extracting and representing the contextual-usage meaning of words by statistical computations applied to a large corpus (Landauer and Dumais 1997).…”
Section: Previous Workmentioning
confidence: 99%
“…Linguistic based text segmentation approaches are derived from the lexical cohesion theory of Halliday and Hasan (1976). They rely on terms repetition to detect topic changes (Reynar 1998;Hearst 1994;Youmans 1991), n-gram word or phrases (Levow 2004) or word frequency (Senda and Yamada 2001;Reynar 1999;Beeferman et al 1997) as well as lexical chaining to identify topic changes (Stokes 2004;Manning 1998) and prosodic clues to mark shifts to new topics (Levow 2004). To improve the accuracy of a text segmentation algorithm Choi et al (2001) applied Latent Semantic Analysis (LSA), a technique aimed at extracting and representing the contextual-usage meaning of words by statistical computations applied to a large corpus (Landauer and Dumais 1997).…”
Section: Previous Workmentioning
confidence: 99%
“…A review of the literature on text segmentation techniques reveals two distinct approaches: statistically based and linguistically driven methods [5,14]. Some statistical approaches are based on probability distributions [2], machine learning techniques ranging from neural networks [4], to support vector machines [18] and Bayesian networks [21], while others treat text as an unlabelled sequence of topics using a hidden Markov model [24]. [8] developed a text segmentation tool called C99 which uses a divisive clustering algorithm developed by [20] to identify topic boundaries.…”
Section: Previous Workmentioning
confidence: 99%
“…By incorporating the factor of character size in determining the likelihood, they showed better performance than without including it. Senda et al published a similar approach to the above method and formulated the problem as a search for the most probable interpretation of character segmentation, recognition and context, but they did not deal with the likelihood of character size [6], [7]. In off-line handwriting recognition, the same problem of character segmentation and recognition occurs.…”
Section: Introductionmentioning
confidence: 99%