Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005.
DOI: 10.1109/icassp.2005.1415290
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Vocabulary Expansion for Automatic Transcription of Broadcast News

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 13 publications
0
8
0
Order By: Relevance
“…We compared the results yielded by our method with those yielded by three conventional OOV scoring methods: 1) frequency order (FQ), 2) concept base (CB) [14] and 3) latent dirichlet allocation (LDA) [20]. For CB, we constructed a concept base with 850K words of transcription from call centers that are different from centers of evaluation.…”
Section: Oov Selection Accuracymentioning
confidence: 99%
See 1 more Smart Citation
“…We compared the results yielded by our method with those yielded by three conventional OOV scoring methods: 1) frequency order (FQ), 2) concept base (CB) [14] and 3) latent dirichlet allocation (LDA) [20]. For CB, we constructed a concept base with 850K words of transcription from call centers that are different from centers of evaluation.…”
Section: Oov Selection Accuracymentioning
confidence: 99%
“…Although these factors can be calculated without any manual efforts, it does not focus on relevances between OOV words in relevant documents and target spoken documents. [14] or latent dirichlet allocation (LDA) [20] can evaluate relevances between OOV words in relevant documents and target spoken documents without any manual efforts. They calculate relevances on the basis of topic distributions of each OOV word and target spoken documents.…”
Section: Semantic Similaritymentioning
confidence: 99%
“…In some models, the approach is to expand the lexicon by adding new words or pronunciations. Ohtsuki et al (2005) propose a two-run model where in the first run, the input speech is recognized by the reference vocabulary and relevant words are extracted from the vocabulary database and added thereafter to the reference vocabulary to build an expanded lexicon. Word recognition is done in the second run based on the lexicon.…”
Section: Related Workmentioning
confidence: 99%
“…The experiments are carried out in the framework of ES-TER evaluation campaign and the Google search engine is used to access Web data 1 . All the tests are performed on about 6 hours of French broadcast news from the test corpus of ESTER 2005.…”
Section: Experimental Frameworkmentioning
confidence: 99%
“…Some papers report experiments on a posteriori search of new words in large external databases [1], but such static approaches fail in contemporary document transcription, where topics and named entities are frequently unexpected. Nervertheless, the web constitutes an immense and continuously updated source of language data, in which most of the possible word-sequences are stored.…”
Section: Introductionmentioning
confidence: 99%