1997
DOI: 10.1117/12.263425
|View full text |Cite
|
Sign up to set email alerts
|

<title>Recognition and visual feature matching of text region in video for conceptual indexing</title>

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

1999
1999
2006
2006

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(7 citation statements)
references
References 0 publications
0
7
0
Order By: Relevance
“…Lienhart and Stuber (1996) assume that characters are drawn in high contrast against the background to be extracted and have no actual results for recognition. Kurakake et al (1997) present results for recognition using adaptive thresholding and color segmentation to extract characters. However, with news captions, we observe characters which have pixel values similar to those in the background.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Lienhart and Stuber (1996) assume that characters are drawn in high contrast against the background to be extracted and have no actual results for recognition. Kurakake et al (1997) present results for recognition using adaptive thresholding and color segmentation to extract characters. However, with news captions, we observe characters which have pixel values similar to those in the background.…”
Section: Introductionmentioning
confidence: 99%
“…Character recognition in videos to make indices is described by Lienhart and Stuber (1996) and Kurakake et al(1997). Lienhart and Stuber (1996) assume that characters are drawn in high contrast against the background to be extracted and have no actual results for recognition.…”
Section: Introductionmentioning
confidence: 99%
“…To construct such systems, both low-level features such as object shape, region intensity, color, texture, motion descriptors, audio measurements, and high-level techniques such as human face detection, speaker identification, and character recognition have been studied for indexing and retrieving image and video information in recent years [3], [4], [10], [11], [13], [19], [21], [24], [27]- [29], [32], [36]. Among these techniques, video caption based methods have attracted particular attention due to the rich content information contained in caption text [1], [2], [6], [9], [11]- [13], [15], [16], [19], [20], [27], [33], [36]. Caption text routinely provides such valuable indexing information as scene locations, speaker names, program introductions, sports scores, special announcements, dates and time.…”
Section: Introductionmentioning
confidence: 99%
“…Even though some address text detection in video frames [1], [5], [11]- [13], [16], [20], [34], they usually treat each video frame as an independent image. When temporal information are utilized, they are used only for text enhancement through multiframe averaging [18] or time-based minimum pixel search [15], [20], [27], [28]. These approaches require text detection and localization for every frame of a video, and careful caption blocks tracing and matching are needed between each frame pair for multiframe enhancement and removal of duplicate captions in different frames.…”
Section: Introductionmentioning
confidence: 99%
“…One problem is that when the caption is superimposed on a background image, it is difficult to apply existing OCR (optical character recognition) techniques, since the resolution of the characters is degraded as a result of the small number of scan lines (525 in NTSC broadcasts). Consequently, there have been several studies focusing on the recognition of caption characters [3,9], which have had some success.…”
Section: Analysis Of News Captions With Reference To Suffix Nounsmentioning
confidence: 99%