2011 International Conference on Document Analysis and Recognition 2011
DOI: 10.1109/icdar.2011.303
|View full text |Cite
|
Sign up to set email alerts
|

Progressive Alignment and Discriminative Error Correction for Multiple OCR Engines

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2011
2011
2020
2020

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 20 publications
(23 citation statements)
references
References 12 publications
0
23
0
Order By: Relevance
“…Klein et al (2002) show that combining complementary result from different OCR models leads to a better output. Lund et al (2011) demonstrate that the overall error rate decreases with the addition of different OCR models, regardless of the performance of each added model. Lund et al (2013a) use machine learning techniques to select the best word recognitions among different OCR outputs.…”
Section: Related Workmentioning
confidence: 86%
See 2 more Smart Citations
“…Klein et al (2002) show that combining complementary result from different OCR models leads to a better output. Lund et al (2011) demonstrate that the overall error rate decreases with the addition of different OCR models, regardless of the performance of each added model. Lund et al (2013a) use machine learning techniques to select the best word recognitions among different OCR outputs.…”
Section: Related Workmentioning
confidence: 86%
“…One direction of work ensembles outputs from multiple OCR engines for the same input and selects the best word recognition as the final output (Klein et al, 2002;Cecotti and Belayd, 2005;Lund and Ringger, 2009;Lund et al, 2011;Lund et al, 2013a;Lund et al, 2013b). Klein et al (2002) show that combining complementary result from different OCR models leads to a better output.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In conclusion this paper has explored the use of in-domain training; where in-domain data is not available, techniques for using out-of-domain training data such as those explored in Lund, Walker, and Ringger (2011) [9] can be used.…”
Section: Discussionmentioning
confidence: 98%
“…In order to further validate the utility of the synthetic datasets, we used the data to train an existing approach to multipleengine OCR error correction, as described by Lund et al 14 The method is motivated by the reasoning that OCR engines have different strengths and weaknesses. Thus, if one OCR engine outputs an incorrect hypothesis for a word token in the source image, another engine might output the correct hypothesis.…”
Section: Ocr Error Correction Taskmentioning
confidence: 99%