2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412447
|View full text |Cite
|
Sign up to set email alerts
|

Multiple Document Datasets Pre-training Improves Text Line Detection With Deep Neural Networks

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 25 publications
(11 citation statements)
references
References 14 publications
0
11
0
Order By: Relevance
“…Their model provided second best results on the well-known cBAD dataset [26] while having a really light post-processing compared to the first best model. Lastly, in one of our previous works we presented Doc-UFCN [6], a FCN inspired by the previously cited systems and by the multi-modal model presented by Yang et al [27]. The encoder is composed of dilated convolutions while the decoder consists in transposed convolutions.…”
Section: Pixel-level Segmentationmentioning
confidence: 99%
See 4 more Smart Citations
“…Their model provided second best results on the well-known cBAD dataset [26] while having a really light post-processing compared to the first best model. Lastly, in one of our previous works we presented Doc-UFCN [6], a FCN inspired by the previously cited systems and by the multi-modal model presented by Yang et al [27]. The encoder is composed of dilated convolutions while the decoder consists in transposed convolutions.…”
Section: Pixel-level Segmentationmentioning
confidence: 99%
“…Their model follows an encoder-decoder architecture where the encoder is a ResNet-50 [28] pre-trained on natural scene images such as ImageNet [29]. Even if this method has shown good results on various tasks (page extraction, baseline detection or layout analysis), the inference time is still significant as shown in one of our early works [6]. Yang et al [27] also presented a model for multiple classes segmentation.…”
Section: Pixel-level Segmentationmentioning
confidence: 99%
See 3 more Smart Citations