2020
DOI: 10.1016/j.patcog.2020.107482
|View full text |Cite
|
Sign up to set email alerts
|

Accurate, data-efficient, unconstrained text recognition with convolutional neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
55
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 104 publications
(55 citation statements)
references
References 13 publications
0
55
0
Order By: Relevance
“…Our method also outperforms the other ones submitted to the competition on every metric (reported in [8]). While, the method proposed by [18] seems relevant when it is trained without many specific examples, our approach reaches state of the art performance on writer adaptation using 16 specific training pages. Table I also highlights the improvement of each model adaptation (optical and LM).…”
Section: Resultsmentioning
confidence: 88%
See 1 more Smart Citation
“…Our method also outperforms the other ones submitted to the competition on every metric (reported in [8]). While, the method proposed by [18] seems relevant when it is trained without many specific examples, our approach reaches state of the art performance on writer adaptation using 16 specific training pages. Table I also highlights the improvement of each model adaptation (optical and LM).…”
Section: Resultsmentioning
confidence: 88%
“…One of them was using Multi-dimensional LSTM (MDLSTM) [17], which has provided good performance when trained on a generic large data set, but has shown difficulty in carrying the writer adaptation process with few samples. After the competition, [18] proposed a fully convolutional network architecture trained with the CTC loss function for text recognition. This system achieves the best results on the READ 2018 data set, significantly outperforming the other systems.…”
Section: A Optical Model Adaptationmentioning
confidence: 99%
“…More recently, Yousef et al propose a FCN architecture [8] with an heavy use of normalization through Batch Normalization [21], Batch Renormalization [22] and Layer Normalization [23]. Their system also contains a gated mechanism derived from Highway Networks [24], residual components and Depthwise Separable Convolutions [25].…”
Section: B Recurrence-free Modelsmentioning
confidence: 99%
“…The gating mechanism defined in our proposed architecture is similar to the one presented in [8] but we only used a 2-part split (one accounting for the gate, the other for the features to be selected by the gate), whereas the authors have proposed a 3-part split with a gate composed of two filters bancs with substraction. The split is carried out over the channel axis.…”
Section: Gating Mechanismmentioning
confidence: 99%
See 1 more Smart Citation