2020 25th International Conference on Pattern Recognition (ICPR) 2021
DOI: 10.1109/icpr48806.2021.9412806
|View full text |Cite
|
Sign up to set email alerts
|

Gaussian Constrained Attention Network for Scene Text Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

4
5

Authors

Journals

citations
Cited by 20 publications
(5 citation statements)
references
References 51 publications
0
5
0
Order By: Relevance
“…The performance on five public benchmarks demonstrates the effectiveness and robustness of our approach. We'd like to combine text recognition [39,40], self-supervised learning [14,31,32,69,75,76], and knowledge distillation [65,66] to build a robust text reading system.…”
Section: Discussionmentioning
confidence: 99%
“…The performance on five public benchmarks demonstrates the effectiveness and robustness of our approach. We'd like to combine text recognition [39,40], self-supervised learning [14,31,32,69,75,76], and knowledge distillation [65,66] to build a robust text reading system.…”
Section: Discussionmentioning
confidence: 99%
“…Then, building polygon prediction module outputs the hidden state h t from the previous input. Furthermore, an attention module with Gaussian constraints [45] calculates the attention weight α t to integrate building feature B and hidden state h t , which can better focus on the local region of the previously predicted vertices y t−2 and y t−1 . Afterward, the region-related coefficient is calculated from attention weight α t and building feature B using an element-wise product.…”
Section: Bidirectional Building Polygon Prediction Modulementioning
confidence: 99%
“…[34] trains a deep BiLSTM encoder to extract broader contextual dependencies. To solve the problem of attention diffusion, [45] proposes a Gaussian constrained refinement module to refine the attention distribution. [86] adopts an external language model to introduce useful context information into the decoding.…”
Section: Autoregressive Text Recognitionmentioning
confidence: 99%