2021
DOI: 10.1007/978-3-030-86334-0_5
|View full text |Cite
|
Sign up to set email alerts
|

SPAN: A Simple Predict & Align Network for Handwritten Paragraph Recognition

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 22 publications
(19 citation statements)
references
References 20 publications
0
19
0
Order By: Relevance
“…They refer to deep architectures composed of many convolutional layers and no recurrent layer. They often include the most recent innovations in deep learning, like gating mechanism or residual connections to obtain state of the art results [5,4,9,19]. These architectures benefit from the computational parallelism offered by convolutional layers, and hence can be trained faster than traditional architectures based on recurrent layers.…”
Section: Standard Approaches For Htrmentioning
confidence: 99%
“…They refer to deep architectures composed of many convolutional layers and no recurrent layer. They often include the most recent innovations in deep learning, like gating mechanism or residual connections to obtain state of the art results [5,4,9,19]. These architectures benefit from the computational parallelism offered by convolutional layers, and hence can be trained faster than traditional architectures based on recurrent layers.…”
Section: Standard Approaches For Htrmentioning
confidence: 99%
“…To alleviate this issue, architectures handling singlecolumn pages (or paragraphs) were recently proposed. In [2], [3], the aim is to reformulate the two-dimensional problem to a one-dimensional problem in order to use the CTC loss, which is only designed for one-dimensional sequence alignment problems. Indeed, this loss enables oneshot predictions, contrary to the cross entropy loss, which implies a recurrent process.…”
Section: Handwritten Text Recognitionmentioning
confidence: 99%
“…This way, prediction times are shorter. While the model from [2] learns a representation transformation, the model from [3] learns to align the predictions on the vertical axis, thanks to a reshaping operation. [1] and [4] proposed attention-based models.…”
Section: Handwritten Text Recognitionmentioning
confidence: 99%
See 1 more Smart Citation
“…1 Then, each line is transcribed independently with line-level HTR. Furthermore, recent works do not explicitly include any LA step and aim to obtain the transcription hypothesis by processing whole pages or paragraphs [4,13,50,40].…”
Section: Introductionmentioning
confidence: 99%