2022
DOI: 10.1007/s10032-022-00422-7
|View full text |Cite
|
Sign up to set email alerts
|

Refocus attention span networks for handwriting line recognition

Abstract: Recurrent neural networks have achieved outstanding recognition performance for handwriting identification despite the enormous variety observed across diverse handwriting structures, and poor-quality scanned documents. Initially, we proposed a BiLSTM baseline model with a sequential architecture well-suited for modeling text lines due to its capacity to learn probability distributions over character or word sequences. However, employing such recurrent paradigms prevents parallelization and suffers from vanish… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 48 publications
0
2
0
Order By: Relevance
“…Sampath and Gomathi [24] used MNIST to train neural networks for English handwritten recognition. RIMES, IAM, and READ were used in [25][26][27] to generate variable-length symbol sequences from the English handwritten text. More related work is shown in Table 1, which includes the datasets' name, size, and the procedure used.…”
Section: English Languagementioning
confidence: 99%
“…Sampath and Gomathi [24] used MNIST to train neural networks for English handwritten recognition. RIMES, IAM, and READ were used in [25][26][27] to generate variable-length symbol sequences from the English handwritten text. More related work is shown in Table 1, which includes the datasets' name, size, and the procedure used.…”
Section: English Languagementioning
confidence: 99%
“…The purpose of sequence modeling is to capture the contextual information of characters in image text for the next prediction, which is more suitable than handling each character separately. Bidirectional Long Short Term Memory (BiLSTM) can better capture long-term dependencies than traditional RNN structures in the sequence modeling stage, and many studies have begun to use BiLSTM in character recognition tasks 18) . However, although BiLSTM can efficiently capture contextual information, its structure itself determines that it is very time-consuming during both training and inference.…”
Section: Non-segmentation-based Character Recognitionmentioning
confidence: 99%