2019 IEEE Winter Conference on Applications of Computer Vision (WACV) 2019
DOI: 10.1109/wacv.2019.00033
|View full text |Cite
|
Sign up to set email alerts
|

TextCaps: Handwritten Character Recognition With Very Small Datasets

Abstract: Many localized languages struggle to reap the benefits of recent advancements in character recognition systems due to the lack of substantial amount of labeled training data. This is due to the difficulty in generating large amounts of labeled data for such languages and inability of deep learning techniques to properly learn from small number of training samples. We solve this problem by introducing a technique of generating new training samples from the existing samples, with realistic augmentations which re… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
40
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 62 publications
(41 citation statements)
references
References 19 publications
1
40
0
Order By: Relevance
“…In comparison with the fully-connected layer decoder [19], this captures more spatial relationships while reconstructing the images. Further, we use binary cross entropy as the loss function for improved performance [12].…”
Section: Class Independent Decoder Networkmentioning
confidence: 99%
“…In comparison with the fully-connected layer decoder [19], this captures more spatial relationships while reconstructing the images. Further, we use binary cross entropy as the loss function for improved performance [12].…”
Section: Class Independent Decoder Networkmentioning
confidence: 99%
“…An interesting work using capsule layers (a concept recently introduced by Sabour et al [76]) is TextCaps [77], a CNN with 3 convolutional layers and two capsule layers, which obtain accuracies of 95.36% in EMNIST Letters and 99.79% in EMNIST Digits. Another interesting work was published by dos Santos et al [78], where they proposed the use of deep convolutional extreme learning machine, where gradient descent is not used to train the network, allowing a very fast learning stage (only 21 min using CPU, according to authors).…”
Section: State Of the Artmentioning
confidence: 99%
“…We also selected two derivatives of the standard capsule network: MS-CapsNet 12 and TextCaps. 13 Both of these derivatives were applied to the Fashion-MNIST and CIFAR-10 datasets, and the main improvements from both variants also lie in their feature extraction processes. In terms of accuracy, both of these models exceed that of the standard capsule network.…”
Section: Resultsmentioning
confidence: 99%
“…TextCaps 13 introduced a technique for generating new training samples from existing samples that simulates the actual changes in human handwriting input by adding random noise to the corresponding instantiation parameters. This strategy is useful in character recognition for localized languages that lack large amounts of labeled training data.…”
Section: Related Workmentioning
confidence: 99%