2021
DOI: 10.1109/access.2021.3133200
|View full text |Cite
|
Sign up to set email alerts
|

Self-Supervised Representation Learning for Document Image Classification

Abstract: Supervised learning, despite being extremely effective, relies on expensive, time-consuming, and error-prone annotations. Self-supervised learning has recently emerged as a strong alternate to supervised learning in a range of different domains as collecting a large amount of unlabeled data can be achieved by simply crawling the internet. These self-supervised methods automatically discover features relevant to represent an input example by using self-defined proxy tasks. In this paper, we question the potenti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 33 publications
0
6
0
Order By: Relevance
“…While supervised learning proves highly effective in various applications, it is not without drawbacks. One significant limitation is its dependency on expensive, time-consuming, and error-prone annotations for training data [30]. This reliance on labeled data can pose a bottleneck, particularly in scenarios where acquiring large-scale labeled datasets is challenging, leading to increased costs and delays in model development.…”
Section: Supervised Learningmentioning
confidence: 99%
“…While supervised learning proves highly effective in various applications, it is not without drawbacks. One significant limitation is its dependency on expensive, time-consuming, and error-prone annotations for training data [30]. This reliance on labeled data can pose a bottleneck, particularly in scenarios where acquiring large-scale labeled datasets is challenging, leading to increased costs and delays in model development.…”
Section: Supervised Learningmentioning
confidence: 99%
“…There are several advantages of our proposed VisFormers model relative to the stateof-the-art approaches. The work conducted by [45,46] achieves an accuracy of less than 90% and requires over 900-1200 s for training using primary image classification neural networks. Nevertheless, our proposed model is designed by combining the transfer learning image classification model and Transformers and obtained an accuracy of over 96%.…”
Section: Comparison With the State Of The Artmentioning
confidence: 99%
“…The training algorithm is inspired by the recently proposed forward-forward algorithm ( 31 ) and local training proposals ( 38 – 41 ) in digital neural networks, which has been extended and adapted to the supervised and unsupervised model-free physical learning of PNNs. Each nonlinear physical system performs a nonlinear transformation on input data (Fig.…”
Section: Phyllmentioning
confidence: 99%
“…Local learning has been extensively studied for training digital neural networks, from early work on Hebbian contrastive learning in Hopfield models ( 35 ) to recent biologically plausible frameworks ( 31 , 34 , 36 , 37 ), blockwise BP ( 38 , 39 ), and contrastive representation learning ( 40 , 41 ). Inspired by this concept and to address the limitations of BP-based PNN training, we proposed a simple and physics-compatible PNN architecture augmented by a physical local learning (PhyLL) algorithm.…”
mentioning
confidence: 99%