2023
DOI: 10.46338/ijetae0223_07
|View full text |Cite
|
Sign up to set email alerts
|

OCR-based Hybrid Image Text Summarizer using Luhn Algorithm with FinetuneTransformer Modelsfor Long Document

Abstract: The accessibility of an enormous number of image text documents on the internet has expanded the opportunities to develop a system for image text recognition with text summarization. Several approaches used in ATS in the literature are based on extractive and abstractive techniques; however, few implementations of the hybrid approach were observed. This paper employed state-of-the-art transformer models with the Luhn algorithm for extracted texts using Tesseract OCR. Nine models were generated and tested using… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 31 publications
0
2
0
Order By: Relevance
“…For many diverse styles of programs in different fields, Optical character recognition is the bottom of it, which we use in our daily life. These days, Optical Character Recognition is being used in many different areas of research [12] [13]. P. Divya et al (2021) developed a web-based optical character recognition application using flask and tesseract [14].…”
Section: B Optical Character Recognitionmentioning
confidence: 99%
“…For many diverse styles of programs in different fields, Optical character recognition is the bottom of it, which we use in our daily life. These days, Optical Character Recognition is being used in many different areas of research [12] [13]. P. Divya et al (2021) developed a web-based optical character recognition application using flask and tesseract [14].…”
Section: B Optical Character Recognitionmentioning
confidence: 99%
“…The transformer model revolution in Natural Language Processing [16] began with the paper "Attention Is All You Need" [17]. Before this, the state-of-the-art approach in NLP relied on gated recurrent neural networks such as Long Short-term Memory (LSTM) [18][19] [20]and gated recurrent units (GRUs) with added.…”
Section: Artificialmentioning
confidence: 99%