2022
DOI: 10.11591/ijece.v12i2.pp1990-2000
|View full text |Cite
|
Sign up to set email alerts
|

Automatic text summarization of konkani texts using pre-trained word embeddings and deep learning

Abstract: <span lang="EN-US">Automatic text summarization has gained immense popularity in research. Previously, several methods have been explored for obtaining effective text summarization outcomes. However, most of the work pertains to the most popular languages spoken in the world. Through this paper, we explore the area of extractive automatic text summarization using deep learning approach and apply it to Konkani language, which is a low-resource language as there are limited resources, such as data, tools, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 20 publications
0
2
0
Order By: Relevance
“…This is done in order to ensure that the split text corresponds to the same way in the corpus of the pre-trained model and also uses the same vocabulary at pre-training time. Tokenization in mBART50 is also based on SentenPiece [31]. SentecePiece can perform sub-word model training directly from raw sentences [32].…”
Section: Tokenizermentioning
confidence: 99%
“…This is done in order to ensure that the split text corresponds to the same way in the corpus of the pre-trained model and also uses the same vocabulary at pre-training time. Tokenization in mBART50 is also based on SentenPiece [31]. SentecePiece can perform sub-word model training directly from raw sentences [32].…”
Section: Tokenizermentioning
confidence: 99%
“…Liu et al [36] used pre trained Word2Vec model and improved them for cross-domain classification by extending the vector to include domain information. Recently D'Silva and Sharma [37] used FastText pre-trained word embeddings and neural networks to classify Konkani texts. Hu et al [38] used BERT to integrate mental features and short text vector to improve topic classification and false detection in short text.…”
Section: Word Embeddings-based Techniques and Deep Learning Modelsmentioning
confidence: 99%