2021
DOI: 10.11591/eei.v10i2.2873
|View full text |Cite
|
Sign up to set email alerts
|

An in-depth exploration of Bangla blog post classification

Abstract: Bangla blog is increasing rapidly in the era of information, and consequently, the blog has a diverse layout and categorization. In such an aptitude, automated blog post classification is a comparatively more efficient solution in order to organize Bangla blog posts in a standard way so that users can easily find their required articles of interest. In this research, nine supervised learning models which are Support Vector Machine (SVM), multinomial naïve Bayes (MNB), multi-layer perceptron (MLP), k-nearest ne… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 18 publications
0
2
0
Order By: Relevance
“…In this method, for keyword matching we used a predefined list of vulgar words with automatically filtered-out words that had no probability of occurrence in vulgar context, as explained in Section 3.3. The method achieved accuracies of 0.2, 0.245, 0.3, 0.324, 0.363, 0.385, 0.427, 0.449, 0.467, and 0.475 within the top 10,20,30,40,50,60,70,80,90, and 100 extracted words, respectively (see Table 4 and Figure 9). Moreover, for the longer word lists, the method, despite filtering out on average only half of the actually non-vulgar words, achieved accuracies close to purely human-based filtering.…”
Section: Baseline 2: Keyword-matching Methods Based On Tf-idf Term Ex...mentioning
confidence: 97%
See 1 more Smart Citation
“…In this method, for keyword matching we used a predefined list of vulgar words with automatically filtered-out words that had no probability of occurrence in vulgar context, as explained in Section 3.3. The method achieved accuracies of 0.2, 0.245, 0.3, 0.324, 0.363, 0.385, 0.427, 0.449, 0.467, and 0.475 within the top 10,20,30,40,50,60,70,80,90, and 100 extracted words, respectively (see Table 4 and Figure 9). Moreover, for the longer word lists, the method, despite filtering out on average only half of the actually non-vulgar words, achieved accuracies close to purely human-based filtering.…”
Section: Baseline 2: Keyword-matching Methods Based On Tf-idf Term Ex...mentioning
confidence: 97%
“…MNB performs well when using features that represent word counts or term frequencies in the context of text classification. Despite its simplicity, MNB frequently performs surprisingly well in tasks like sentiment analysis and document categorization [80].…”
Section: Multinomial Naive Bayesmentioning
confidence: 99%
“…In Table 7, the sentences are equivalent to each other in words and meanings. The BLEU score of this particular example is measured at 0.67 using a 1 gram probability weight [24]. 1 gram BLEU score is 1 when at least one reference sentence equals the candidate sentence in tokens (Table 4 where output sentences are already meaningful candidates.)…”
Section: Bidirectional Translatormentioning
confidence: 99%
“…The data are labeled into three, i.e., abusive, not abusive and abusive but not offensive. Islam et al introduced a bangla blog article classification system in [17]. Abusive language detection of Twitter text content [18] is implemented utilizing the bidirectional recurrent neural network (BiRNN) method.…”
Section: Introductionmentioning
confidence: 99%