2009
DOI: 10.1080/18756891.2009.9727671
|View full text |Cite
|
Sign up to set email alerts
|

Text Categorization Based on Topic Model

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 23 publications
(5 citation statements)
references
References 5 publications
0
5
0
Order By: Relevance
“…In order to operationalize the evaluation of selection bias, we use topic models to capture latent semantics. Regularly used topic modeling techniques such as Latent Dirichlet Allocation (LDA) (Blei et al, 2003) have proven their efficiency to handle several NLP applications such as data exploration (Rodriguez and Storer, 2020), Twitter hashtag recommendation (Godin et al, 2013), authorship attribution (Seroussi et al, 2014), and text categorization (Zhou et al, 2009). In order to evaluate the consistency of the generated topics, Newman et al (2010) Lau and Baldwin (2016) investigated the effect of cardinality on topic generation.…”
Section: Related Workmentioning
confidence: 99%
“…In order to operationalize the evaluation of selection bias, we use topic models to capture latent semantics. Regularly used topic modeling techniques such as Latent Dirichlet Allocation (LDA) (Blei et al, 2003) have proven their efficiency to handle several NLP applications such as data exploration (Rodriguez and Storer, 2020), Twitter hashtag recommendation (Godin et al, 2013), authorship attribution (Seroussi et al, 2014), and text categorization (Zhou et al, 2009). In order to evaluate the consistency of the generated topics, Newman et al (2010) Lau and Baldwin (2016) investigated the effect of cardinality on topic generation.…”
Section: Related Workmentioning
confidence: 99%
“…Apart from the mentioned studies, a recent direction suggests using the topic model notion as a feature reduction method, in which the text is represented as a mixture of hidden topics, where the extracted latent topics from text documents form the features ( Onan et al, 2016 ). In other words, the textual data is represented as a bag of topics ( Zhou et al, 2009 ; Yousef et al, 2020a ) rather than a bag of words. However, in the short-text corpus, an advanced approach must be developed ( Al Qundus et al, 2020 ).…”
Section: Introductionmentioning
confidence: 99%
“…Topic models, such as PLSA (Hofmann, 1999) and LDA (Blei et al, 2003), have shown great success in discovering latent topics in text collections. They have considerable applications in the information retrieval, text clustering and categorization (Zhou et al, 2009), word sense disambiguation (Boyd-Graber et al, 2007), etc.…”
Section: Introductionmentioning
confidence: 99%