2015
DOI: 10.1016/j.eswa.2014.09.031
|View full text |Cite
|
Sign up to set email alerts
|

Learning to classify short text from scientific documents using topic models with various types of knowledge

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0
5

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 66 publications
(29 citation statements)
references
References 21 publications
0
24
0
5
Order By: Relevance
“…One of the most common techniques for text transformation is enrichment, Phan et al [10] propose to use Latent Dirichelet Allocation (LDA) to generate several clusters of words (topics) from a set of external documents, and then to add the words in these topics to the texts in the training dataset based on the semantic links between the words in the texts and the words in the topics. In a similar way, Vo and Ock [19], multiply the sources of external documents to better enrich scientific document titles. In a previous work [3], we proposed a twolevel enrichment method which, on the one hand, adds to short texts words from topics generated with LDA and, on the other hand, adds words from topics according to the similarities between the topics and the texts as whole entities.…”
Section: Training Data Preprocessingmentioning
confidence: 97%
“…One of the most common techniques for text transformation is enrichment, Phan et al [10] propose to use Latent Dirichelet Allocation (LDA) to generate several clusters of words (topics) from a set of external documents, and then to add the words in these topics to the texts in the training dataset based on the semantic links between the words in the texts and the words in the topics. In a similar way, Vo and Ock [19], multiply the sources of external documents to better enrich scientific document titles. In a previous work [3], we proposed a twolevel enrichment method which, on the one hand, adds to short texts words from topics generated with LDA and, on the other hand, adds words from topics according to the similarities between the topics and the texts as whole entities.…”
Section: Training Data Preprocessingmentioning
confidence: 97%
“…Text categorisation [19], vocabulary reduction [20], visual encoding [21], image recognition [22] or even video retrieval [23] are some of the applications where topic models have been successfully used.…”
Section: Current Limitations and Trendsmentioning
confidence: 99%
“…Zhang and Zhong collected large-scale external data to build the topic model according to the LDA, which enables word topics to enrich feature representations of short text [22]. Vo and Ock employed a LDA-based method to discover hidden topic from universal datasets (e.g., Computer Science Bibliography), Lecture Notes in Computer Science book series (LNCS), and Wikipedia [23].…”
Section: Related Workmentioning
confidence: 99%