2017
DOI: 10.1371/journal.pone.0174341
|View full text |Cite
|
Sign up to set email alerts
|

Relevance popularity: A term event model based feature selection scheme for text classification

Abstract: Feature selection is a practical approach for improving the performance of text classification methods by optimizing the feature subsets input to classifiers. In traditional feature selection methods such as information gain and chi-square, the number of documents that contain a particular term (i.e. the document frequency) is often used. However, the frequency of a given term appearing in each document has not been fully investigated, even though it is a promising feature to produce accurate classifications. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(1 citation statement)
references
References 33 publications
0
1
0
Order By: Relevance
“…The fifth part is to illustrate the effectiveness of MFS-MIRF in terms of the N, HL, and RL indices for the selected features under the MLKNN classifier. Seven state-of-the-art feature selection methods for comparison are MFS [51], MD [52], RP [53], VGFSS [54], NDM [55], AIPSO [56], and EGA+CDM [57], where EGA describes the enhanced genetic algorithm. Following the designed technologies in [51], the nine datasets are selected from Table 3 for comparison with the above eight algorithms.…”
Section: Classification Results Of Various Feature Selection Methomentioning
confidence: 99%
“…The fifth part is to illustrate the effectiveness of MFS-MIRF in terms of the N, HL, and RL indices for the selected features under the MLKNN classifier. Seven state-of-the-art feature selection methods for comparison are MFS [51], MD [52], RP [53], VGFSS [54], NDM [55], AIPSO [56], and EGA+CDM [57], where EGA describes the enhanced genetic algorithm. Following the designed technologies in [51], the nine datasets are selected from Table 3 for comparison with the above eight algorithms.…”
Section: Classification Results Of Various Feature Selection Methomentioning
confidence: 99%