2021
DOI: 10.1007/s40009-021-01043-0
|View full text |Cite
|
Sign up to set email alerts
|

Feature Selection for Text Classification Using Machine Learning Approaches

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(12 citation statements)
references
References 12 publications
1
11
0
Order By: Relevance
“…Combining the paired-input nonlinear knockoff filter with the MLP in [27]. Random forest in [28], Naive Bayes, and SVM classifiers [29] are other algorithms used in recent studies of FS problems.…”
Section: Literature Reviewmentioning
confidence: 99%
“…Combining the paired-input nonlinear knockoff filter with the MLP in [27]. Random forest in [28], Naive Bayes, and SVM classifiers [29] are other algorithms used in recent studies of FS problems.…”
Section: Literature Reviewmentioning
confidence: 99%
“…As early as the 1970s, Salton et al [13] proposed vector space model (VSM), which was successfully applied to the famous SMART system. In the following 50 years, text classification has been mainly based on shallow learning model, for example, naive Bayes-based text classification method, K-nearest neighbor method, and support vector machine method [14][15][16]. Although these methods have improved accuracy, they all rely on complex feature processing engineering and do not take into account the semantic information of the text.…”
Section: Related Workmentioning
confidence: 99%
“…As Camastra and Vinciarelli mention [ 35 ], using more features than is strictly necessary leads to several problems, pointing out that one of the main problems was the space needed to store the data. As the amount of available information increases, the compression for storage becomes even more critical [ 12 , 36 , 37 ]. Additionally, for the scope of this work, it cannot be ignored that the application of dimensional reduction techniques for reducing pre-computed embedding dimensions neither improves the runtime nor the memory requirement for running the models.…”
Section: Related Workmentioning
confidence: 99%
“…Similarly, the use of dimension reduction techniques is likewise interesting in Semantic Similarity [ 27 , 37 ]. As discussed previously, in the Semantic Similarity task, the linear O ( N ) complexity of cosine similarity is one of the reasons why this distance metric is widely used in the community and this study.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation