Proceedings of the Tenth International Conference on Information and Knowledge Management 2001
DOI: 10.1145/502585.502605
|View full text |Cite
|
Sign up to set email alerts
|

Using LSI for text classification in the presence of background text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
41
0

Year Published

2002
2002
2011
2011

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 107 publications
(41 citation statements)
references
References 9 publications
0
41
0
Order By: Relevance
“…For data set 2, we compared the performance by varying the number of SVD dimensions k from 10,20,40,60,80,100,120,150,200,250,300,350,400,450, 500, 550 to 600. The neural network's input nodes number is equal to the dimensions of the document vectors.…”
Section: Experimental Results and Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…For data set 2, we compared the performance by varying the number of SVD dimensions k from 10,20,40,60,80,100,120,150,200,250,300,350,400,450, 500, 550 to 600. The neural network's input nodes number is equal to the dimensions of the document vectors.…”
Section: Experimental Results and Analysismentioning
confidence: 99%
“…LSA has been applied to text categorization in many previous works. Yang [19] used SVD for noise reduction to improve the computational efficiency in text categorization, Zelikovitz and Hirsh [20] performed LSA expanded term by document matrix in conjunction with background knowledge in text categorization, more recently, the supervised LSA [21] has been proposed to improve the performance in text categorization, and the idea of using latent semantic indexing and recurrent neural network for text classification has been mentioned by Mitra, et al [22].…”
Section: Introductionmentioning
confidence: 99%
“…Unlabeled examples have been used in conjunction with positive examples to improve classification (Li and Liu, 2003;Yu et al, 2003). Empirically it has been shown that combining labeled and unlabeled examples can improve accuracy on unseen test sets in certain situations (Blum and Mitchell, 1998;Mitchell, 1999;Joachims, 1999;Jaakkola et al, 1999;Zelikovitz and Hirsh, 2001;Bruce, 2001;Zhang and Oles, 2000).…”
Section: Related Workmentioning
confidence: 99%
“…Principal components analysis is a simple method and has been proved to be useful to reduce features (Lam and Lee 1999). Latent semantic indexing is an amazing approach both for feature selection and reduction (Zelikovitz andHirsh 2001, Song and). More recently, natural language processing techniques, e.g.…”
Section: Introductionmentioning
confidence: 99%