2006
DOI: 10.1016/j.knosys.2005.11.014
|View full text |Cite
|
Sign up to set email alerts
|

The effect of principal component analysis on machine learning accuracy with high-dimensional spectral data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
47
0
2

Year Published

2009
2009
2022
2022

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 111 publications
(49 citation statements)
references
References 12 publications
0
47
0
2
Order By: Relevance
“…Both FE and FS have been found in many studies (Babaoğlu, Fındık, & Bayrak, 2010; Chandrashekar & Sahin, 2014; Howley, Madden, O'Connell, & Ryder, 2006) to improve classification accuracy, and it was therefore expected that predictions following these methods would be more accurate than predictions made on the full dataset. However, in the ecologically redundant Kwongan, these methods led to statistically significant classification improvements in only three scenarios, whereas in the nonredundant Woodland, they improved accuracies 12 times.…”
Section: Discussionmentioning
confidence: 99%
“…Both FE and FS have been found in many studies (Babaoğlu, Fındık, & Bayrak, 2010; Chandrashekar & Sahin, 2014; Howley, Madden, O'Connell, & Ryder, 2006) to improve classification accuracy, and it was therefore expected that predictions following these methods would be more accurate than predictions made on the full dataset. However, in the ecologically redundant Kwongan, these methods led to statistically significant classification improvements in only three scenarios, whereas in the nonredundant Woodland, they improved accuracies 12 times.…”
Section: Discussionmentioning
confidence: 99%
“…According to [58], for datasets with very low complexity (few PCs), the relevant information has been excluded during the process of PCA, which resulted in a lower classification accuracy for datasets after PCA. The PCA could give a higher classification accuracy to datasets with very high complexity (many PCs), where the dataset before PCA does not only have relevant information, but also contains noise [59]. With the presence of noise, the classifier over-fits the training data and thus does not generalize well.…”
Section: Supervised Machine Learning Analysis For Pattern Recognitionmentioning
confidence: 99%
“…Previous studies [116] have shown how the number of PCs used affects classification error among several different classification methods for high dimensional data such as…”
Section: Selection Of Pcs Based On Cumulative Percent Variancementioning
confidence: 99%