2013
DOI: 10.1007/978-3-642-40991-2_41
|View full text |Cite
|
Sign up to set email alerts
|

Hub Co-occurrence Modeling for Robust High-Dimensional kNN Classification

Abstract: Abstract. The emergence of hubs in k-nearest neighbor (kNN) topologies of intrinsically high dimensional data has recently been shown to be quite detrimental to many standard machine learning tasks, including classification. Robust hubness-aware learning methods are required in order to overcome the impact of the highly uneven distribution of influence. In this paper, we have adapted the Hidden Naive Bayes (HNB) model to the problem of modeling neighbor occurrences and co-occurrences in high-dimensional data. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0
3

Year Published

2013
2013
2019
2019

Publication Types

Select...
4
3
1

Relationship

5
3

Authors

Journals

citations
Cited by 15 publications
(16 citation statements)
references
References 23 publications
0
13
0
3
Order By: Relevance
“…This is both a geometric consequence of their emergence, as well as a consequence of frequent and severe cluster assumption violation in many real-world data sets. Therefore, hubnessaware methods have been proposed for instance selection [10,66], metric learning [35,36,58,64,70] and classification [67,69,74,76]-in order to reduce the negative impact of detrimental hub points in supervised learning. Hubness has also been shown to affect the performance of learning methods under class imbalance in highdimensional data [68].…”
Section: Application Domains and Other Types Of Hubness-aware Methodsmentioning
confidence: 99%
“…This is both a geometric consequence of their emergence, as well as a consequence of frequent and severe cluster assumption violation in many real-world data sets. Therefore, hubnessaware methods have been proposed for instance selection [10,66], metric learning [35,36,58,64,70] and classification [67,69,74,76]-in order to reduce the negative impact of detrimental hub points in supervised learning. Hubness has also been shown to affect the performance of learning methods under class imbalance in highdimensional data [68].…”
Section: Application Domains and Other Types Of Hubness-aware Methodsmentioning
confidence: 99%
“…Points that lie closer to those local cluster centers tend to be closer on average to other points in the cluster, which gives them a higher probability of being included in k-nearest neighbor sets. This is why some hub linkage occurs, where pairs of hub points in proximity of the same cluster center often co-occur as neighbors [37].…”
Section: Data Hubness In Many Dimensionsmentioning
confidence: 99%
“…In classification, hubness-aware k-nearest neighbor voting frameworks have been shown to significantly improve kNN classification accuracy [40][39][36] [37]. Hubness information was also shown to be potentially useful in certain types of support-vector machines (SVM) [15].…”
Section: Introductionmentioning
confidence: 99%
“…As nearest neighbor classifiers are attractive both from the theoretical and practical points of view, considerable research was performed to enhance nearest neighbor classification. Some of the most promising recent methods were based on the observation that a few time-series tend to be nearest neighbors of surprisingly large amount of other time-series [62]. We refer to this phenomenon as the presence of hubs or hubness for short, and the classifiers that take this phenomenon into account are called hubness-aware classifiers.…”
Section: Introductionmentioning
confidence: 99%