2011
DOI: 10.1016/j.patcog.2010.11.019
|View full text |Cite
|
Sign up to set email alerts
|

Semi-supervised classification and betweenness computation on large, sparse, directed graphs

Abstract: a b s t r a c tThis work addresses graph-based semi-supervised classification and betweenness computation in large, sparse, networks (several millions of nodes). The objective of semi-supervised classification is to assign a label to unlabeled nodes using the whole topology of the graph and the labeling at our disposal. Two approaches are developed to avoid explicit computation of pairwise proximity between the nodes of the graph, which would be impractical for graphs containing millions of nodes. The first ap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
22
0

Year Published

2011
2011
2020
2020

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 34 publications
(23 citation statements)
references
References 48 publications
1
22
0
Order By: Relevance
“…Curiously, the spectral method applied to the three kernels (the Markov diffusion kernel (MD), the regularized commute time kernel (RCT) and the regularized Laplacian kernel (RL)) provides bad performances (all three kernels perform significantly worse than the two baselines). This is especially odd, as these kernels obtained good results when used in a sum-of-similarities context [31,66] -see the results obtained by the sum-of-similarities based on the RCT Table 4: Ranking of the different classification methods according to Borda's method (the higher score, the better). Table 3 which is not statistically different from the best method.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Curiously, the spectral method applied to the three kernels (the Markov diffusion kernel (MD), the regularized commute time kernel (RCT) and the regularized Laplacian kernel (RL)) provides bad performances (all three kernels perform significantly worse than the two baselines). This is especially odd, as these kernels obtained good results when used in a sum-of-similarities context [31,66] -see the results obtained by the sum-of-similarities based on the RCT Table 4: Ranking of the different classification methods according to Borda's method (the higher score, the better). Table 3 which is not statistically different from the best method.…”
Section: Resultsmentioning
confidence: 99%
“…‱ A sum-of-similarities (SoS) algorithm based on the regularized commute time kernel, which provided good results on large datasets in [66]; see this paper for details. This is our second baseline method, denoted as SoS.…”
Section: Compared Methodsmentioning
confidence: 99%
“…21 for the definition of RWR), and applied it to address the classification problem [74]. With this method both complexities of time and space can be reduced.…”
Section: Aucmentioning
confidence: 99%
“…Instead of considering samples as I.I.D observations, graph-based learning takes the relationships/correlations between samples to ensure effective learning. For example, graph-based approaches have been popularly used to propagate labels in semi-supervised learning [24][25][26], where training samples are connected through one or multiple graphs. A recent method [27] considers preserving global and local structures inside the training data for feature selection.…”
Section: Graph-based Learningmentioning
confidence: 99%