2010
DOI: 10.1007/978-3-642-13657-3_12
|View full text |Cite
|
Sign up to set email alerts
|

Online Sampling of High Centrality Individuals in Social Networks

Abstract: Abstract. In this work, we investigate the use of online or "crawling" algorithms to sample large social networks in order to determine the most influential or important individuals within the network (by varying definitions of network centrality). We describe a novel sampling technique based on concepts from expander graphs. We empirically evaluate this method in addition to other online sampling strategies on several realworld social networks. We find that, by sampling nodes to maximize the expansion of the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
35
0

Year Published

2011
2011
2024
2024

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 52 publications
(35 citation statements)
references
References 20 publications
0
35
0
Order By: Relevance
“…The algorithm computes, for each vertex v, the shortest path to every other vertex and then traverses these paths backwards to efficiently compute the contribution of the shortest paths from v to the betweenness of other vertices. For very large networks, the cost of this algorithm would still be prohibitive in practice, so many approximation algorithms were developed (Jacob et al 2005;Brandes and Pich 2007;Bader et al 2007;Geisberger et al 2008;Maiya and Berger-Wolf 2010;Lim et al 2011). The use of random sampling was one of the more natural approaches to speed up the computation of betweenness.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The algorithm computes, for each vertex v, the shortest path to every other vertex and then traverses these paths backwards to efficiently compute the contribution of the shortest paths from v to the betweenness of other vertices. For very large networks, the cost of this algorithm would still be prohibitive in practice, so many approximation algorithms were developed (Jacob et al 2005;Brandes and Pich 2007;Bader et al 2007;Geisberger et al 2008;Maiya and Berger-Wolf 2010;Lim et al 2011). The use of random sampling was one of the more natural approaches to speed up the computation of betweenness.…”
Section: Related Workmentioning
confidence: 99%
“…Bader et al (2007) present an adaptive sampling algorithm which computes good estimations for the betweenness of high-centrality vertices, by keeping track of the partial contribution of each sampled vertex, obtained by performing a single-source shortest paths computation to all other vertices. Maiya and Berger-Wolf (2010) use concepts from expander graphs to select a connected sample of vertices. They estimate the betweenness from the sample, which includes the vertices with high centrality.…”
Section: Related Workmentioning
confidence: 99%
“…In [20], the authors propose the Expansion Sampling algorithm which constructs a sample subgraph through maximal expansion-a neighboring node of the current sample is selected into the sample if it has the most number of neighbors that are neither within the current sample nor are neighbors of the nodes in the current sample. In [21], the authors propose the BackLink Count (BLC) algorithm which includes the node that is most connected to the current sample.…”
Section: A Subgraph Samplingmentioning
confidence: 99%
“…Another approach to extracting a subgraph for estimating the spectral radius is one based on finding the set of nodes that have the largest eigenvalue centrality within the network [20], [21]. In [20], the authors propose the Expansion Sampling algorithm which constructs a sample subgraph through maximal expansion-a neighboring node of the current sample is selected into the sample if it has the most number of neighbors that are neither within the current sample nor are neighbors of the nodes in the current sample.…”
Section: A Subgraph Samplingmentioning
confidence: 99%
“…In this setting the concept of hub has been widely studied, and is at the basis of many important applications, ranging from analysis of the structure of the Internet to web searches, from peer-to-peer network analysis to social networks, from Viral Marketing to analysis of the Blogosphere, from outbreaks of epidemics to metabolic network analysis [4,14,1,13,11,24,15,17].…”
Section: Introductionmentioning
confidence: 99%