2010
DOI: 10.1007/978-3-642-15277-1_46
|View full text |Cite
|
Sign up to set email alerts
|

Sampling Bias in BitTorrent Measurements

Abstract: Abstract. Real-world measurements play an important role in understanding the characteristics and in improving the operation of BitTorrent, which is currently a popular Internet application. Much like measuring the Internet, the complexity and scale of the BitTorrent network make a single, complete measurement impractical. While a large number of measurements have already employed diverse sampling techniques to study parts of BitTorrent network, until now there exists no investigation of their sampling bias, t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2011
2011
2019
2019

Publication Types

Select...
5
2
1

Relationship

3
5

Authors

Journals

citations
Cited by 15 publications
(14 citation statements)
references
References 25 publications
0
14
0
Order By: Relevance
“…A number of studies have focused on measuring and characterising specific properties of BitTorrent (e.g., [5,7]); other work has introduced improvements to the measuring process itself (e.g., [24,26,27]). The limitations of the evidence collected through indirect monitoring for legal cases motivated Bauer et al [1] to design BitStalker, an active probing mechanism for identifying hosts using BitTorrent to download files.…”
Section: Related Workmentioning
confidence: 99%
“…A number of studies have focused on measuring and characterising specific properties of BitTorrent (e.g., [5,7]); other work has introduced improvements to the measuring process itself (e.g., [24,26,27]). The limitations of the evidence collected through indirect monitoring for legal cases motivated Bauer et al [1] to design BitStalker, an active probing mechanism for identifying hosts using BitTorrent to download files.…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, log traces are often problematic to obtain as they require agreement from content providers. The second group uses crawling (or spidering) techniques [8], [9], [16], [18], [20]. In order to obtain a comprehensive picture of the entire set of BitTorrent swarms, a crawler periodically contacts each tracker to obtain all the peer addresses that have been tracked.…”
Section: Tracker-level Measurementsmentioning
confidence: 99%
“…However, each swarm includes many unreachable peers (NATed, firewalled, or inactive at the time of the measurement). Owing to their inability to connect the crawler to unreachable peers, conventional peer-level measurement studies can measure only a part of all the peers in the swarm [9]. Second, the peer-level measurement method may degrade the performance of BitTorrent systems.…”
Section: Introductionmentioning
confidence: 99%
“…Another reason is the difficulty of validating BitTorrent models using synthetic data or (only) small datasets-the complexity and heterogeneity of BitTorrent increase [30] the risk of biased models and over-fitting. Since OpenBitTorrent and PublicBitTorrent are nowadays two of the most populated public trackers, we assume that the results derived from the datasets of these trackers are representative for public BitTorrent trackers.…”
Section: B Datasetsmentioning
confidence: 99%