2016
DOI: 10.1109/tnet.2015.2466453
|View full text |Cite
|
Sign up to set email alerts
|

Joint Latency and Cost Optimization for Erasure-Coded Data Center Storage

Abstract: Modern distributed storage systems offer large capacity to satisfy the exponentially increasing need of storage space. They often use erasure codes to protect against disk and node failures to increase reliability, while trying to meet the latency requirements of the applications and clients. This paper provides an insightful upper bound on the average service delay of such erasure-coded storage with arbitrary service time distribution and consisting of multiple heterogeneous files. Not only does the result su… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
115
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 83 publications
(117 citation statements)
references
References 40 publications
1
115
1
Order By: Relevance
“…In the above applications, load-balancing policies (see, e.g., [19,20,31,37,40] are usually used for assigning tasks to servers. For scenarios where either low-overhead is desired or information accessibility is constrained (such as in a distributed setting), workload arXiv:1710.00296v3 [cs.PF] 16 Sep 2018 agnostic assignment policies [19,31,37] can be preferred. Our limited fork-join model assumes a random task assignment policy, which is suitable for such application scenarios.…”
Section: Motivationmentioning
confidence: 99%
“…In the above applications, load-balancing policies (see, e.g., [19,20,31,37,40] are usually used for assigning tasks to servers. For scenarios where either low-overhead is desired or information accessibility is constrained (such as in a distributed setting), workload arXiv:1710.00296v3 [cs.PF] 16 Sep 2018 agnostic assignment policies [19,31,37] can be preferred. Our limited fork-join model assumes a random task assignment policy, which is suitable for such application scenarios.…”
Section: Motivationmentioning
confidence: 99%
“…Since the problem of finding optimal assignment is challenging, we use a probabilistic scheduling approach where the VMs are assigned to jobs with certain probabilities which can be optimized for improved performance. Such scheduling approaches have been used in [16]. This approach results in the arrival distribution for the different types of jobs in the networking phase also being a Poisson process.…”
Section: Introductionmentioning
confidence: 99%
“…By optimizing these probabilities, we quantify SDTP through a closed-form, tight upper bound for CDNbased video streaming with arbitrary cache content placement and network resource allocation. We note that the analysis in this paper is fundamentally different from those for distributed file storage, e.g., [7], [8], because the stall duration of a video relies on the download times of all its chunks, rather than simply the time to download the last chunk of a file. Further, since video chunks are downloaded and played sequentially, the download times and playback times of different video chunks are highly correlated and thus jointly determine the SDTP metric.…”
Section: Introductionmentioning
confidence: 99%
“…Mean latency and tail latency have been characterized in [7], [8] and [22], [23], respectively, for a system with multiple files using probabilistic scheduling. However, these papers consider only file downloading rather than video streaming.…”
Section: Introductionmentioning
confidence: 99%