Abstract | ||
---|---|---|
Scale-out applications have emerged as the dominant Internet services today. A request in a scale-out workload generally involves task partitioning and merging with barrier synchronization, making it difficult to predict the request tail latency to meet stringent tail Service Level Objectives (SLOs). In this paper, we find that the request tail latency can be faithfully predicted, in the high load region, by a prediction model using only the mean and variance of the task response time as input. The prediction errors for the 99th percentile request latency are found to be consistently within 10% at the load of 90%for both model and measurement-based testing cases. Consequently, the work in this paper establishes an important link between the request tail SLOs and the low order task statistics in a high load region, where the resource provisioning is desired. Finally, we discuss how the prediction model may facilitate highly scalable, tail-constrained resource provisioning for scaleout workloads. |
Year | Venue | Field |
---|---|---|
2016 | HotCloud | Synchronization,Service level objective,Computer science,Latency (engineering),Workload,Response time,Real-time computing,Provisioning,Scalability,Distributed computing,The Internet |
DocType | Citations | PageRank |
Conference | 4 | 0.39 |
References | Authors | |
13 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
minh nguyen | 1 | 7 | 2.05 |
Zhongwei Li | 2 | 5 | 1.41 |
Feng Duan | 3 | 34 | 4.43 |
Hao Che | 4 | 31 | 4.70 |
Hong Jiang | 5 | 304 | 24.60 |