2019 IEEE International Conference on Web Services (ICWS) 2019
DOI: 10.1109/icws.2019.00023
|View full text |Cite
|
Sign up to set email alerts
|

Microscaler: Automatic Scaling for Microservices with an Online Learning Approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
41
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 78 publications
(41 citation statements)
references
References 10 publications
0
41
0
Order By: Relevance
“…The authors of this study focused on microservices scheduling to minimize the end-to-end response delay under a pre-specified budget constraint. Another effort by [15] used response latency as a driving factor to trigger autoscaling cloud microservices. Moreover, [16] focused to investigate the impact of the heap size, garbage collection, concurrency and service demand on the tail latency of Java microservices.…”
Section: End-to-end Latency Predictionmentioning
confidence: 99%
“…The authors of this study focused on microservices scheduling to minimize the end-to-end response delay under a pre-specified budget constraint. Another effort by [15] used response latency as a driving factor to trigger autoscaling cloud microservices. Moreover, [16] focused to investigate the impact of the heap size, garbage collection, concurrency and service demand on the tail latency of Java microservices.…”
Section: End-to-end Latency Predictionmentioning
confidence: 99%
“…Since the offline supervised learning algorithm cannot output path indexes in real time, it is necessary to introduce online machine learning algorithms (for example, reinforcement learning) to adaptively adjust the model according to any major events. Secondly, we may also need to apply machine learning methods to dynamically build resource models under different workloads, that is, upgrade the mesh components studied to Istio-like service grid [37]. Furthermore, the throughput performance of the transport network is constrained by the request rate of the service, the stream access control strategy, the mesh network topology, and the machine learning models.…”
Section: ) Comparison Of the Number Of Outages With And Without Retransmissionmentioning
confidence: 99%
“…To set up the auto scaling thresholds, it normally involves initial guess work and couple of rounds of fine tuning based on the real time experiences in production environments. During this guessing process, it might lead to situations where either the autoscaling thresholds are under configured there by leading to scenarios where the service becomes unavailable to the callers during a peak load situation (or) extra compute resources are provisioned which might go unused, due to the over configured auto-scaling threshold values, causing monetary losses [6] [7]. Some of the researchers like Muhammad Abdullah and others [24] have solved this problem for the micro services which are already functional in the production environments by using a resource prediction model which is trained based on the historical autoscaling performance traces.…”
Section: Related Workmentioning
confidence: 99%
“…That makes the value of m, n in equation 7as 10. Using these values, the resource utilization at any point of time T can be calculated as per equation (7). To calculate the total resource utilization, we used equation (9) based on Reiman sum, by splitting the time intervals as 10,20,30,40,50 and 60 seconds.…”
Section: Figure 6 Cpu Utilization Trend For Large Functionmentioning
confidence: 99%