AutoScaling (ResearchPPT)
AutoScaling (ResearchPPT)
J I N G J I A N G , J I E LU , G U A N G Q U A N Z H A N G , G U O D O N G LO N G
2 0 1 3 1 3 T H I E E E / A C M I N T E R N AT I O N A L S Y M P O S I U M O N
C LU S T E R , C LO U D , A N D G R I D C O M P U T I N G
Presented By Nayan,Naveen,Digvijay
INTRODUCTION
Web
System Modeling
Web application providers can specify their budgetary
constraints and SLA in respect of latency for their
applications.
In practical applications, an unpredictable burst of
number of requests will happen.
Proposed scheme:
This scheme monitors the waiting queue of requests to be
processed in real-time. Once the length of the queue is
bigger than a threshold, the scheme could dynamically
append VMs to process the exceeding number of requests.
Steps followed in
Scheme
To collect request records as the history data.
To analyze the history data hourly and predict the
number of requests for the next time-unit.
To discover the optimal number of VMs by utilizing the
Optimization Model.
To scale the resource(VMs) up or down or NOP from a
public cloud platform
The current and past observed values :- (X(t k), X(t k + 1), ...X(t 1),
X(t)),
Future value :- X(t + p)
where k is the length of the history data used for prediction and p is the
predictive time fragment
Markov chain
Optimization Model
Cost, latency tradeoff
If virtual machines are reduced then cost reduced but waiting
time increases
If VM increased then cost increase
Optimization Model
Cont..
Due to the different scale of the number of VMs and latency, we can
normalize the latency by equation
G = L/T
where T is the latency threshold, which is defined in SLAs.
Experiment
Real-world datasets:
1. AOL 1 search log dataset
2. Sogou 2 search log dataset
3. A real-world dataset collected by the UTS (The
University of Technology, Sydney) library
Time-unit of re-allocation: on hourly basis.
Experiment Steps
1. Investigate how the seasonal characters
affect the selection of features for
prediction modeling
2. Evaluate the prediction model through three datasets
3. Visualize the performance of the prediction model
4. Evaluate the allocation performance for the given
number of requests
5. Compare our scheme with other approaches
Features Selection
Evaluation
Investigate how the seasonal characters
affect the selection of features for
prediction modeling
Number of requests in each hour as a vector <v1, ..., vi,
...v60>,
(vi - volume within one minute)
Apply the Symmetrizing Kullback-Leibler (KL)
divergence to measure the difference between two
distribution probabilities.
Cont..
Conclusion:
If SKL divergence varies from 0 to 1.
All the SKL divergences in Table I are small, which
demonstrates that the three datasets have highly seasonal
characters and the number of requests can be predicted by
using the history data.
Cont..
select the key features for the linear regression model
Experiment Steps
1. Investigate how the seasonal characters affect the
selection of features for prediction modeling
Evaluation Methods
Used 10-fold cross validation as the evaluation method
Root Mean Squared Error (RMSE)
Relative Squared Error (RSE)
Mean Absolute Error (MAE)
Relative Absolute Error (RAE)
coefficient of determination (R2)
Experiment Steps
1. Investigate how the seasonal characters affect the
selection of features for prediction modeling
2. Evaluate the prediction model through three datasets
Prediction Model
Evaluation
U = (1 + padding) prediction
To select a good padding value, we measure the
relationship between the padding value and the
confidence interval
Conclusion:
our scheme achieves good prediction on both number of
requests and resource demands, and that the padding
value can be dynamically adjusted well in each time
interval
Cont..
Experiment Steps
1. Investigate how the seasonal characters affect the
selection of features for prediction modeling
2. Evaluate the prediction model through three datasets
3. Visualize the performance of the prediction model
Allocation Evaluation
Cont..
Experiment Steps
1. Investigate how the seasonal characters affect the
selection of features for prediction modeling
2. Evaluate the prediction model through three datasets
3. Visualize the performance of the prediction model
4. Evaluate the allocation performance for the given
number of requests
Performance Evaluation
for a Web Application
Problem identified
Proposed algorithm adds virtual machines only if SLA
doesnt violate. But in case of sudden traffic burst there
may be need of additional resources beyond the limit
given in SLA. So we need a better trade off between SLA
violation and poor latency in burst traffic.
Thank You