Dynamic Resource Allocation Method for Load Balance Scheduling Over Cloud Data Center Networks

Sakshi Chhabra* and Ashutosh Kumar Singh

National Institute of Technology Kurukshetra Haryana, India

E-mail: sakshichhabra555@gmail.com; ashutosh@nitkkr.ac.in

*Corresponding Author

Received 19 May 2021; Accepted 02 August 2021; Publication 27 October 2021

Abstract

The cloud datacenter has numerous hosts as well as application requests where resources are dynamic. The demands placed on the resource allocation are diverse. These factors could lead to load imbalances, which affect scheduling efficiency and resource utilization. A scheduling method called Dynamic Resource Allocation for Load Balancing (DRALB) is proposed. The proposed solution constitutes two steps: First, the load manager analyzes the resource requirements such as CPU, Memory, Energy and Bandwidth usage and allocates an appropriate number of VMs for each application. Second, the resource information is collected and updated where resources are sorted into four queues according to the loads of resources i.e. CPU intensive, Memory intensive, Energy intensive and Bandwidth intensive. We demonstarate that SLA-aware scheduling not only facilitates the cloud consumers by resources availability and improves throughput, response time etc. but also maximizes the cloud profits with less resource utilization and SLA (Service Level Agreement) violation penalties. This method is based on diversity of client’s applications and searching the optimal resources for the particular deployment. Experiments were carried out based on following parameters i.e. average response time; resource utilization, SLA violation rate and load balancing. The experimental results demonstrate that this method can reduce the wastage of resources and reduces the traffic upto 44.89% and 58.49% in the network.

Keywords: Cloud computing, resource configuration, dynamic allocation, optimization.

1 Introduction

The rapid development of cloud computing has increased the traffic rate exponentially in the data center networks. In order to save storage, energy consumption, bandwidth and computing capacity we need efficient cloud traffic engineering within each data datacenter [1]. Therefore, how to handle such a growing amount of workload in a scalable manner and optimized the Virtual Machine (VM) Placement to accommodate the traffic growth. To maximize the resource utilization and satisfying the Service Level Agreement (SLA) for tenants are important research issues. It requires effective management of cloud resource provisioning. Cloud provisioning is the allocation of resources to the clients according to their requirements, some require powerful CPU computing capacity or high bandwidth or large amount of storage which leads the load imbalance problem. In our daily scenario, clients could experience many difficulties like long system immediate responses during bank deposits or withdrawls, real-time temperature measurements, delays etc. For these application tasks, the real-time resource manager and load balancer signify the decision about the computing resources and the load estimation [2]. Hence, task scheduling and resource management play a key role in cloud computing to maximize the diversity in client’s applications and the uncertain factors of resources. On the other hand, Service Level Agreement (SLA) is a part of service contract and one of the major considerations for every cloud’s customer. However, it becomes challenging for cloud providers to meet SLA due to dynamic multiresource sharing. For some sensitive applications, a minimum of 99.9% availability is required which difficult for various cloud computing services [3]. In this work, we aim to provide SLA aware cloud resource provision framework which ensures the Quality of Service (QOS) with least violation rate. We demonstarate that SLA-aware scheduling not only facilitates the cloud consumers by resources availability and improves throughput, response time etc. but also maximizes the cloud profit with less resources utilization and SLA violation penalty. The paper aims to propose a dynamic resource allocation method called DRALB for scheduling the workload over cloud data cenetrs. This method is based on diversity of client’s applications and searching the optimal resources for the particular deployment.

The rest of the paper is organized with following sections: Section 2 introduces the related work. Section 3 shows a proposed framework for cloud data centers. Section 4 includes performance evaluation and analysis of implementing the proposed work followed by conclusion in Section 5.

2 Related Work

To study the optimal resource scheduling during task deployment in cloud computing, several techniques have been proposed by various authors and a few of them are explained. Fung Po Tso et al. [4] have discovered a technique of improving data center utilization based on two network topologies: canonical and fat tree. This model was found effective to improve utilization by using near optimal traffic engineering and it reduces Maximum Link Utilization (MLU) and increase overall network capacity through Penalizing Exponential Flow-spliTing (PEFT) routing. Liang Quan et al. [5] have presented a reconfiguration framework based on request predication that determines the objective of relatively optimal configuration. They have evaluated their algorithm with request prediction and deal with App_VM_Configuration, Assignment Shifting and Deployment Shifting. An optimal VM Placement method for traffic scalability have been explained in [6] whch formulates the idea of Marcov Chain (MC) based solution to optimize the VM placements. This algorithm decreases the rate of exchanging traffic among racks and avoids the traffic overflow. Liyun Zuo et al. [7] have proposed a multique interlacing method based on task’s classification where resources are sorted into three queues: CPU, I/O, Memory-intensive according to their task’s requirements. This method was found effective to balance the load that were added to improve the resource utilization and performance. According to the evolved results, it was found that their algorithm is always better than previous solutions especially for large number of tasks. This application type based VM placement and allocation strategy is proposed in [8]. By comparing and analyzing the resource usage efficiency and improves the application execution, which represents better performance. Experimental results menifest the two strategies i.e. VMAllocationPolicy (VAPS) and LoadBalanceVMAllocation. The proposed method in [9] are implemented in Cloudsim have presented a dynamic hierarchical load balancing model which helps to solves the traffic scalability issue. This framework selects the most approporiate host that satisfies the multi-dimensional resource constraints over random and sequential. DHLB improves upto 66% and outperforms the existing solutions. Wanchun Dou et al. [10] have designed a hierarchical control framework for leveraging task scheduling and Resource Co-Allocation (RCA) method for the big data platform. This framework consists of four steps: (1) Meta service preprocessing (2) Resource usage monitoring (3) Resource co-allocation for meta services (4) Global resouce co-allocation. The cloud computing benefits cloud service consumers in terms of cost and helps to reduce temporal and monetary costs. By evaluation and analyzed the performance of cost optimization parameters is explained in [11]. The dynamic hyper-heuristic technique that can effectively optimize and save the cost and time of cloud service providers has been provided. Completion Time Driven Hyper-Heuristic (CTDHH) has achieved the optimal results for Scientific Workflow Applications (SWFA) datasets. As cloud consumers are requesting SLAs in order to use services with acceptable QOS. Some work has [12], presented a SLA-aware resource scheduling framework i.e. dynamic hybrid metaheuristic algorithm to maximize the profits based on Parallel SA (PSA) and simulated annealing.

All reseachers mentioned above worked against the resource scheduling and load balancing problems. Through different from the previous studies, we focus on a real-time application request types with resource configuration and optimization. The model allocates an appropriate resources to the VMs based on its types. Through analysis of our proposed algorithm, the VM allocation scheme is found suitable and guarantees that load among physical machines is well balanced and comparatively superior to the above works.

3 DRALB

In IaaS cloud data centers, when users submit the task requests, the cloud DC choose the hosts randomly to deploy the tasks. But it becomes optimal if we choose the optimal hosts for the particular task deployment. The problem of task deployment is formalized as follows: consider a set of n resources ={r1,r2,,rn} and t task requests T={T1,T2,,Tt} in the current system of cloud. The set of VMs represented by v virtual machines ={VM1,VM2,,VMv} need to be placed into physical hosts ={ph1,ph2,,php} formulated with n clients ={Cl1,Cl2,,Cln}, m server set ={s1,s2,,sm}. Given such scenario, the CSPs always desire to obtain an optimal mapping of VMs and servers to maximize the resource utilization. A mapping between physical host and clients with specific requests Θ:T× allocates each physical host from each user with specific task requests, if hosts one or more VMs, it is active, ΘT××={ΘTt,php,Cln|ΘTt,php,Cln=1 if task request T of client n is allocated to optimal physical host}. Similarly, if VMs were assigned to optimal hosts then Θi=1 otherwise Θi=0. Let i represents the resources which has iC, iM, iE and iB capacity of CPU, Memory, Energy and Bandwidth respectively. The utilization of resources for VMj are VMjC, VMjM, VMjE and VMjB. At the time of allocating tasks, firstly we check whether available memory is greater or close to the requested ones then it can only deploy the tasks. As Equation (1) ensures that the total required consumption of processors resource amount should not exceed its total capacity.

i=1VMVMjCiCVMjMiMVMjEiEVMjBiB,
sm (1)

The resource utilization of each resource is calculated independently, CPU utilization of one physical host is computed using Equation (2), memory, energy and bandwidth respectively in Equations (3)–(5). The average resource utilization of datacenter ΦDC that needs to be maximized is demonstrated using Equation (6). |N| represents the number of resources to be considered i.e. |N|=4.

ΦiC=i=1nΘi×VMjCiC (2)
ΦiM=i=1nΘi×VMjMiM (3)
ΦiE=i=1nΘi×VMjEiE (4)
ΦiB=i=1nΘi×VMjBiB (5)
ΦDC=i=1tΦi=1C+i=1tΦi=1M+i=1tΦi=1E+i=1tΦi=1B|N|×i=1tΘi (6)

images

Figure 1 DRALB framework.

The model chooses the more appropriate physical machine for the VM migration and placement to the respective upcoming applications. The effective VM placement optimizes the resource utilization, performance and energy consumption without SLA violation.

f{ph1,ph2,,phn|}=f(ph1|)×f(ph2|)××f(phn|). (7)

where are the parameters i.e. CPU, memory, energy consumption and bandwidth. The upcoming application requests are classified according to their type and resource requirements. By analyzing the resource need for the respective application, search the appropriate VM and PM where task can be deployed. For example, if the weights of an unknown application Wci=1, Wmi=0, Wei=0, Wbwi=0 is obtained through an application classifier algorithm, it means the application is CPU-intensive. Generally the usage of CPU is more for the upcoming applications as per the proposed VM allocation and placement strategy. More CPU resource should be allocated to the VM, while the other resources such as memory, energy, bandwidth, etc. can be relatively less.

3.1 SLA-aware Cloud Provision Model

This framework monitors the upcoming applications and demonstrate the SLA aware scheduling policy. The two parameters are mainly focused: Response time (RT) and Resource Utilization criteria (RUC). The response time is defined in SLA agreement to calculate the time of upcoming applications. The threshold is fixed that defines the maximum average response time to deal with tasks. As RUC threshold represents the maximum utilization of resources that every VM instance can have. If the utilization or response time exceeds the threshold number then apply penalty on it.

PF=i=0n[Pri×i-(Pnli×Ui+Cost×i)] (8)

where Pri defines the pricing, Pnli is penalty. The objective of this SLA-aware cloud load balancer is to maximize the benefit of cloud services and minimizes the penalty function. The cost and penalty should be less during cloud service delivery. The two performance parameters are focused to meet fairness constraints as per SLA violation while monitoring the applications. Figure 1 describes the architecture of proposed DRALB approach in the cloud environment. It shows the interaction between Task Scheduler and Resource Manager which plays an important role in the whole architecture. Resource manager monitors the link capacity and link state routing utilization for every host. We adopt task classifier to classify the resource requirements (CPU, memory, energy and bandwidth utilization) for each application as per SLA. When clients are requesting new batch of jobs arriving at cloud system, then load balancing is activated. The test data is monitored and initialized by load managers. Then optimal VM placement is applied and selects the most favorable machine which accomplishes and fulfill the requirement of deploying tasks. The SLA aware load balancer is divided into two types: local and global load balancer. Local load balancer monitors the load which are in same physical host to meet the SLA aware requirements. Global balancer transfer the upcoming requests to the under-loaded physical hosts as per need of SLA. Moreover, the energy consumption is calculated:

k={×[k×ðkwork+(1-k)×ðkidle],Hoststate:on×ðkstandby,Host-state:off (9)

The total energy consumption of a cloud data center is defined as =k=0nk. If the host state is on then it comes either in work state or in idle state. It is assumed that these hosts are heterogeneous and implemented in dynamic environment.

4 Performance Evaluation

The experiments are conducted using Cloudsim 3.0 and Java-Eclipse IDE on a machine equipped with Intel® Core™ I5-3230M processor of 2.60 GHZ clock speed and 8 GB of main memory to confirm the performance of DRALB.

4.1 Experimental Setup

The simulation cloud network is carried out which were generated by Cloudsim and had clear CPU, Memory, Energy and bandwidth. The results are performed on 0 to 200 physical hosts with different configurations and had 10 virtual machines on each host. We have compared our model’s performance to that of widely adopted well-known methods employed in literature, in terms of well-defined set of metrics. A series of task requests has been generated on each experiment and every task request has obvious need for computing resource amount of CPU, Memory, Energy and Bandwidth. DRALB model is compared with other three algorithms [9]: random, sequential and DHLB. The number of tasks used in this experiment to verify the model’s efficacy is T=400 (T>R) and R=1000 (T>R). The parameters used in simulation is illustrated in Table 1.

Table 1 Parameters used in simulation

Parameters Value
VM Setup of Data Center
CPU Computing ability 1860 MIPs, 2660 MIPs
Disk I/O 8 GB
RAM 4096 MB
Bandwidth 100 M/s
Storage 10 G
Task Setup of Data Center
Length (CPU) [250-1000] MIPs
File Size [100-2000] MB
Output size (Memory) [20-40] MB

4.2 Experimental Results

The following experiment first tests the makespan time which defines the total time needed for processing the tasks from begining to the end. It is mainly used in the context of scheduling when the job requests to physical hosts are assigned. The first parameter of effective load balancing is time where:

TTi=Pt+Rt+Wt (10)

where Pt, Rt and Wt are the processing time, recieving time and waiting time respectively. The second is resource utilization, DRALB minimises the wastage of resources while maximising the performance and its usage of their data centres. The third is average response time which is used to evaluate the scheduling performance. Then the load balancing verifies the load and achieves the overload avoidance for systems with multi-resource constraints. The finding of the failure nodes in the dynamic environment can only be possible if the chosen physical machine can’t fulfill some of the demands of requested tasks. Then, we calculated the SLA violation rate to measure the availability of services for customers in cloud systems.

SLAVrate=Number of Requests ViolatedNumber of Total Requests (11)

Table 2 Makespan for requested tasks

Number of VMs

Requested 100 200
Tasks RND SEQ DHLB DRALB RND SEQ DHLB DRALB
40 27.786 26.564 18.998 10.439 14.532 20.718 8.674 4.721
80 22.742 28.948 21.730 11.299 18.529 19.630 10.087 7.248
120 31.589 31.632 23.321 13.895 12.630 23.751 11.929 9.659
160 25.857 34.704 28.779 15.075 13.652 25.859 14.840 12.840
200 39.059 37.736 31.729 16.639 21.933 26.259 18.692 16.552

Table 3 Resource Utilization (in %age)

Allocation Average Average
Policies When, T < R Wastage When, T > R Wastage
RND 66.42 69.61 89.62 70.33 49.53% 69.67 78.72 70.99 85.68 40.73%
SEQ 71.14 78.41 75.61 80.52 33.24% 73.42 76.14 79.52 80.42 30.51%
DHLB 73.62 77.52 80.01 76.21 27.98% 76.42 79.32 83.67 79.42 22.61%
DRALB 70.33 79.39 75.52 77.67 20.67% 75.11 80.12 82.55 81.52 18.31%

images

Figure 2 Comparison of RND, SEQ, DHLB and DRALB in terms of Average Response Time when T=400 and T<R.

images

Figure 3 Comparison of RND, SEQ, DHLB and DRALB in terms of Average Response Time when T=1000 and T>R

images

Figure 4 Comparision of RND, SEQ, DHLB and DRALB Resource Allocation in Failures

images

Figure 5 Proportion of overflow traffic.

Table 2 illustrates the makespan which defines the total time needed for processing the tasks from begining to the end. It is mainly used in the context of scheduling when the job requests to physical hosts are assigned. Table 3 depicts the %age of resource utilisation and DRALB minimises the wastage of resources while maximising the performance and its usage of their data centres. DRALB improves the utilisation of resources up to 38.71%, 33.24% and 21.98% for random, sequential and DHLB respectively. Figure 2 shows the performance of evaluated algorithm DRALB which produces the average response time than other two heuristics when Tasks (T)=400 and Tasks < Resources (R), as it considers both the current system state and the future VM placement request arrivals. Figure 3 evaluates the average response time when (T)=1000 and T>R to analyze and evaluate whether it is good or bad in terms of its service performance. These effective measures mainly include the ability of dealing with tasks, the response time to calculate a task request and number of completed service per unit time etc. The finding of these failure nodes in the dynamic environment can only be possible if the chosen physical machine can’t fulfill some of the demands of requested tasks as demonstrated in Figure 4. Figure 5 shows the proportion of traffic overflow with other existing techniques in the simulated cloud network data center. By finding the optimal resources during load balancing, the model distributes client requests or network load efficiently across multiple servers that results in reduced traffic. During simulation, we have calculated the proportion of traffic that are sent to a server. The traffic percentage is calculated by Traffic percentage = (Assigned weight/Total weight) *100. It has been found that the proposed work reduced the traffic upto 58.49% during experimental evaluation. It achieves the overload avoidance for systems with multi-resource constraints.

5 Conclusion

This paper presents an application SLA aware resource allocation scheme to analyse the resource requirements and allocate an appropriate number of physical machines for the particular deployment. This framework monitors the upcoming applications and demonstrate the SLA aware scheduling policy. The two performance parameters are focused to meet fairness constraints as per SLA violation while monitoring the applications. Experiments show that DRALB can improve effective load balancing in contrast to sequential, random placement and DHLB method. This SLA-aware cloud load balancer maximizes the benefit of cloud services, reduces the number of failures and minimizes the penalty function. Performance evaluation demonstrates that DRALB achieves satisfactory outputs which reduces the wastage of resources and reduces the traffic upto 44.89% and 58.49% respectively for the experimental data while pointing out the observations of existing well-known algorithms.

In future work, we will study the adaptive methods to better balance the tradeoff between SLA violation and the resource efficiency. Besides, adding more objectives into our model and then implementing the algorithms in a real cloud data centre constitute our future work.

Acknowledgment

The authors would like a wonderful thanks to the DeitY (Department of Electronics and Information technology) for this research project. And we are extremely fortunate to get constant encouragement, support as well as appreciation.

References

[1] Yu, Rong, Yan Zhang, Stein Gjessing, Wenlong Xia, and Kun Yang. ‘Toward cloud-based vehicular networks with efficient resource management.’ IEEE Network 27, no. 5, 2013: 48–55.

[2] Chhabra Sakshi, and Ashutosh Kumar Singh. ‘A Probabilistic Model for Finding an Optimal Host Framework and Load Distribution in Cloud Environment.’ Procedia Computer Science 125 (2018): 683–690.

[3] Dou, Wanchun, Xiaolong Xu, Xiang Liu, Laurence T. Yang, and Yiping Wen. ‘A Resource Co-Allocation method for load-balance scheduling over big data platforms.’ Future Generation Computer Systems 86 (2018): 1064–1075.

[4] Tso, Fung Po, and Dimitrios P. Pezaros. ‘Improving data center network utilization using near-optimal traffic engineering.’ IEEE transactions on parallel and distributed systems 24, no. 6 (2013): 1139–1148.

[5] Liang, Quan, Jing Zhang, Yong-hui Zhang, and Jiu-mei Liang. ‘The placement method of resources and applications based on request prediction in cloud data center.’ Information Sciences 279 (2014): 735–745.

[6] Ma, Teng, Jiangxing Wu, Yuxiang Hu, and Wanwei Huang. ‘Optimal VM placement for traffic scalability using Markov chain in cloud data centre networks.’ Electronics Letters 53, no. 9 (2017): 602–604.

[7] Zuo, Liyun, Shoubin Dong, Lei Shu, Chunsheng Zhu, and Guangjie Han. ‘A multiqueue interlacing peak scheduling method based on tasks classification in cloud computing.’ IEEE Systems Journal 12, no. 2 (2016): 1518–1530.

[8] Peng, Jun-jie, Xiao-fei Zhi, and Xiao-lan Xie. ‘Application type based resource allocation strategy in cloud environment.’ Microprocessors and Microsystems 47 (2016): 385–391.

[9] S. Chhabra and A.K. Singh. ‘Dynamic Hierarchical Load Balancing Model for Cloud Data Center Networks.’ IET Digital Library, Volume 55, Issue 2, 24 January 2019, pp. 94–96.

[10] Dou, Wanchun, Xiaolong Xu, Xiang Liu, Laurence T. Yang, and Yiping Wen. ‘A Resource Co-Allocation method for load-balance scheduling over big data platforms.’ Future Generation Computer Systems 86 (2018): 1064–1075.

[11] Alkhanak, Ehab Nabiel, and Sai Peck Lee. ‘A hyper-heuristic cost optimisation approach for Scientific Workflow Scheduling in cloud computing.’ Future Generation Computer Systems 86 (2018): 480–506.

[12] Zhao, Yangming, Yifan Huang, Kai Chen, Minlan Yu, Sheng Wang, and DongSheng Li. ‘Joint VM placement and topology optimization for traffic scalability in dynamic datacenter networks.’ Computer Networks 80 (2015): 109–123.

Biographies

images

Sakshi Chhabra. She received the BCA degree in Computer Applications from the Punjab University, Chandigarh in 2012, and the MCA degree in Computer Applications in 2015 (INDIA). She just completed her Ph.D degree in 2020 from National Institute of Technology, Kurukshetra in Department of Computer Applications. Currently, she is working as an Assistant Professor in the above institute. Her main research interests include Cloud Computing, Load Balancing and Information Security. She has published the research papers in SCI, Scopus journals and International Conferences.

images

Ashutosh Kumar Singh. He is working as a Professor in National Institute of Technology, Kurukshetra, India. He has more than 15 years research and teaching experience in various Universities of India, UK, and Malaysia. Prior to this appointment, he has worked as an Associate Professor and Head of Department Electrical and Computer Engineering in School of Engineering Curtin University Australia offshore Campus Malaysia, Sr. Lecturer and Deputy Dean (Research and Graduate Studies) in Faculty of Information Technology, University Tun Abdul Razak Kuala Lumpur Malaysia, Post Doc RA in the Department of Computer Science, University of Bristol, Faculty of Information Science and Technology, Multimedia University Malaysia and Sr. Lecturer in Electronics and Communication Department at NIST, India. His research area includes Web Technology, Big Data, Verification, Synthesis, Design and Testing of Digital Circuits. He has published more than 300 research papers now in different journals, conferences and news magazines.

Abstract

1 Introduction

2 Related Work

3 DRALB

images

3.1 SLA-aware Cloud Provision Model

4 Performance Evaluation

4.1 Experimental Setup

4.2 Experimental Results

images

images

images

images

5 Conclusion

Acknowledgment

References

Biographies