scholarly journals Novel Approach to Task Scheduling and Load Balancing Using the Dominant Sequence Clustering and Mean Shift Clustering Algorithms

2019 ◽  
Vol 11 (5) ◽  
pp. 109 ◽  
Author(s):  
Amer Al-Rahayfeh ◽  
Saleh Atiewi ◽  
Abdullah Abuhussein ◽  
Muder Almiani

Cloud computing (CC) is fast-growing and frequently adopted in information technology (IT) environments due to the benefits it offers. Task scheduling and load balancing are amongst the hot topics in the realm of CC. To overcome the shortcomings of the existing task scheduling and load balancing approaches, we propose a novel approach that uses dominant sequence clustering (DSC) for task scheduling and a weighted least connection (WLC) algorithm for load balancing. First, users’ tasks are clustered using the DSC algorithm, which represents user tasks as graph of one or more clusters. After task clustering, each task is ranked using Modified Heterogeneous Earliest Finish Time (MHEFT) algorithm. where the highest priority task is scheduled first. Afterwards, virtual machines (VM) are clustered using a mean shift clustering (MSC) algorithm using kernel functions. Load balancing is subsequently performed using a WLC algorithm, which distributes the load based on server weight and capacity as well as client connectivity to server. A highly weighted or least connected server is selected for task allocation, which in turn increases the response time. Finally, we evaluate the proposed architecture using metrics such as response time, makespan, resource utilization, and service reliability.

Author(s):  
Shailendra Raghuvanshi ◽  
Priyanka Dubey

Load balancing of non-preemptive independent tasks on virtual machines (VMs) is an important aspect of task scheduling in clouds. Whenever certain VMs are overloaded and remaining VMs are under loaded with tasks for processing, the load has to be balanced to achieve optimal machine utilization. In this paper, we propose an algorithm named honey bee behavior inspired load balancing, which aims to achieve well balanced load across virtual machines for maximizing the throughput. The proposed algorithm also balances the priorities of tasks on the machines in such a way that the amount of waiting time of the tasks in the queue is minimal. We have compared the proposed algorithm with existing load balancing and scheduling algorithms. The experimental results show that the algorithm is effective when compared with existing algorithms. Our approach illustrates that there is a significant improvement in average execution time and reduction in waiting time of tasks on queue using workflowsim simulator in JAVA.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Mahfooz Alam ◽  
Mahak ◽  
Raza Abbas Haidri ◽  
Dileep Kumar Yadav

Purpose Cloud users can access services at anytime from anywhere in the world. On average, Google now processes more than 40,000 searches every second, which is approximately 3.5 billion searches per day. The diverse and vast amounts of data are generated with the development of next-generation information technologies such as cryptocurrency, internet of things and big data. To execute such applications, it is needed to design an efficient scheduling algorithm that considers the quality of service parameters like utilization, makespan and response time. Therefore, this paper aims to propose a novel Efficient Static Task Allocation (ESTA) algorithm, which optimizes average utilization. Design/methodology/approach Cloud computing provides resources such as virtual machine, network, storage, etc. over the internet. Cloud computing follows the pay-per-use billing model. To achieve efficient task allocation, scheduling algorithm problems should be interacted and tackled through efficient task distribution on the resources. The methodology of ESTA algorithm is based on minimum completion time approach. ESTA intelligently maps the batch of independent tasks (cloudlets) on heterogeneous virtual machines and optimizes their utilization in infrastructure as a service cloud computing. Findings To evaluate the performance of ESTA, the simulation study is compared with Min-Min, load balancing strategy with migration cost, Longest job in the fastest resource-shortest job in the fastest resource, sufferage, minimum completion time (MCT), minimum execution time and opportunistic load balancing on account of makespan, utilization and response time. Originality/value The simulation result reveals that the ESTA algorithm consistently superior performs under varying of batch independent of cloudlets and the number of virtual machines’ test conditions.


2019 ◽  
Vol 16 (4) ◽  
pp. 627-637
Author(s):  
Sanaz Hosseinzadeh Sabeti ◽  
Maryam Mollabgher

Goal: Load balancing policies often map workloads on virtual machines, and are being sought to achieve their goals by creating an almost equal level of workload on any virtual machine. In this research, a hybrid load balancing algorithm is proposed with the aim of reducing response time and processing time. Design / Methodology / Approach: The proposed algorithm performs load balancing using a table including the status indicators of virtual machines and the task list allocated to each virtual machine. The evaluation results of response time and processing time in data centers from four algorithms, ESCE, Throttled, Round Robin and the proposed algorithm is done. Results: The overall response time and data processing time in the proposed algorithm data center are shorter than other algorithms and improve the response time and data processing time in the data center. The results of the overall response time for all algorithms show that the response time of the proposed algorithm is 12.28%, compared to the Round Robin algorithm, 9.1% compared to the Throttled algorithm, and 4.86% of the ESCE algorithm. Limitations of the investigation: Due to time and technical limitations, load balancing has not been achieved with more goals, such as lowering costs and increasing productivity. Practical implications: The implementation of a hybrid load factor policy can improve the response time and processing time. The use of load balancing will cause the traffic load between virtual machines to be properly distributed and prevent bottlenecks. This will be effective in increasing customer responsiveness. And finally, improving response time increases the satisfaction of cloud users and increases the productivity of computing resources. Originality/Value: This research can be effective in optimizing the existing algorithms and will take a step towards further research in this regard.


Internet of Things (IoT) and Internet of Mobile Things (IoMT) acquired widespread popularity by its ease of deployment and support for innovative applications. The sensed and aggregated data from IoT and IoMT are transferred to Cloud through Internet for analysis, interpretation and decision making. In order to generate timely response and sending back the decisions to the end users or Administrators, it is important to select appropriate cloud data centers which would process and produce responses in a shorter time. Beside several factors that determine the performance of the integrated 6LOWPAN and Cloud Data Centers, we analyze the available bandwidth between various user bases (IoT and IoMT networks) and the cloud data centers. Amidst of various services offered in cloud, problems such as congestion, delay and poor response time arises when the number of user request increases. Load balancing/sharing algorithms are the popularly used techniques to improve the performance of the cloud system. Load refers to the number of user requests (Data) from different types of networks such as IoT and IoMT which are IPv6 compliant. In this paper we investigate the impact of homogeneous and heterogeneous bandwidth between different regions in load balancing algorithms for mapping user requests (Data) to various virtual machines in Cloud. We investigate the influence of bandwidth across different regions in determining the response time for the corresponding data collected from data harvesting networks. We simulated the cloud environment with various bandwidth values between user base and data centers and presented the average response time for individual user bases. We used Cloud- Analyst an open source tool to simulate the proposed work. The obtained results can be used as a reference to map the mass data generated by various networks to appropriate data centers to produce the response in an optimal time.


2021 ◽  
Vol 13 (2) ◽  
pp. 423-438
Author(s):  
B. Lakhani ◽  
A. Agrawal

One of the key challenges in the domain of cloud computing is task scheduling and estimation of cloud workloads for time critical applications pertaining to constrained cloud resources. While effective task scheduling is necessary for balancing the load, workload forecasting is necessary to plan in advance the requirements of cloud platforms based on previous data so as to effectively utilize cloud resources. Often it is challenging to gather sufficient information about the tasks and hence allocating the tasks to virtual machines (VMs) in the most optimal way is non-trivial. In this paper, a hybrid task scheduling approach is proposed based on evolutionary algorithms. The first approach is the amalgamation of bat and particle swarm optimization (PSO) techniques. The scheduling approach also combines the processing time preemption (PTP) approach to schedule the source intensive tasks which allows to reduce the response time of the proposed system.  The second approach is a machine learning based approach employing gradient descent with momentum (GDM). The evaluation of the proposed system has been done based on the response time and mean square error of the system.


Author(s):  
Er. Ruchi ◽  
Harish Kumar

Cloud computing is referred to as biggest technology of today’s environment that provide access to distributed resources on the basis of pay-per-use. Everyone try to use cloud to reduce the cost and maintenance of infrastructure due to which lots of load is increasing day by day. Therefore, there is need to balance that load since resources of cloud are limited but usage is increasing at every moment. This paper discuss how the resources are allocated and how the tasks are scheduled among those resources. Task scheduling mainly focuses on enhancing the utilization of resources and hence reduction in response time. There are various static and dynamic load balancing algorithms to balance the load, this paper discusses comparative study of these algorithms.


Author(s):  
Noha G. Elnagar ◽  
Ghada F. Elkabbany ◽  
Amr A. Al-Awamry ◽  
Mohamed B. Abdelhalim

<span lang="EN-US">Load balancing is crucial to ensure scalability, reliability, minimize response time, and processing time and maximize resource utilization in cloud computing. However, the load fluctuation accompanied with the distribution of a huge number of requests among a set of virtual machines (VMs) is challenging and needs effective and practical load balancers. In this work, a two listed throttled load balancer (TLT-LB) algorithm is proposed and further simulated using the CloudAnalyst simulator. The TLT-LB algorithm is based on the modification of the conventional TLB algorithm to improve the distribution of the tasks between different VMs. The performance of the TLT-LB algorithm compared to the TLB, round robin (RR), and active monitoring load balancer (AMLB) algorithms has been evaluated using two different configurations. Interestingly, the TLT-LB significantly balances the load between the VMs by reducing the loading gap between the heaviest loaded and the lightest loaded VMs to be 6.45% compared to 68.55% for the TLB and AMLB algorithms. Furthermore, the TLT-LB algorithm considerably reduces the average response time and processing time compared to the TLB, RR, and AMLB algorithms.</span>


Cloud computing becoming one of the most advanced and promising technologies in these days for information technology era. It has also helped to reduce the cost of small and medium enterprises based on cloud provider services. Resource scheduling with load balancing is one of the primary and most important goals of the cloud computing scheduling process. Resource scheduling in cloud is a non-deterministic problem and is responsible for assigning tasks to virtual machines (VMs) by a servers or service providers in a way that increases the resource utilization and performance, reduces response time, and keeps the whole system balanced. So in this paper, we presented a model deep learning based resource scheduling and load balancing using multidimensional queuing load optimization (MQLO) algorithm with the concept of for cloud environment Multidimensional Resource Scheduling and Queuing Network (MRSQN) is used to detect the overloaded server and migrate them to VMs. Here, ANN is used as deep learning concept as a classifier that helps to identify the overloaded or under loaded servers or VMs and balanced them based on their basis parameters such as CPU, memory and bandwidth. In particular, the proposed ANN-based MQLO algorithm has improved the response time as well success rate. The simulation results show that the proposed ANN-based MQLO algorithm has improved the response time compared to the existing algorithms in terms of Average Success Rate, Resource Scheduling Efficiency, Energy Consumption and Response Time.


Sign in / Sign up

Export Citation Format

Share Document