scholarly journals Support Vector Regression based Mapreduce Throttled Load Balancer for Data Centers

The process of analyzing big data and other valuable information is a significant process in the cloud. Since big data processing utilizes a large number of resources for completing certain tasks. Therefore, the incoming tasks are allocated with better utilization of resources to minimize the workload across the server in the cloud. The conventional load balancing technique failed to balance the load effectively among data centers and dynamic QoS requirements of big data application. In order to improve the load balancing with maximum throughput and minimum makespan, Support Vector Regression based MapReduce Throttled Load Balancing (SVR-MTLB) technique is introduced. Initially, a large number of cloud user requests (data/file) are sent to the cloud server from different locations. After collecting the cloud user request, the SVR-MTLB technique balances the workload of the virtual machine with the help of support vector regression. The load balancer uses the index table for maintaining the virtual machines. Then, map function performs the regression analysis using optimal hyperplane and provides three resource status of the virtual machine namely overloaded, less loaded and balanced load. After finding the less loaded VM, the load balancer sends the ID of the virtual machine to the data center controller. The controller performs migration of the task from an overloaded VM to a less loaded VM at run time. This in turn assists to minimize the response time. Experimental evaluation is carried out on the factors such as throughput, makespan, migration time and response time with respect to a number of tasks. The experimental results reported that the proposed SVR-MTLB technique obtains high throughput with minimum response time, makespan as well as migration time than the state -of -the -art methods.

2019 ◽  
Vol 16 (4) ◽  
pp. 627-637
Author(s):  
Sanaz Hosseinzadeh Sabeti ◽  
Maryam Mollabgher

Goal: Load balancing policies often map workloads on virtual machines, and are being sought to achieve their goals by creating an almost equal level of workload on any virtual machine. In this research, a hybrid load balancing algorithm is proposed with the aim of reducing response time and processing time. Design / Methodology / Approach: The proposed algorithm performs load balancing using a table including the status indicators of virtual machines and the task list allocated to each virtual machine. The evaluation results of response time and processing time in data centers from four algorithms, ESCE, Throttled, Round Robin and the proposed algorithm is done. Results: The overall response time and data processing time in the proposed algorithm data center are shorter than other algorithms and improve the response time and data processing time in the data center. The results of the overall response time for all algorithms show that the response time of the proposed algorithm is 12.28%, compared to the Round Robin algorithm, 9.1% compared to the Throttled algorithm, and 4.86% of the ESCE algorithm. Limitations of the investigation: Due to time and technical limitations, load balancing has not been achieved with more goals, such as lowering costs and increasing productivity. Practical implications: The implementation of a hybrid load factor policy can improve the response time and processing time. The use of load balancing will cause the traffic load between virtual machines to be properly distributed and prevent bottlenecks. This will be effective in increasing customer responsiveness. And finally, improving response time increases the satisfaction of cloud users and increases the productivity of computing resources. Originality/Value: This research can be effective in optimizing the existing algorithms and will take a step towards further research in this regard.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Mahfooz Alam ◽  
Mahak ◽  
Raza Abbas Haidri ◽  
Dileep Kumar Yadav

Purpose Cloud users can access services at anytime from anywhere in the world. On average, Google now processes more than 40,000 searches every second, which is approximately 3.5 billion searches per day. The diverse and vast amounts of data are generated with the development of next-generation information technologies such as cryptocurrency, internet of things and big data. To execute such applications, it is needed to design an efficient scheduling algorithm that considers the quality of service parameters like utilization, makespan and response time. Therefore, this paper aims to propose a novel Efficient Static Task Allocation (ESTA) algorithm, which optimizes average utilization. Design/methodology/approach Cloud computing provides resources such as virtual machine, network, storage, etc. over the internet. Cloud computing follows the pay-per-use billing model. To achieve efficient task allocation, scheduling algorithm problems should be interacted and tackled through efficient task distribution on the resources. The methodology of ESTA algorithm is based on minimum completion time approach. ESTA intelligently maps the batch of independent tasks (cloudlets) on heterogeneous virtual machines and optimizes their utilization in infrastructure as a service cloud computing. Findings To evaluate the performance of ESTA, the simulation study is compared with Min-Min, load balancing strategy with migration cost, Longest job in the fastest resource-shortest job in the fastest resource, sufferage, minimum completion time (MCT), minimum execution time and opportunistic load balancing on account of makespan, utilization and response time. Originality/value The simulation result reveals that the ESTA algorithm consistently superior performs under varying of batch independent of cloudlets and the number of virtual machines’ test conditions.


Load balancing algorithms and service broker policies plays a crucial role in determining the performance of cloud systems. User response time and data center request servicing time are largely affected by the load balancing algorithms and service broker policies. Several load balancing algorithms and service broker polices exist in the literature to perform the data center allocation and virtual machine allocation for the given set of user requests. In this paper, we investigate the performance of equally spread current execution (ESCE) based load balancing algorithm with closest data center(CDC) service broker policy in a cloud environment that consists of homogeneous and heterogeneous device characteristics in data centers and heterogeneous communication bandwidth that exist between different regions where cloud data centers are deployed. We performed a simulation using CloudAnalyst an open source software with different settings of device characteristics and bandwidth. The user response time and data center request servicing time are found considerably less in heterogeneous environment.


Author(s):  
K. Balaji, Et. al.

The evolution of IT led Cloud computing technology emerge as a new prototype in providing the services to its users on rented basis at any time or place. Considering the flexibility of cloud services, innumerable organizations switched their businesses to the cloud technology by setting up more data centers. Nevertheless, it has become mandatory to provide profitable execution of tasks and appropriate  resource utilization. A few approaches were outlined in literature to enhance performance, job scheduling, storage resources, QoS and load distribution. Load balancing concept permits data centers to avert over-loading or under-loading in virtual machines that as such is an issue in cloud computing domain. Consequently, it necessitate the researchers to layout and apply a proper load balancer for cloud environment. The respective study represents a view of problems and threats faced by the current load balancing techniques and make the researchers find more efficient algorithms.


Author(s):  
K. Balaji , Et. al.

The evolution of IT led Cloud computing technology emerge as a new prototype in providing the services to its users on rented basis at any time or place. Considering the flexibility of cloud services, innumerable organizations switched their businesses to the cloud technology by setting up more data centers. Nevertheless, it has become mandatory to provide profitable execution of tasks and appropriate  resource utilization. A few approaches were outlined in literature to enhance performance, job scheduling, storage resources, QoS and load distribution. Load balancing concept permits data centers to avert over-loading or under-loading in virtual machines that as such is an issue in cloud computing domain. Consequently, it necessitate the researchers to layout and apply a proper load balancer for cloud environment. The respective study represents a view of problems and threats faced by the current load balancing techniques and make the researchers find more efficient algorithms.


Internet of Things (IoT) and Internet of Mobile Things (IoMT) acquired widespread popularity by its ease of deployment and support for innovative applications. The sensed and aggregated data from IoT and IoMT are transferred to Cloud through Internet for analysis, interpretation and decision making. In order to generate timely response and sending back the decisions to the end users or Administrators, it is important to select appropriate cloud data centers which would process and produce responses in a shorter time. Beside several factors that determine the performance of the integrated 6LOWPAN and Cloud Data Centers, we analyze the available bandwidth between various user bases (IoT and IoMT networks) and the cloud data centers. Amidst of various services offered in cloud, problems such as congestion, delay and poor response time arises when the number of user request increases. Load balancing/sharing algorithms are the popularly used techniques to improve the performance of the cloud system. Load refers to the number of user requests (Data) from different types of networks such as IoT and IoMT which are IPv6 compliant. In this paper we investigate the impact of homogeneous and heterogeneous bandwidth between different regions in load balancing algorithms for mapping user requests (Data) to various virtual machines in Cloud. We investigate the influence of bandwidth across different regions in determining the response time for the corresponding data collected from data harvesting networks. We simulated the cloud environment with various bandwidth values between user base and data centers and presented the average response time for individual user bases. We used Cloud- Analyst an open source tool to simulate the proposed work. The obtained results can be used as a reference to map the mass data generated by various networks to appropriate data centers to produce the response in an optimal time.


Author(s):  
Junlin Sun ◽  
Yi Zhang

In the big data platform, because of the large amount of data, the problem of load imbalance is prominent. Most of the current load balancing methods have problems such as high data flow loss rate and long response time; therefore, more effective load balancing method is urgently needed. Taking HBase as the research subject, the study analyzed the dynamic load balancing method of data flow. First, the HBase platform was introduced briefly, and then the dynamic load-balancing algorithm was designed. The data flow was divided into blocks, and then the load of nodes was predicted based on the grey prediction GM(1,1) model. Finally, the load was migrated through the dynamic adjustable method to achieve load balancing. The experimental results showed that the accuracy of the method for load prediction was high, the average error percentage was 0.93%, and the average response time was short; under 3000 tasks, the response time of the method designed in this study was 14.17% shorter than that of the method combining TV white space (TVWS) and long-term evolution (LTE); the average flow of nodes with the largest load was also smaller, and the data flow loss rate was basically 0%. The experimental results show the effectiveness of the proposed method, which can be further promoted and applied in practice.


Sign in / Sign up

Export Citation Format

Share Document