scholarly journals Layer-4 Load Balancer for Flow Size Prediction with TCP/UDP Separation Using P4

2021 ◽  
pp. 3152-3166
Author(s):  
Sanaa Alaa Hussein ◽  
Mustafa Ismael Salman

        Nowadays, datacenters become more complicated and handle many more users’ requests. Custom protocols are becoming more demanded, and an advanced load balancer to distribute the  requests among servers is essential to serve the users quickly and efficiently. P4 introduced a new way to manipulate all packet headers. Therefore, by making use of the P4 ability to decapsulate the transport layer header, a new algorithm of load balancing is proposed. The algorithm has three main parts. First, a TCP/UDP separation  is used to separate the flows based on the network layer information about the used protocol in the transport layer. Second, a flow size prediction technique is adopted, which relies on the service port number of the transport layer. Lastly, a probing system is considered to detect and solve the failure of the link and server. The proposed load balancer enhances response time of both resources usage and packet processing of the datacenter. Also, our load balancer improves link failure detection by developing a custom probing protocol.

The process of analyzing big data and other valuable information is a significant process in the cloud. Since big data processing utilizes a large number of resources for completing certain tasks. Therefore, the incoming tasks are allocated with better utilization of resources to minimize the workload across the server in the cloud. The conventional load balancing technique failed to balance the load effectively among data centers and dynamic QoS requirements of big data application. In order to improve the load balancing with maximum throughput and minimum makespan, Support Vector Regression based MapReduce Throttled Load Balancing (SVR-MTLB) technique is introduced. Initially, a large number of cloud user requests (data/file) are sent to the cloud server from different locations. After collecting the cloud user request, the SVR-MTLB technique balances the workload of the virtual machine with the help of support vector regression. The load balancer uses the index table for maintaining the virtual machines. Then, map function performs the regression analysis using optimal hyperplane and provides three resource status of the virtual machine namely overloaded, less loaded and balanced load. After finding the less loaded VM, the load balancer sends the ID of the virtual machine to the data center controller. The controller performs migration of the task from an overloaded VM to a less loaded VM at run time. This in turn assists to minimize the response time. Experimental evaluation is carried out on the factors such as throughput, makespan, migration time and response time with respect to a number of tasks. The experimental results reported that the proposed SVR-MTLB technique obtains high throughput with minimum response time, makespan as well as migration time than the state -of -the -art methods.


Author(s):  
Noha G. Elnagar ◽  
Ghada F. Elkabbany ◽  
Amr A. Al-Awamry ◽  
Mohamed B. Abdelhalim

<span lang="EN-US">Load balancing is crucial to ensure scalability, reliability, minimize response time, and processing time and maximize resource utilization in cloud computing. However, the load fluctuation accompanied with the distribution of a huge number of requests among a set of virtual machines (VMs) is challenging and needs effective and practical load balancers. In this work, a two listed throttled load balancer (TLT-LB) algorithm is proposed and further simulated using the CloudAnalyst simulator. The TLT-LB algorithm is based on the modification of the conventional TLB algorithm to improve the distribution of the tasks between different VMs. The performance of the TLT-LB algorithm compared to the TLB, round robin (RR), and active monitoring load balancer (AMLB) algorithms has been evaluated using two different configurations. Interestingly, the TLT-LB significantly balances the load between the VMs by reducing the loading gap between the heaviest loaded and the lightest loaded VMs to be 6.45% compared to 68.55% for the TLB and AMLB algorithms. Furthermore, the TLT-LB algorithm considerably reduces the average response time and processing time compared to the TLB, RR, and AMLB algorithms.</span>


2019 ◽  
Vol 28 (2) ◽  
pp. 298-339
Author(s):  
Dima Mansour ◽  
Haidar Osman ◽  
Christian Tschudin

AbstractLoad balancing is a mechanism to distribute client requests among several service instances. It enables resource utilization, lowers response time, and increases user satisfaction. In Named-Data Networking (NDN) and NDN-like architectures, load balancing becomes crucial when dynamic services are present, where relying solely on forwarding strategies can overload certain service instances while others are underutilized especially with the limited benefit of on-path caching when it comes to services. To understand the challenges and opportunities of load balancing in NDN, we analyze conventional load balancing in IP networks, and three closely related fields in NDN: congestion control, forwarding strategies, and data center management. We identify three possible scenarios for load balancing in NDN: facade load balancer, controller for Interest queues, and router-based load balancing. These different solutions use different metrics to identify the load on replicas, have different compliance levels with NDN, and place the load balancing functionality in different network components. From our findings, we propose and implement a new lightweight router-based load balancing approach called the communicating vessels and experimentally show how it reduces service response time and senses server capabilities without probing.


2020 ◽  
Vol 10 (2) ◽  
pp. 22-35 ◽  
Author(s):  
Sudipta Sahana ◽  
Tanmoy Mukherjee ◽  
Debabrata Sarddar

Cloud load balancing has become one of the most vital aspects of Cloud computing that has captured the attention of IT organizations and business firms in recent years. Among the issues related to this particular aspect, one such issue which needs to be addressed is the issue of effectively serving the clients' requests among multiple servers using an appropriate load balancer. Previous survey papers discussed various issues of cloud load balancing and accordingly devised various methods and techniques to address those issues with the objectives of reduction of processing time and response time along with optimization of costs. In this article, we have discussed an effective load balancing technique using the weighted Round-Robin algorithm which can process the client requests among multiple servers with minimal response time. Considering all these aspects, a cloud-based dynamic load balancer is being used to solve the problem of load balancing in the cloud infrastructure.


2021 ◽  
Vol 6 (1) ◽  
pp. 103
Author(s):  
Hardiyan Kesuma Ramadhan ◽  
Sukma Wardhana

In the digital era and the outbreak of the COVID-19 pandemic, all activities are online. If the number of users accessing the server exceeds IT infrastructure, server down occurs. A load balancer device is required to share the traffic request load. This study compares four algorithms on Citrix ADC VPX load balancer: round-robin, least connection, least response time and least packet using GNS3. The results of testing response time and throughput parameters show that the least connection algorithm is superior. There were a 33% reduction in response time and a 53% increase in throughput. In the service hits parameter, the round-robin algorithm has the evenest traffic distribution. While least packet superior in CPU utilization with 76% reduction. So algorithm with the best response time and throughput is the least connection. The algorithm with the best service hits is round-robin. Large scale implementation is recommended using the least connection algorithm regarding response time and throughput. When emphasizing evenest distribution, use a round-robin algorithm.


Electronics ◽  
2021 ◽  
Vol 10 (3) ◽  
pp. 323
Author(s):  
Marwa A. Abdelaal ◽  
Gamal A. Ebrahim ◽  
Wagdy R. Anis

The widespread adoption of network function virtualization (NFV) leads to providing network services through a chain of virtual network functions (VNFs). This architecture is called service function chain (SFC), which can be hosted on top of commodity servers and switches located at the cloud. Meanwhile, software-defined networking (SDN) can be utilized to manage VNFs to handle traffic flows through SFC. One of the most critical issues that needs to be addressed in NFV is VNF placement that optimizes physical link bandwidth consumption. Moreover, deploying SFCs enables service providers to consider different goals, such as minimizing the overall cost and service response time. In this paper, a novel approach for the VNF placement problem for SFCs, called virtual network functions and their replica placement (VNFRP), is introduced. It tries to achieve load balancing over the core links while considering multiple resource constraints. Hence, the VNF placement problem is first formulated as an integer linear programming (ILP) optimization problem, aiming to minimize link bandwidth consumption, energy consumption, and SFC placement cost. Then, a heuristic algorithm is proposed to find a near-optimal solution for this optimization problem. Simulation studies are conducted to evaluate the performance of the proposed approach. The simulation results show that VNFRP can significantly improve load balancing by 80% when the number of replicas is increased. Additionally, VNFRP provides more than a 54% reduction in network energy consumption. Furthermore, it can efficiently reduce the SFC placement cost by more than 67%. Moreover, with the advantages of a fast response time and rapid convergence, VNFRP can be considered as a scalable solution for large networking environments.


2021 ◽  
Author(s):  
Rashid Khogali

We synthesize online scheduling algorithms to optimally assign a set of arriving heterogeneous tasks to heterogeneous speed-scalable processors under the single threaded computing architecture. By using dynamic speed-scaling, where each processor's speed is able to dynamically change within hardware and software processing constraints, the goal of our algorithms is to minimize the total financial cost (in dollars) of response time and energy consumption (TCRTEC) of the tasks. In our work, the processors are heterogeneous in that they may differ in their hardware specifications with respect to maximum processing rate, power function parameters and energy sources. Tasks are heterogeneous in terms of computation volume, memory and minimum processing requirements. We also consider that the unit price of response time for each task is heterogeneous because the user may be willing to pay higher/lower unit prices for certain tasks, thereby increasing/decreasing their optimum processing rates. We model the overhead loading time incurred when a task is loaded by a given processor prior to its execution and assume it to be heterogeneous as well. Under the single threaded, single buffered computing architecture, we synthesize the SBDPP algorithm and its two other versions. Its first two versions allow the user to specify the unit price of energy and response time for executing each arriving task. The algorithm's second version extends the functionality of the first by allowing the user or the OS of the computing device to further modify a task's unit price of time or energy in order to achieve a linearly controlled operation point that lies somewhere in the economy-performance mode continuum of a task's execution. The algorithm's third version operates exclusively on the latter. We briefly extend the algorithm and its versions to consider migration, where an unfinished task is paused and resumed on another processor. The SBDPP algorithm is qualitatively compared against its two other versions. The SBDPP dispatcher is analytically shown to perform better than the well known Round Robin dispatcher in terms of the TCRTEC performance metric. Through simulations we deduce a relationship between the arrival rate of tasks, number of processors and response time of tasks. Under the Single threaded, multi-buffered computing architecture we have four contributions that constitute the SMBSPP algorithm. First, we propose a novel task dispatching strategy for assigning the tasks to the processors. Second, we propose a novel preemptive service discipline called Smallest remaining Computation Volume Per unit Price of response Time (SCVPPT) to schedule the tasks on the assigned processor. Third, we propose a dynamic speed-scaling function that explicitly determines the optimum processing rate of each task. Most of the simulations consider both stochastic and deterministic traffic conditions. Our simulation results show that SCVPPT outperforms the two known service disciplines, Shortest Remaining Processing Time (SRPT) and the First Come First Serve (FCFS), in terms of minimizing the TCRTEC performance metric. The results also show that the algorithm's dispatcher drastically outperforms the well known Round Robin dispatcher with cost savings exceeding 100% even when the processors are mildly heterogeneous. Finally, analytical and simulation results show that our speed scaling function performs better than a comparable speed scaling function in current literature. Under a fixed budget of energy, we synthesize the SMBAD algorithm which uses the micro-economic laws of Supply and Demand (LSD) to heuristically adjust the unit price of energy in order to extend battery life and execute more than 50% of tasks on a single processor (under the single threaded, multi buffered computing architecture). By extending all our multiprocessor algorithms to factor independent (battery) energy sources that is associated with each processor, we analytically show that load balancing effects are induced on hetergeneous parallel processors. This happens when the unit price of energy is adjusted by the battery level of each processor in accordance with LSD. Furthermore, we show that a variation of this load balancing effect also occurs when the heterogeneous processors use a single battery as long as they operate at unconstrained processing rates.


Sign in / Sign up

Export Citation Format

Share Document