scholarly journals Integration of 6LOWPAN and Cloud Data Centers: The Role of Link Capacity between Sensor Networks and Cloud Data Centers

Internet of Things (IoT) and Internet of Mobile Things (IoMT) acquired widespread popularity by its ease of deployment and support for innovative applications. The sensed and aggregated data from IoT and IoMT are transferred to Cloud through Internet for analysis, interpretation and decision making. In order to generate timely response and sending back the decisions to the end users or Administrators, it is important to select appropriate cloud data centers which would process and produce responses in a shorter time. Beside several factors that determine the performance of the integrated 6LOWPAN and Cloud Data Centers, we analyze the available bandwidth between various user bases (IoT and IoMT networks) and the cloud data centers. Amidst of various services offered in cloud, problems such as congestion, delay and poor response time arises when the number of user request increases. Load balancing/sharing algorithms are the popularly used techniques to improve the performance of the cloud system. Load refers to the number of user requests (Data) from different types of networks such as IoT and IoMT which are IPv6 compliant. In this paper we investigate the impact of homogeneous and heterogeneous bandwidth between different regions in load balancing algorithms for mapping user requests (Data) to various virtual machines in Cloud. We investigate the influence of bandwidth across different regions in determining the response time for the corresponding data collected from data harvesting networks. We simulated the cloud environment with various bandwidth values between user base and data centers and presented the average response time for individual user bases. We used Cloud- Analyst an open source tool to simulate the proposed work. The obtained results can be used as a reference to map the mass data generated by various networks to appropriate data centers to produce the response in an optimal time.

Load balancing is an important aspect in cloud to share load among different virtual machines running on various physical nodes. The user response time which is an important performance metric is being highly influenced by the efficient load balancing algorithm for cloud data centers. Virtual machines which are part of the cloud data centers consist of various types of physical devices. The user response time is affected significantly by the capacity of physical devices that exist as part of the data centers. Several load balancing algorithms exist in the literature to allocate task effectively on various virtual machines running in data centers. We investigate the performance of round robin based load balancing algorithm with closest data center as service broker policy in cloud data centers. We have performed a simulation with data centers that consist of devices with different physical characteristics such as memory, storage, bandwidth, processor speed and scheduling policy using Round Robin load balancing algorithm with closest data centers as service broker policy. We present the merits of heterogeneous device characteristics in reducing the user response time and the data center request service time. We used Cloud Analyst, an open source simulation tool for cloud computing environment


Load balancing algorithms and service broker policies plays a crucial role in determining the performance of cloud systems. User response time and data center request servicing time are largely affected by the load balancing algorithms and service broker policies. Several load balancing algorithms and service broker polices exist in the literature to perform the data center allocation and virtual machine allocation for the given set of user requests. In this paper, we investigate the performance of equally spread current execution (ESCE) based load balancing algorithm with closest data center(CDC) service broker policy in a cloud environment that consists of homogeneous and heterogeneous device characteristics in data centers and heterogeneous communication bandwidth that exist between different regions where cloud data centers are deployed. We performed a simulation using CloudAnalyst an open source software with different settings of device characteristics and bandwidth. The user response time and data center request servicing time are found considerably less in heterogeneous environment.


2017 ◽  
Vol 16 (3) ◽  
pp. 6247-6253
Author(s):  
Ashima Ashima ◽  
Mrs Navjot Jyoti

Cloud computing is a vigorous technology by which a user can get software, application, operating system and hardware as a service without actually possessing it and paying only according to the usage. Cloud Computing is a hot topic of research for the researchers these days. With the rapid growth of Interne technology cloud computing have become main source of computing for small as well big IT companies. In the cloud computing milieu the cloud data centers and the users of the cloud-computing are globally situated, therefore it is a big challenge for cloud data centers to efficiently handle the requests which are coming from millions of users and service them in an efficient manner. Load balancing is a critical aspect that ensures that all the resources and entities are well balanced such that no resource or entity neither is under loaded nor overloaded. The load balancing algorithms can be static or dynamic.  Load balancing in this environment means equal distribution of workload across all the nodes. Load balancing provides a way of achieving the proper utilization of resources and better user satisfaction. Hence, use of an appropriate load balancing algorithm is necessary for selecting the virtual machines or servers. This paper focuses on the load balancing algorithm which distributes the incoming jobs among VMs optimally in cloud data centers. In this paper, we have reviewed several existing load balancing mechanisms and we have tried to address the problems associated with them.


Author(s):  
Mehran Tarahomi ◽  
Mohammad Izadi

<p>There are several physical data centers in cloud environment with hundreds or thousands of computers. Virtualization is the key technology to make cloud computing feasible. It separates virtual machines in a way that each of these so-called virtualized machines can be configured on a number of hosts according to the type of user application. It is also possible to dynamically alter the allocated resources of a virtual machine. Different methods of energy saving in data centers can be divided into three general categories: 1) methods based on load balancing of resources; 2) using hardware facilities for scheduling; 3) considering thermal characteristics of the environment. This paper focuses on load balancing methods as they act dynamically because of their dependence on the current behavior of system. By taking a detailed look on previous methods, we provide a hybrid method which enables us to save energy through finding a suitable configuration for virtual machines placement and considering special features of virtual environments for scheduling and balancing dynamic loads by live migration method.</p>


2022 ◽  
Author(s):  
Arezoo Ghasemi ◽  
Abolfazl Toroghi Haghighat ◽  
Amin Keshavarzi

Abstract The process of mapping Virtual Machines (VMs) to Physical Ma- chines (PMs), which is defined as VM placement, affects Cloud Data Centers (DCs) performance. To enhance the performance, optimal placement of VMs regarding conflicting objectives has been proposed in some research, such as Multi-Objective VM reBalance (MOVMrB) and Reinforcement Learning VM reBalance (RLVMrB) in recent years. The MOVMrB algorithm is based on the BBO meta-heuristic algorithm and the RLVMrB algorithm inspired by reinforcement learning, which in both of them the non-dominance method is used to evaluate generated solutions. Although this approach reaches accept- able results, it fails to consider other solutions which are optimal regarding all objectives, when it meets the best solution based on one of these objectives. In this paper, we propose two enhanced multi-objective algorithms, Fuzzy- RLVMrB and Fuzzy-MOVMrB, that are able to consider all objectives when evaluating candidate solutions in solution space. All four algorithms aim to balance the load between VMs in terms of processor, bandwidth, and memory as well as horizontal and vertical load balance. We simulated all algorithms using the CloudSim simulator and compared them in terms of horizontal and vertical load balance and execution time. The simulation results show that Fuzzy-RLVMrB and Fuzzy-MOVMrB algorithms outperform RLVMrB and MOVMrB algorithms in terms of vertical load balancing and horizontal load balancing. Also, the RLVMrB and Fuzzy-RLVMrB algorithms are better in execution time than the MOVMrB and Fuzzy-MOVMrB algorithms.


2021 ◽  
Vol 11 (13) ◽  
pp. 5849
Author(s):  
Nimra Malik ◽  
Muhammad Sardaraz ◽  
Muhammad Tahir ◽  
Babar Shah ◽  
Gohar Ali ◽  
...  

Cloud computing is a rapidly growing technology that has been implemented in various fields in recent years, such as business, research, industry, and computing. Cloud computing provides different services over the internet, thus eliminating the need for personalized hardware and other resources. Cloud computing environments face some challenges in terms of resource utilization, energy efficiency, heterogeneous resources, etc. Tasks scheduling and virtual machines (VMs) are used as consolidation techniques in order to tackle these issues. Tasks scheduling has been extensively studied in the literature. The problem has been studied with different parameters and objectives. In this article, we address the problem of energy consumption and efficient resource utilization in virtualized cloud data centers. The proposed algorithm is based on task classification and thresholds for efficient scheduling and better resource utilization. In the first phase, workflow tasks are pre-processed to avoid bottlenecks by placing tasks with more dependencies and long execution times in separate queues. In the next step, tasks are classified based on the intensities of the required resources. Finally, Particle Swarm Optimization (PSO) is used to select the best schedules. Experiments were performed to validate the proposed technique. Comparative results obtained on benchmark datasets are presented. The results show the effectiveness of the proposed algorithm over that of the other algorithms to which it was compared in terms of energy consumption, makespan, and load balancing.


2017 ◽  
Vol 16 (6) ◽  
pp. 6953-6961
Author(s):  
Kavita Redishettywar ◽  
Prof. Rafik Juber Thekiya

Cloud computing is a vigorous technology by which a user can get software, application, operating system and hardware as a service without actually possessing it and paying only according to the usage. Cloud Computing is a hot topic of research for the researchers these days. With the rapid growth of Interne technology cloud computing have become main source of computing for small as well big IT companies. In the cloud computing milieu the cloud data centers and the users of the cloud-computing are globally situated, therefore it is a big challenge for cloud data centers to efficiently handle the requests which are coming from millions of users and service them in an efficient manner. Load balancing ensures that no single node will be overloaded and used to distribute workload among multiple nodes. It helps to improve system performance and proper utilization of resources. We propose an improved load balancing algorithm for job scheduling in the cloud environment using K-Means clustering of cloudlets and virtual machines in the cloud environment. All the cloudlets given by the user are divided into 3 clusters depending upon client’s priority, cost and instruction length of the cloudlet. The virtual machines inside the datacenter hosts are also grouped into multiple clusters depending upon virtual machine capacity in terms of processor, memory, and bandwidth. Sorting is applied at both the ends to reduce the latency. Multiple number of experiments have been conducted by taking different configurations of cloudlets and virtual machine. Various parameters like waiting time, execution time, turnaround time and the usage cost have been computed inside the cloudsim environment to demonstrate the results. Compared with the other job scheduling algorithms, the improved load balancing algorithm can outperform them according to the experimental results.


2018 ◽  
Vol 17 (2) ◽  
pp. 7261-7272 ◽  
Author(s):  
Ishaan Chawla

Cloud computing is a vigorous technology by which a user can get software, application, operating system and hardware as a service without actually possessing it and paying only according to the usage. Cloud Computing is a hot topic of research for the researchers these days. With the rapid growth of Internet technology cloud computing have become main source of computing for small as well big IT companies. In the cloud computing milieu the cloud data centers and the users of the cloud-computing are globally situated, therefore it is a big challenge for cloud data centers to efficiently handle the requests which are coming from millions of users and service them in an efficient manner.Cloud computing is Internet based development and use of computer technology. It is a style of computing in which dynamically scalable and often virtualized resources are provided as a service over the Internet. Users need not have knowledge of, expertise in, or control over the technology infrastructure "in the cloud" that supports them. Scheduling is one of the core steps to efficiently exploit the capabilities of heterogeneous computing systems. On cloud computing platform, load balancing of the entire system can be  dynamically handled  by  using  virtualization  technology through which it  becomes  possible  to  remap  virtual  machine  and physical resources  according  to  the  change  in  load. However, in order to improve performance, the virtual machines have to fully utilize its resources and services by adapting to computing environment dynamically.  The  load balancing  with  proper  allocation  of  resources  must  be guaranteed  in  order  to  improve  resource  utility.  Load balancing is a critical aspect that ensures that all the resources and entities are well balanced such that no resource or entity neither is under loaded nor overloaded. The load balancing algorithms can be static or dynamic.  Load balancing in this environment means equal distribution of workload across all the nodes. Load balancing provides a way of achieving the proper utilization of resources and better user satisfaction. Hence, use of an appropriate load balancing algorithm is necessary for selecting the virtual machines or servers. This paper focuses on the load balancing algorithm which distributes the incoming jobs among VMs optimally in cloud data centers. In this paper, we have reviewed several existing load balancing mechanisms and we have tried to address the problems associated with them.


2017 ◽  
Vol 26 (1) ◽  
pp. 113-128
Author(s):  
Gamal Eldin I. Selim ◽  
Mohamed A. El-Rashidy ◽  
Nawal A. El-Fishawy

2021 ◽  
Vol 11 (9) ◽  
pp. 3870
Author(s):  
Jeongsu Kim ◽  
Kyungwoon Lee ◽  
Gyeongsik Yang ◽  
Kwanhoon Lee ◽  
Jaemin Im ◽  
...  

This paper investigates the performance interference of blockchain services that run on cloud data centers. As the data centers offer shared computing resources to multiple services, the blockchain services can experience performance interference due to the co-located services. We explore the impact of the interference on Fabric performance and develop a new technique to offer performance isolation for Hyperledger Fabric, the most popular blockchain platform. First, we analyze the characteristics of the different components in Hyperledger Fabric and show that Fabric components have different impacts on the performance of Fabric. Then, we present QiOi, component-level performance isolation technique for Hyperledger Fabric. The key idea of QiOi is to dynamically control the CPU scheduling of Fabric components to cope with the performance interference. We implement QiOi as a user-level daemon and evaluate how QiOi mitigates the performance interference of Fabric. The evaluation results demonstrate that QiOi mitigates performance degradation of Fabric by 22% and improves Fabric latency by 2.5 times without sacrificing the performance of co-located services. In addition, we show that QiOi can support different ordering services and chaincodes with negligible overhead to Fabric performance.


Sign in / Sign up

Export Citation Format

Share Document