scholarly journals Presenting Migration Scheduling Algorithm for Virtual Machines to Optimize Energy Consumption Simultaneously and Generate Pollutants in super Computing Networks

2019 ◽  
2014 ◽  
Vol 1046 ◽  
pp. 508-511
Author(s):  
Jian Rong Zhu ◽  
Yi Zhuang ◽  
Jing Li ◽  
Wei Zhu

How to reduce energy consumption while improving utility of datacenter is one of the key technologies in the cloud computing environment. In this paper, we use energy consumption and utility of data center as objective functions to set up a virtual machine scheduling model based on multi-objective optimization VMSA-MOP, and design a virtual machine scheduling algorithm based on NSGA-2 to solve the model. Experimental results show that compared with other virtual machine scheduling algorithms, our algorithm can obtain relatively optimal scheduling results.


2014 ◽  
Vol 24 (3) ◽  
pp. 535-550 ◽  
Author(s):  
Jiaqi Zhao ◽  
Yousri Mhedheb ◽  
Jie Tao ◽  
Foued Jrad ◽  
Qinghuai Liu ◽  
...  

Abstract Scheduling virtual machines is a major research topic for cloud computing, because it directly influences the performance, the operation cost and the quality of services. A large cloud center is normally equipped with several hundred thousand physical machines. The mission of the scheduler is to select the best one to host a virtual machine. This is an NPhard global optimization problem with grand challenges for researchers. This work studies the Virtual Machine (VM) scheduling problem on the cloud. Our primary concern with VM scheduling is the energy consumption, because the largest part of a cloud center operation cost goes to the kilowatts used. We designed a scheduling algorithm that allocates an incoming virtual machine instance on the host machine, which results in the lowest energy consumption of the entire system. More specifically, we developed a new algorithm, called vision cognition, to solve the global optimization problem. This algorithm is inspired by the observation of how human eyes see directly the smallest/largest item without comparing them pairwisely. We theoretically proved that the algorithm works correctly and converges fast. Practically, we validated the novel algorithm, together with the scheduling concept, using a simulation approach. The adopted cloud simulator models different cloud infrastructures with various properties and detailed runtime information that can usually not be acquired from real clouds. The experimental results demonstrate the benefit of our approach in terms of reducing the cloud center energy consumption


2018 ◽  
Vol 7 (1) ◽  
pp. 16-19
Author(s):  
Anupama Gupta ◽  
Kulveer Kaur ◽  
Rajvir Kaur

Cloud computing is the architecture in which cloudlets are executed by the virtual machines. The most applicable virtual machines are selected on the basis of execution time and failure rate. Due to virtual machine overloading, the execution time and energy consumption is increased at steady rate. In this paper, BFO technique is applied in which weight of each virtual machine is calculated and the virtual machine which has the maximum weight is selected on which cloudlet will be migrated. The performance of proposed algorithm is tested by implementing it in CloudSim and analyzing it in terms of execution time, energy consumption.


Author(s):  
Weiwei Lin ◽  
Chao Yang ◽  
Chaoyue Zhu ◽  
James Z. Wang ◽  
Zhiping Peng

Nowadays, how to improve energy efficiency has become a challenging problem in cloud computing. However, most existing efforts in improving the energy efficiency of a cloud system only focus on resource allocation at the system level like managing physical nodes or virtual machines. This paper tries to address the energy efficiency problem of a heterogeneous cloud system at the task scheduling level. A novel task scheduling algorithm is proposed to reduce the energy consumption of the system while maintaining its performance, without closing or consolidating any system resources such as virtual machines or storage systems. The algorithm dynamically monitors CPU and memory load information of participating nodes on a heterogeneous Hadoop platform with Ganglia, then selects and submits an appropriate task to the node with relatively low workload to avoid excessive energy consumption on some nodes. Experimental results show that this novel scheduling algorithm can effectively improve the energy-saving ratio of a heterogeneous cloud platform while maintaining a high system performance.


2014 ◽  
Vol 2014 ◽  
pp. 1-7 ◽  
Author(s):  
Guangyu Du ◽  
Hong He ◽  
Qinggang Meng

Data centers, as resource providers, take advantage of virtualization technology to achieve excellent resource utilization, scalability, and high availability. However, large numbers of computing servers containing virtual machines of data centers consume a tremendous amount of energy. Thus, it is necessary to significantly improve resource utilization. Among the many issues associated with energy, scheduling plays a very important role in successful task execution and energy consumption in virtualized environments. This paper seeks to implement an energy-efficient task scheduling algorithm for virtual machines with changeless speed comprised of two main steps: assigning as many tasks as possible to virtual machines with lower energy consumption and keeping the makespan of each virtual machine within a deadline. We propose a novel scheduling algorithm for heterogeneous virtual machines in virtualized environments to effectively reduce energy consumption and finish all tasks before a deadline. The new scheduling strategy is simulated using the CloudSim toolkit package. Experimental results show that our approach outperforms previous scheduling methods by a significant margin in terms of energy consumption.


Author(s):  
Shailendra Raghuvanshi ◽  
Priyanka Dubey

Load balancing of non-preemptive independent tasks on virtual machines (VMs) is an important aspect of task scheduling in clouds. Whenever certain VMs are overloaded and remaining VMs are under loaded with tasks for processing, the load has to be balanced to achieve optimal machine utilization. In this paper, we propose an algorithm named honey bee behavior inspired load balancing, which aims to achieve well balanced load across virtual machines for maximizing the throughput. The proposed algorithm also balances the priorities of tasks on the machines in such a way that the amount of waiting time of the tasks in the queue is minimal. We have compared the proposed algorithm with existing load balancing and scheduling algorithms. The experimental results show that the algorithm is effective when compared with existing algorithms. Our approach illustrates that there is a significant improvement in average execution time and reduction in waiting time of tasks on queue using workflowsim simulator in JAVA.


Author(s):  
Gurpreet Singh ◽  
Manish Mahajan ◽  
Rajni Mohana

BACKGROUND: Cloud computing is considered as an on-demand service resource with the applications towards data center on pay per user basis. For allocating the resources appropriately for the satisfaction of user needs, an effective and reliable resource allocation method is required. Because of the enhanced user demand, the allocation of resources has now considered as a complex and challenging task when a physical machine is overloaded, Virtual Machines share its load by utilizing the physical machine resources. Previous studies lack in energy consumption and time management while keeping the Virtual Machine at the different server in turned on state. AIM AND OBJECTIVE: The main aim of this research work is to propose an effective resource allocation scheme for allocating the Virtual Machine from an ad hoc sub server with Virtual Machines. EXECUTION MODEL: The execution of the research has been carried out into two sections, initially, the location of Virtual Machines and Physical Machine with the server has been taken place and subsequently, the cross-validation of allocation is addressed. For the sorting of Virtual Machines, Modified Best Fit Decreasing algorithm is used and Multi-Machine Job Scheduling is used while the placement process of jobs to an appropriate host. Artificial Neural Network as a classifier, has allocated jobs to the hosts. Measures, viz. Service Level Agreement violation and energy consumption are considered and fruitful results have been obtained with a 37.7 of reduction in energy consumption and 15% improvement in Service Level Agreement violation.


Electronics ◽  
2021 ◽  
Vol 10 (11) ◽  
pp. 1320
Author(s):  
Vijay Prakash ◽  
Seema Bawa ◽  
Lalit Garg

Workflow scheduling is one of the significant issues for scientific applications among virtual machine migration, database management, security, performance, fault tolerance, server consolidation, etc. In this paper, existing time-based scheduling algorithms, such as first come first serve (FCFS), min–min, max–min, and minimum completion time (MCT), along with dependency-based scheduling algorithm MaxChild have been considered. These time-based scheduling algorithms only compare the burst time of tasks. Based on the burst time, these schedulers, schedule the sub-tasks of the application on suitable virtual machines according to the scheduling criteria. During this process, not much attention was given to the proper utilization of the resources. A novel dependency and time-based scheduling algorithm is proposed that considers the parent to child (P2C) node dependencies, child to parent node dependencies, and the time of different tasks in the workflows. The proposed P2C algorithm emphasizes proper utilization of the resources and overcomes the limitations of these time-based schedulers. The scientific applications, such as CyberShake, Montage, Epigenomics, Inspiral, and SIPHT, are represented in terms of the workflow. The tasks can be represented as the nodes, and relationships between the tasks can be represented as the dependencies in the workflows. All the results have been validated by using the simulation-based environment created with the help of the WorkflowSim simulator for the cloud environment. It has been observed that the proposed approach outperforms the mentioned time and dependency-based scheduling algorithms in terms of the total execution time by efficiently utilizing the resources.


2020 ◽  
Author(s):  
João Luiz Grave Gross ◽  
Cláudio Fernando Fernando Resin Geyer

In a scenario with increasingly mobile devices connected to the Internet, data-intensive applications and energy consumption limited by battery capacity, we propose a cost minimization model for IoT devices in a Mobile Edge Computing (MEC) architecture with the main objective of reducing total energy consumption and total elapsed times from task creation to conclusion. The cost model is implemented using the TEMS (Time and Energy Minimization Scheduler) scheduling algorithm and validated with simulation. The results show that it is possible to reduce the energy consumed in the system by up to 51.61% and the total elapsed time by up to 86.65% in the simulated cases with the parameters and characteristics defined in each experiment.


Sign in / Sign up

Export Citation Format

Share Document