SMART GRID AND CLOUD COMPUTING : MINIMIZING POWER CONSUMPTION AND UTILITY EXPENDITURE IN DATA CENTERS SU MITKUMAR BOSE , M ICH A EL SA LSBU RG, SCO T T BROCK , AND RONA L D SK EOCH

2016 ◽  
pp. 74-101 ◽  
Author(s):  
Deepika T. ◽  
Prakash P.

The flourishing development of the cloud computing paradigm provides several services in the industrial business world. Power consumption by cloud data centers is one of the crucial issues for service providers in the domain of cloud computing. Pursuant to the rapid technology enhancements in cloud environments and data centers augmentations, power utilization in data centers is expected to grow unabated. A diverse set of numerous connected devices, engaged with the ubiquitous cloud, results in unprecedented power utilization by the data centers, accompanied by increased carbon footprints. Nearly a million physical machines (PM) are running all over the data centers, along with (5 – 6) million virtual machines (VM). In the next five years, the power needs of this domain are expected to spiral up to 5% of global power production. The virtual machine power consumption reduction impacts the diminishing of the PM’s power, however further changing in power consumption of data center year by year, to aid the cloud vendors using prediction methods. The sudden fluctuation in power utilization will cause power outage in the cloud data centers. This paper aims to forecast the VM power consumption with the help of regressive predictive analysis, one of the Machine Learning (ML) techniques. The potency of this approach to make better predictions of future value, using Multi-layer Perceptron (MLP) regressor which provides 91% of accuracy during the prediction process.


2021 ◽  
Vol 39 (1B) ◽  
pp. 203-208
Author(s):  
Haider A. Ghanem ◽  
Rana F. Ghani ◽  
Maha J. Abbas

Data centers are the main nerve of the Internet because of its hosting, storage, cloud computing and other services. All these services require a lot of work and resources, such as energy and cooling. The main problem is how to improve the work of data centers through increased resource utilization by using virtual host simulations and exploiting all server resources. In this paper, we have considered memory resources, where Virtual machines were distributed to hosts after comparing the virtual machines with the host from where the memory and putting the virtual machine on the appropriate host, this will reduce the host machines in the data centers and this will improve the performance of the data centers, in terms of power consumption and the number of servers used and cost.


Author(s):  
Mahendra Kumar Gourisaria ◽  
S. S. Patra ◽  
P. M. Khilar

<p>Cloud computing is an emerging field of computation. As the data centers consume large amount of power, it increases the system overheads as well as the carbon dioxide emission increases drastically. The main aim is to maximize the resource utilization by minimizing the power consumption. However, the greatest usages of resources does not mean that there has been a right use of energy.  Various resources which are idle, also consumes a significant amount of energy. So we have to keep minimum resources idle. Current studies have shown that the power consumption due to unused computing resources is nearly 1 to 20%. So, the unused resources have been assigned with some of the tasks to utilize the unused period. In the present paper, it has been suggested that the energy saving with task consolidation which has been saved the energy by minimizing the number of idle resources in a cloud computing environment. It has been achieved far-reaching experiments to quantify the performance of the proposed algorithm. The same has also been compared with the FCFSMaxUtil and Energy aware Task Consolidation (ETC) algorithm. The outcomes have shown that the suggested algorithm surpass the FCFSMaxUtil and ETC algorithm in terms of the CPU utilization and energy consumption.</p>


2019 ◽  
Vol 5 ◽  
pp. e211
Author(s):  
Hadi Khani ◽  
Hamed Khanmirza

Cloud computing technology has been a game changer in recent years. Cloud computing providers promise cost-effective and on-demand resource computing for their users. Cloud computing providers are running the workloads of users as virtual machines (VMs) in a large-scale data center consisting a few thousands physical servers. Cloud data centers face highly dynamic workloads varying over time and many short tasks that demand quick resource management decisions. These data centers are large scale and the behavior of workload is unpredictable. The incoming VM must be assigned onto the proper physical machine (PM) in order to keep a balance between power consumption and quality of service. The scale and agility of cloud computing data centers are unprecedented so the previous approaches are fruitless. We suggest an analytical model for cloud computing data centers when the number of PMs in the data center is large. In particular, we focus on the assignment of VM onto PMs regardless of their current load. For exponential VM arrival with general distribution sojourn time, the mean power consumption is calculated. Then, we show the minimum power consumption under quality of service constraint will be achieved with randomize assignment of incoming VMs onto PMs. Extensive simulation supports the validity of our analytical model.


Energies ◽  
2018 ◽  
Vol 11 (12) ◽  
pp. 3345 ◽  
Author(s):  
Rasool Bukhsh ◽  
Nadeem Javaid ◽  
Zahoor Ali Khan ◽  
Farruh Ishmanov ◽  
Muhammad Afzal ◽  
...  

The integration of the smart grid with the cloud computing environment promises to develop an improved energy-management system for utility and consumers. New applications and services are being developed which generate huge requests to be processed in the cloud. As smart grids can dynamically be operated according to consumer requests (data), so, they can be called Data-Driven Smart Grids. Fog computing as an extension of cloud computing helps to mitigate the load on cloud data centers. This paper presents a cloud–fog-based system model to reduce Response Time (RT) and Processing Time (PT). The load of requests from end devices is processed in fog data centers. The selection of potential data centers and efficient allocation of requests on Virtual Machines (VMs) optimize the RT and PT. A New Service Broker Policy (NSBP) is proposed for the selection of a potential data center. The load-balancing algorithm, a hybrid of Particle Swarm Optimization and Simulated Annealing (PSO-SA), is proposed for the efficient allocation of requests on VMs in the potential data center. In the proposed system model, Micro-Grids (MGs) are placed near the fogs for uninterrupted and cheap power supply to clusters of residential buildings. The simulation results show the supremacy of NSBP and PSO-SA over their counterparts.


2012 ◽  
Vol 13 (03n04) ◽  
pp. 1250010
Author(s):  
SHAOSHAN LIU ◽  
WON W. RO ◽  
CHEN LIU ◽  
ALFREDO CRISTOBAL-SALAS ◽  
CHRISTOPHE CÉRIN ◽  
...  

The computer industry is moving towards two extremes: extremely high-performance high-throughput cloud computing, and low-power mobile computing. Cloud computing, while providing high performance, is very costly. Google and Microsoft Bing spend billions of dollars each year to maintain their server farms, mainly due to the high power bills. On the other hand, mobile computing is under a very tight energy budget, but yet the end users demand ever increasing performance on these devices. This trend indicates that conventional architectures are not able to deliver high-performance and low power consumption at the same time, and we need a new architecture model to address the needs of both extremes. In this paper, we thus introduce our Extremely Heterogeneous Architecture (EHA) project: EHA is a novel architecture that incorporates both general-purpose and specialized cores on the same chip. The general-purpose cores take care of generic control and computation. On the other hand, the specialized cores, including GPU, hard accelerators (ASIC accelerators), and soft accelerators (FPGAs), are designed for accelerating frequently used or heavy weight applications. When acceleration is not needed, the specialized cores are turned off to reduce power consumption. We demonstrate that EHA is able to improve performance through acceleration, and at the same time reduce power consumption. Since EHA is a heterogeneous architecture, it is suitable for accelerating heterogeneous workloads on the same chip. For example, data centers and clouds provide many services, including media streaming, searching, indexing, scientific computations. The ultimate goal of the EHA project is two-fold: first, to design a chip that is able to run different cloud services on it, and through this design, we would be able to greatly reduce the cost, both recurring and non-recurring, of data centers\clouds; second, to design a light-weight EHA that runs on mobile devices, providing end users with improved experience even under tight battery budget constraints.


Author(s):  
Ahmed Abdul Hassan Al-Fatlawi ◽  
Seifedine Kadry

Green Cloud computing is envisioned to achieve not only efficient processing and utilization of computing but also to minimize energy consumption. This is essential for ensuring that the future growth of Cloud computing is sustainable. Otherwise, Cloud computing with increasingly pervasive client devices interacting with data centers will cause an enormous escalation of energy usage. To address this problem, data center resources need to be managed in an energy-efficient manner to drive Green Cloud computing. The management of power consumption in data centers has led to a number of substantial improvements in energy efficiency. Techniques such as ON/OFF mode on server of data centers improve the energy efficiency of Cloud computing. In this chapter, the authors present how to calculate power consumption in Cloud computing and how power consumption in a data center can be reduced when its storage is used in a way that decreases the time needed to access it.


Author(s):  
Mahendra Kumar Gourisaria ◽  
S. S. Patra ◽  
P. M. Khilar

<p>Cloud computing is an emerging field of computation. As the data centers consume large amount of power, it increases the system overheads as well as the carbon dioxide emission increases drastically. The main aim is to maximize the resource utilization by minimizing the power consumption. However, the greatest usages of resources does not mean that there has been a right use of energy.  Various resources which are idle, also consumes a significant amount of energy. So we have to keep minimum resources idle. Current studies have shown that the power consumption due to unused computing resources is nearly 1 to 20%. So, the unused resources have been assigned with some of the tasks to utilize the unused period. In the present paper, it has been suggested that the energy saving with task consolidation which has been saved the energy by minimizing the number of idle resources in a cloud computing environment. It has been achieved far-reaching experiments to quantify the performance of the proposed algorithm. The same has also been compared with the FCFSMaxUtil and Energy aware Task Consolidation (ETC) algorithm. The outcomes have shown that the suggested algorithm surpass the FCFSMaxUtil and ETC algorithm in terms of the CPU utilization and energy consumption.</p>


Sensors ◽  
2020 ◽  
Vol 20 (13) ◽  
pp. 3635 ◽  
Author(s):  
Guoming Zhang ◽  
Xiaoyu Ji ◽  
Yanjie Li ◽  
Wenyuan Xu

As a critical component in the smart grid, the Distribution Terminal Unit (DTU) dynamically adjusts the running status of the entire smart grid based on the collected electrical parameters to ensure the safe and stable operation of the smart grid. However, as a real-time embedded device, DTU has not only resource constraints but also specific requirements on real-time performance, thus, the traditional anomaly detection method cannot be deployed. To detect the tamper of the program running on DTU, we proposed a power-based non-intrusive condition monitoring method that collects and analyzes the power consumption of DTU using power sensors and machine learning (ML) techniques, the feasibility of this approach is that the power consumption is closely related to the executing code in CPUs, that is when the execution code is tampered with, the power consumption changes accordingly. To validate this idea, we set up a testbed based on DTU and simulated four types of imperceptible attacks that change the code running in ARM and DSP processors, respectively. We generate representative features and select lightweight ML algorithms to detect these attacks. We finally implemented the detection system on the windows and ubuntu platform and validated its effectiveness. The results show that the detection accuracy is up to 99.98% in a non-intrusive and lightweight way.


Sign in / Sign up

Export Citation Format

Share Document