energy efficient computing
Recently Published Documents


TOTAL DOCUMENTS

158
(FIVE YEARS 47)

H-INDEX

16
(FIVE YEARS 2)

Author(s):  
Alireza Souri ◽  
Vincenzo Piuri ◽  
Mohammad Shojafar ◽  
Eyhab Al‐Masri ◽  
Saru Kumari

Author(s):  
Kethavath Prem Kumar ◽  
◽  
Thirumalaisamy Ragunathan ◽  
Devara Vasumathi ◽  
◽  
...  

Cloud Computing is rapidly being utilized to operate informational technological services by outstanding technologies for a variety of benefits, including dynamically improved resources planning and a new service delivery method. The Cloud computing process is occurred by allowing the client devices for data access through the internet from a remote server, computers, and the databases. An internet connection is linked among the front end users such as client device, network, browser, and software application with the back end that constitutes of servers, computers, and database. For satisfying the demands of the Service Level Agreement (SLA), providers of cloud service should reduce the usage of energy. Capacity reservations oriented system is available by clouds’ providers to permit users for customizing Virtual Machines (VMs) having specified age and geographic resources, reduces the amount to be paid for cloud services. To overcome the aforementioned issue, an Improved Spider Monkey Optimization (ISMO) approach is proposed for cloud center optimization. The VM consolidation architecture based on the proposed ISMO algorithm decreases energy usage while attempting to prevent Service Level Agreement breaches. The accessibility of hosts or virtual machines (VMs) for task performance is measured by fitness. If the number of tasks to be handled increases the hosts of VMs available at right state. The proposed VM consolidation architecture decreases energy usage while also attempting to prevent Service Level Agreement breaches and also provide energy-efficient computing in data centers. The proposed approach may be utilized to provide energy-efficient computing in data centers. The energy efficiency of the proposed ISMO method is achieved 28266 whereas, the existing algorithm showed an energy efficiency of 6009 and 10001.


2021 ◽  
Author(s):  
Rob Aitken ◽  
Yorie Nakahira ◽  
John Strachan ◽  
Kirk Bresniker ◽  
Ian Young ◽  
...  

2021 ◽  
Vol 20 (5) ◽  
pp. 1-21
Author(s):  
Vasileios Leon ◽  
Theodora Paparouni ◽  
Evangelos Petrongonas ◽  
Dimitrios Soudris ◽  
Kiamal Pekmestzi

Approximate computing has emerged as a promising design alternative for delivering power-efficient systems and circuits by exploiting the inherent error resiliency of numerous applications. The current article aims to tackle the increased hardware cost of floating-point multiplication units, which prohibits their usage in embedded computing. We introduce AFMU (Approximate Floating-point MUltiplier), an area/power-efficient family of multipliers, which apply two approximation techniques in the resource-hungry mantissa multiplication and can be seamlessly extended to support dynamic configuration of the approximation levels via gating signals. AFMU offers large accuracy configuration margins, provides negligible logic overhead for dynamic configuration, and detects unexpected results that may arise due to the approximations. Our evaluation shows that AFMU delivers energy gains in the range 3.6%–53.5% for half-precision and 37.2%–82.4% for single-precision, in exchange for mean relative error around 0.05%–3.33% and 0.01%–2.20%, respectively. In comparison with state-of-the-art multipliers, AFMU exhibits up to 4–6× smaller error on average while delivering more energy-efficient computing. The evaluation in image processing shows that AFMU provides sufficient quality of service, i.e., more than 50 db PSNR and near 1 SSIM values, and up to 57.4% power reduction. When used in floating-point CNNs, the accuracy loss is small (or zero), i.e., up to 5.4% for MNIST and CIFAR-10, in exchange for up to 63.8% power gain.


2021 ◽  
pp. 2000275
Author(s):  
Chenghao Feng ◽  
Zhoufeng Ying ◽  
Zheng Zhao ◽  
Jiaqi Gu ◽  
David Z. Pan ◽  
...  

2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Daniele D’Agostino ◽  
Ivan Merelli ◽  
Marco Aldinucci ◽  
Daniele Cesini

Energy consumption is one of the major issues in today’s computer science, and an increasing number of scientific communities are interested in evaluating the tradeoff between time-to-solution and energy-to-solution. Despite, in the last two decades, computing which revolved around centralized computing infrastructures, such as supercomputing and data centers, the wide adoption of the Internet of Things (IoT) paradigm is currently inverting this trend due to the huge amount of data it generates, pushing computing power back to places where the data are generated—the so-called fog/edge computing. This shift towards a decentralized model requires an equivalent change in the software engineering paradigms, development environments, hardware tools, languages, and computation models for scientific programming because the local computational capabilities are typically limited and require a careful evaluation of power consumption. This paper aims to present how these concepts can be actually implemented in scientific software by presenting the state of the art of powerful, less power-hungry processors from one side and energy-aware tools and techniques from the other one.


2021 ◽  
Vol 64 (6) ◽  
Author(s):  
Liang Chang ◽  
Chenglong Li ◽  
Zhaomin Zhang ◽  
Jianbiao Xiao ◽  
Qingsong Liu ◽  
...  

2021 ◽  
Vol 11 (2) ◽  
pp. 14
Author(s):  
Jennifer Hasler ◽  
Eric Black

Physical computing unifies real value computing including analog, neuromorphic, optical, and quantum computing. Many real-valued techniques show improvements in energy efficiency, enable smaller area per computation, and potentially improve algorithm scaling. These physical computing techniques suffer from not having a strong computational theory to guide application development in contrast to digital computation’s deep theoretical grounding in application development. We consider the possibility of a real-valued Turing machine model, the potential computational and algorithmic opportunities of these techniques, the implications for implementation applications, and the computational complexity space arising from this model. These techniques have shown promise in increasing energy efficiency, enabling smaller area per computation, and potentially improving algorithm scaling.


Sign in / Sign up

Export Citation Format

Share Document