elastic computing
Recently Published Documents


TOTAL DOCUMENTS

48
(FIVE YEARS 20)

H-INDEX

8
(FIVE YEARS 2)

2021 ◽  
Author(s):  
Chunming Rong ◽  
Jiahui Geng ◽  
Thomas J. Hacker ◽  
Haakon Bryhni ◽  
Martin G. Jaatun

Abstract Modern information systems are built fron a complex composition of networks, infrastructure, devices, services, and applications, interconnected by data flows that are often private and financially sensitive. The 5G networks, which can create hyperlocalized services, have highlighted many of the deficiencies of current practices in use today to create and operate information systems. Emerging cloud computing techniques, such as Infrastructure-as-Code (IaC) and elastic computing, o↵er a path for a future re-imagining of how we create, deploy, secure, operate, and retire information systems. In this paper, we articulate the position that a comprehensive new approach is needed for all OSI layers from layer 2 up to applications that are built on underlying principles that include reproducibility, continuous integration/continuous delivery, auditability, and versioning. There are obvious needs to redesign and optimize the protocols from the network layer to the application layer. Our vision seeks to augment existing Cloud Computing and Networking solutions with support for multiple cloud infrastructures and seamless integration of cloud-based microservices. To address these issues, we propose an approach named Open Infrastructure as Code (OpenIaC), which is an attempt to provide a common open forum to integrate and build on advances in cloud computing and blockchain to address the needs of modern information architectures. The main mission of our OpenIaC approach is to provide services based on the principles of Zero Trust Architecture (ZTA) among the federation of connected resources based on Decentralized Identity (DID). Our objectives include the creation of an open-source hub with fine-grained access control for an open and connected infrastructure of shared resources (sensing, storage, computing, 3D printing, etc.) managed by blockchains and federations. Our proposed approach has the potential to provide a path for developing new platforms, business models, and a modernized information ecosystem necessary for 5G networks.


Author(s):  
Mr. N. B. Kadu

With increasing network virtualization, data centre's workloads are modified in depth to serve various service-oriented applications, often defined by a time-bound service response, which, in turn, places a heavy demand on data center networks. Network virtualization in computing is the technique of integrating network resources and network functions in hardware and software into one virtual network, the software-based administration entity. Number of people ask for the server simultaneously, thereby slowing down the service.It is so costly to buy a new server that we developed a virtual system by creating a virtual system. With a trend to increase the number of cloud apps in the datacenter. There are numerous physical machines (PMs) linked via switches in the datacenter. Hardware PM resources for adaptable and elastic computing capabilities are usually shared via virtualization technology. Usually a cloud application is implemented in a virtual cluster that includes many virtual machines which occupy PM resources on request.


Author(s):  
Shashikant Ilager ◽  
Vlado Stankovski ◽  
Shrideep Pallickarar ◽  
Rajkumar Buyya

Sensors ◽  
2021 ◽  
Vol 21 (9) ◽  
pp. 2971
Author(s):  
Wei Huang ◽  
Jianzhong Zhou ◽  
Dongying Zhang

Remotely-sensed satellite image fusion is indispensable for the generation of long-term gap-free Earth observation data. While cloud computing (CC) provides the big picture for RS big data (RSBD), the fundamental question of the efficient fusion of RSBD on CC platforms has not yet been settled. To this end, we propose a lightweight cloud-native framework for the elastic processing of RSBD in this study. With the scaling mechanisms provided by both the Infrastructure as a Service (IaaS) and Platform as a Services (PaaS) of CC, the Spark-on-Kubernetes operator model running in the framework can enhance the efficiency of Spark-based algorithms without considering bottlenecks such as task latency caused by an unbalanced workload, and can ease the burden to tune the performance parameters for their parallel algorithms. Internally, we propose a task scheduling mechanism (TSM) to dynamically change the Spark executor pods’ affinities to the computing hosts. The TSM learns the workload of a computing host. Learning from the ratio between the number of completed and failed tasks on a computing host, the TSM dispatches Spark executor pods to newer and less-overwhelmed computing hosts. In order to illustrate the advantage, we implement a parallel enhanced spatial and temporal adaptive reflectance fusion model (PESTARFM) to enable the efficient fusion of big RS images with a Spark aggregation function. We construct an OpenStack cloud computing environment to test the usability of the framework. According to the experiments, TSM can improve the performance of the PESTARFM using only PaaS scaling to about 11.7%. When using both the IaaS and PaaS scaling, the maximum performance gain with the TSM can be even greater than 13.6%. The fusion of such big Sentinel and PlanetScope images requires less than 4 min in the experimental environment.


2021 ◽  
pp. 676-689
Author(s):  
Damiano Perri ◽  
Marco Simonetti ◽  
Sergio Tasso ◽  
Federico Ragni ◽  
Osvaldo Gervasi

Sign in / Sign up

Export Citation Format

Share Document