5 Performance Assessment of High Availability Clustered Computing using LVS-NAT

2011 ◽  
Vol 1 (1) ◽  
pp. 6
Author(s):  
Muhammad Kashif Shaikh ◽  
Muzammil Ahmad Khan ◽  
Mumtaz -ul- Imam

High availability cluster computing environment attempts to provide high availability to computing services. This paper evaluates building and investigating a highly available computing environment that provides a solution to achieve high availability. A prototype of cluster computing environment is developed in Linux environment to provide a single but highly available point of entry. The cluster of computers run web based application to provide services to HTTP users.

Sensors ◽  
2021 ◽  
Vol 21 (6) ◽  
pp. 1987 ◽  
Author(s):  
BoSun Park ◽  
Seog Chung Seo

In edge computing service, edge devices collect data from a number of embedded devices, like sensors, CCTVs (Closed-circuit Television), and so on, and communicate with application servers. Since a large portion of communication in edge computing services are conducted in wireless, the transmitted data needs to be properly encrypted. Furthermore, the application servers (resp. edge devices) are responsible for encrypting or decrypting a large amount of data from edge devices (resp. terminal devices), the cryptographic operation needs to be optimized on both server side and edge device side. Actually, the confidentiality and integrity of data are essential for secure communication. In this paper, we present two versions of security software which can be used on edge device side and server side for secure communication between them in edge computing environment. Our softwares are basically web-based application because of its universality where the softwares can be executed on any web browsers. Our softwares make use of ESTATE (Energy efficient and Single-state Tweakable block cipher based MAC-Then-Encrypt)algorithm, which is a promising candidate of NIST LWC (National Institute of Standards and Technology LightWeight Cryptography) competition and it provides not only data confidentiality but also data authentication. It also implements the ESTATE algorithm using Web Assembly for efficient use on edge devices, and optimizes the performance of the algorithm using the properties of the underlying block cipher. Several methods are applied to efficiently operate the ESTATE algorithm. We use conditional statements to XOR the extended tweak values during the operation of the ESTATE algorithm. To eliminate this unnecessary process, we use a method of expanding and storing the tweak value through pre-computation. The measured results of the ESTATE algorithm implemented with Web Assembly and the reference C/C++ ESTATE algorithm are compared. ESTATE implemented as Web Assembly is measured in web browsers Chrome, FireFox, and Microsoft Edge. For efficiency on server side, we make use of OpenCL which is parallel computing framework in order to process a number of data simultaneously. In addition, when implementing with OpenCL, using conditional statements causes performance degradation. We eliminated the conditional statement using the loop unrolling method to eliminate the performance degradation. In addition, OpenCL operates by moving the data to be encrypted to the local memory because the local memory has a high operation speed. TweAES-128 and TweAES-128-6, which have the same structure as AES algorithm, can apply the previously existing studied T-table method. In addition, the input value 16-byte is processed in parallel and calculated. In addition, since it may be vulnerable to cache-timing attack, it is safely operated by applying the previously existing studied T-table shuffling method. Our softwares cover the necessary security service from edge devices to servers in edge computing services and they can be easily used for various types of edge computing devices because they are all web-based applications.


2012 ◽  
Vol 4 (4) ◽  
pp. 68-88
Author(s):  
Chao-Tung Yang ◽  
Wen-Feng Hsieh

This paper’s objective is to implement and evaluate a high-performance computing environment by clustering idle PCs (personal computers) with diskless slave nodes on campuses to obtain the effectiveness of the largest computer potency. Two sets of Cluster platforms, BCCD and DRBL, are used to compare computing performance. It’s to prove that DRBL has better performance than BCCD in this experiment. Originally, DRBL was created to facilitate instructions for a Free Software Teaching platform. In order to achieve the purpose, DRBL is applied to the computer classroom with 32 PCs so to enable PCs to be switched manually or automatically among different OS (operating systems). The bioinformatics program, mpiBLAST, is executed smoothly in the Cluster architecture as well. From management’s view, the state of each Computation Node in Clusters is monitored by “Ganglia”, an existing Open Source. The authors gather the relevant information of CPU, Memory, and Network Load for each Computation Node in every network section. Through comparing aspects of performance, including performance of Swap and different network environment, they attempted to find out the best Cluster environment in a computer classroom at the school. Finally, HPL of HPCC is used to demonstrate cluster performance.


2012 ◽  
Vol 3 (2) ◽  
pp. 51-59 ◽  
Author(s):  
Nawsher Khan ◽  
A. Noraziah ◽  
Elrasheed I. Ismail ◽  
Mustafa Mat Deris ◽  
Tutut Herawan

Cloud computing is fundamentally altering the expectations for how and when computing, storage, and networking resources should be allocated, managed, consumed, and allow users to utilize services globally. Due to the powerful computing and storage, high availability and security, easy accessibility and adaptability, reliable scalability and interoperability, cost and time effective cloud computing is the top, needed for current fast growing business world. A client, organization or a trade that adopting emerging cloud environment can choose a well suitable infrastructure, platform, software, and a network resource, for any business, where each one has some exclusive features and advantages. The authors first develop a comprehensive classification for describing cloud computing architecture. This classification help in survey of several existing cloud computing services developed by various projects globally such as Amazon, Google, Microsoft, Sun and Force.com and by using this survey’s results the authors identified similarities and differences of the architecture approaches of cloud computing.


2014 ◽  
Vol 631-632 ◽  
pp. 1053-1056
Author(s):  
Hui Xia

The paper addressed the issues of limited resource for data optimization for efficiency, reliability, scalability and security of data in distributed, cluster systems with huge datasets. The study’s experimental results predicted that the MapReduce tool developed improved data optimization. The system exhibits undesired speedup with smaller datasets, but reasonable speedup is achieved with a larger enough datasets that complements the number of computing nodes reducing the execution time by 30% as compared to normal data mining and processing. The MapReduce tool is able to handle data growth trendily, especially with larger number of computing nodes. Scaleup gracefully grows as data and number of computing nodes increases. Security of data is guaranteed at all computing nodes since data is replicated at various nodes on the cluster system hence reliable. Our implementation of the MapReduce runs on distributed cluster computing environment of a national education web portal and is highly scalable.


2021 ◽  
Vol 27 (2) ◽  
Author(s):  
H. Hamza ◽  
A.F.D Kana ◽  
M.Y. Tanko ◽  
S. Aliyu

Cloud computing is a model that aims to deliver a reliable, customizable and scalable computing environment for end-users. Cloud computing is one of the most widely used technologies embraced by sectors and academia, offering a versatile and effective way to store and retrieve documents. The performance and efficiency of cloud computing services always depend upon the performance of the execution of user tasks submitted to the cloud system. Scheduling of user tasks plays a significant role in improving the performance of cloud services. Accordingly, many dependent task scheduling algorithms have been proposed to improve the performance of cloud services and resource utilization; however, most of the techniques for determining which task should be scheduled next are inefficient. This research provided an enhanced algorithm for scheduling dependent tasks in cloud that aims at improving the overall performance of the system. The Dependent tasks were represented as a directed acyclic graph (DAG) and the number of dependent tasks and their total running time were used as a heuristic for determining which path should be explored first. Best first search approach based on the defined heuristic was used to traverse the graph to determine which task should be scheduled next. The results of the simulation using WorkflowSim toolkit showed an average improvement of 18% and 19% on waiting time and turnaround time were achieved respectively.


Author(s):  
Hanne Mawhinney

Recent evidence of the adoption of technologically mediated systems of knowledge management as part of the trend to accountability in the Institutions of Higher Education around the world has been widely disseminated in World Bank and UNESCO reports, and the effects of assessment driven accountability on preparation programs extensively debated in scholarly communities. Less scholarly attention has been paid to the institutional effects of the technology enhanced performance assessment evidentiary demands on university programs undergoing review by national accrediting bodies. The lack of scholarly attention is addressed by presenting a case study examining the institutional dynamics of accreditation review experienced by faculty in one department that offered graduate programs leading to certification for education leaders. Drawing from institutional analysis (Scott, 2008a, 2008b) a conceptual framework is established in a discussion of conditions of enactment of the regulative, normative and cognitive facets of the institutional dynamics evident in the implementation ecology of accountability systems. The case study analysis outlines four phases of development of the essential elements of a web-based assessment system, and describes the questions raised by faculty about performance evidence, the assessment of that evidence, and the nature of measures of program outcome effectiveness. Classic theories of organizations fail to fully explain the concerns and questions that were raised by faculty. In contrast, Engeström’s (1999, 2001, 2008) theory of expansive learning grounded in Cultural Historical Activity Theory (CHAT) provides insights into faculty responses to questions raised by the criteria for program review established by the accrediting body. Artifacts of expansive learning evident in the development of a performance assessment system can be viewed as reflecting institutionalization of regulative, normative, and cognitive dimensions of the emergent evaluative state of leadership preparation around the world. Implications are suggested for understanding the development of information technology (IT) enhanced knowledge management systems (KMS).


Sign in / Sign up

Export Citation Format

Share Document