scholarly journals MICROSERVICE ARCHITECTURE FOR BUILDING HIGH-AVAILABILITY DISTRIBUTED AUTOMATED COMPUTING SYSTEM IN A CLOUD INFRASTRUCTURE

Author(s):  
Denis Zolotariov

The article is devoted to the research and development of a highly available distributed automated computing system by iterative algorithms based on the microservice architecture in a cloud infrastructure. The subject of the research is the practical foundations of building high-availability automated computing systems based on microservice architecture in a cloud-based distributed infrastructure. The purpose of the article is to develop and to substantiate practical recommendations for the formation of the infrastructure of a high-availability automated computing system based on the microservice architecture, the choice of its constituent elements and their components. The task of the work: to identify the necessary structural elements of a microservice automated computing system and to analyze the constituent components and functional load for each of them, set specific tasks for building each of them and justify the choice of tools for their creation. In the course of the research, methods of system analysis were used to decompose a complex system into elements and each element into functional components, and tools: information technologies Apache Kafka, Kafkacat, Wolfram Mathematica, nginx, Lumen, Telegram, Dropbox, and MySQL. As a result of the study, it was found that the system infrastructure should consist of: fault-tolerant interservice transport, a high-availability computing microservice, and communication microservices with end customers, which save or process the results. For each of them, recommendations are provided regarding the formation and selection of implementation tools. According to the recommendations, one variant of implementation of such system has been developed, the principles of its operation are shown and the results are presented. It has been proven that when using a Kafka queue it is efficient to publish batches of results rather than one at a time, which results to significant overhead on queue servers and data latency for its clients. Recommendations are given on the implementation of the CI/CD system to build a continuous cycle of adding and improving microservices. Conclusions. Practical foundations have been developed for the implementation of high availability distributed automated computing systems based on microservice architecture in a cloud infrastructure. The flexibility in processing the results of such a system is shown due to the possibility of adding microservices and using third-party analytical applications that support connection to the Kafka queue. The economic benefit of using the described system is shown. Future ways of its improvement are given.

1974 ◽  
Vol 3 (32) ◽  
Author(s):  
L. Phillip Caillouet ◽  
Bruce D. Shriver

This paper offers an introduction to a research effort in fault tolerant computer architecture which has been organized at the University of Southwestern Louisiana (USL). It is intended as an overview of several topics which have been isolated for study, and as an indication of preliminary undertakings with regards to one particular topic. This first area of concentration lnvolves the systematic design of fault tolerant computing systems via a multi-level approach. Efforts are being initiated also in the areas of diagnosis of microprogrammable processors via firmware, fault data management across levels of virtual machines, development of a methodology for realizing a firmware hardcore on a variety of hosts, and delineation of a minimal set of resources for the design of a practical host for a multi-level fault tolerant computing system. The research is being conducted under the auspices of Project Beta at USL.


Author(s):  
Denis Zolotariov

The article is devoted to the development and substantiation of practical recommendations regarding the formation of a mechanism for deploying a software environment for creating and executing microservices in a rapidly changing technological stack. The subject of the research is the basics of building a system for automated deployment of a software environment for the development and execution of microservices. The purpose of the article is to develop and substantiate practical recommendations for the formation of a mechanism for deploying a software environment for creating and executing microservices in a rapidly changing technological stack. The task of the work: to determine the necessary elements of the deployment mechanism of the software environment and provide an analysis of the functional load for each of them, set specific tasks that must be solved when building each of them, propose and justify the choice of tools for their solution. In the course of the study, the methods of system analysis were used to decompose a complex system into elements and each element into functional components. As of the study, it was established that such a mechanism should consist of the following elements: a universal server initialization a result subsystem for any technological stack and a software environment deployment subsystem for developing or executing an application of a certain type on a certain technological stack. Each element is described in detail, its functional load is shown and its role in the overall system is substantiated. It is shown that such a standardized approach to the deployment of the development and runtime environment allows, among other things, to solve the problem of operating microservices in a tested environment. Conclusions. Practical recommendations for the formation of a mechanism for deploying a software environment for creating and executing microservices in a rapidly changing technological stack have been developed and substantiated. This mechanism is automated. It shows its flexibility and versatility in relation to programming languages and other features of the software environment. It is pointed out that when implemented in the shell language, bash does not need any third-party applications for its work. The economic benefit of using the proposed mechanism is shown. The ways of its improvement are shown.


Author(s):  
Rosangela Melo ◽  
Vicente Sobrinho ◽  
Ivanildo Filho ◽  
Fábio Feliciano ◽  
Paulo Maciel

Cloud computing offer IT services to the users worldwide based on pay-as-you-go model. Furthermore, architectures and services that are provided by cloud computing system must have high availability, scalability, security and furthermore be fault tolerant. Plan these environments is not an easy task, it is necessary to ensure that undesirable situations or errors do not occur or can be minimized. The use of Sensitivity Analysis combined with the use of modeling hierarchy and the study of models for representing redundancy mechanisms arise in order to investigate the possible changes that these systems suffer and identify their shortcomings and propose improvements solutions for the planning of these systems. The use of redundancy techniques has been used to improve availability. In this work the redundancy mechanisms is a solution to improve the performance of environments in cloud computing.


2012 ◽  
Vol 4 (1) ◽  
pp. 37-51 ◽  
Author(s):  
Hodjat Hamidi ◽  
Abbas Vafaei ◽  
Seyed Amir Hassan Monadjemi

In this paper, the authors present a new approach to algorithm based fault tolerance (ABFT) for High Performance computing system. The Algorithm Based Fault Tolerance approach transforms a system that does not tolerate a specific type of fault, called the fault-intolerant system, to a system that provides a specific level of fault tolerance, namely recovery. The ABFT techniques that detect errors rely on the comparison of parity values computed in two ways, the parallel processing of input parity values produce output parity values comparable with parity values regenerated from the original processed outputs, can apply convolution codes for the redundancy. This method is a new approach to concurrent error correction in fault-tolerant computing systems. This paper proposes a novel computing paradigm to provide fault tolerance for numerical algorithms. The authors also present, implement, and evaluate early detection in ABFT.


Life ◽  
2021 ◽  
Vol 11 (4) ◽  
pp. 310
Author(s):  
Shih-Chia Chang ◽  
Ming-Tsang Lu ◽  
Tzu-Hui Pan ◽  
Chiao-Shan Chen

Although the electronic health (e-health) cloud computing system is a promising innovation, its adoption in the healthcare industry has been slow. This study investigated the adoption of e-health cloud computing systems in the healthcare industry and considered security functions, management, cloud service delivery, and cloud software for e-health cloud computing systems. Although numerous studies have determined factors affecting e-health cloud computing systems, few comprehensive reviews of factors and their relations have been conducted. Therefore, this study investigated the relations between the factors affecting e-health cloud computing systems by using a multiple criteria decision-making technique, in which decision-making trial and evaluation laboratory (DEMATEL), DANP (DEMATEL-based Analytic Network Process), and modified VIKOR (VlseKriterijumska Optimizacija I Kompromisno Resenje) approaches were combined. The intended level of adoption of an e-health cloud computing system could be determined by using the proposed approach. The results of a case study performed on the Taiwanese healthcare industry indicated that the cloud management function must be primarily enhanced and that cost effectiveness is the most significant factor in the adoption of e-health cloud computing. This result is valuable for allocating resources to decrease performance gaps in the Taiwanese healthcare industry.


2021 ◽  
Vol 11 (12) ◽  
pp. 5458
Author(s):  
Sangjun Kim ◽  
Kyung-Joon Park

A cyber-physical system (CPS) is the integration of a physical system into the real world and control applications in a computing system, interacting through a communications network. Network technology connecting physical systems and computing systems enables the simultaneous control of many physical systems and provides intelligent applications for them. However, enhancing connectivity leads to extended attack vectors in which attackers can trespass on the network and launch cyber-physical attacks, remotely disrupting the CPS. Therefore, extensive studies into cyber-physical security are being conducted in various domains, such as physical, network, and computing systems. Moreover, large-scale and complex CPSs make it difficult to analyze and detect cyber-physical attacks, and thus, machine learning (ML) techniques have recently been adopted for cyber-physical security. In this survey, we provide an extensive review of the threats and ML-based security designs for CPSs. First, we present a CPS structure that classifies the functions of the CPS into three layers: the physical system, the network, and software applications. Then, we discuss the taxonomy of cyber-physical attacks on each layer, and in particular, we analyze attacks based on the dynamics of the physical system. We review existing studies on detecting cyber-physical attacks with various ML techniques from the perspectives of the physical system, the network, and the computing system. Furthermore, we discuss future research directions for ML-based cyber-physical security research in the context of real-time constraints, resiliency, and dataset generation to learn about the possible attacks.


Author(s):  
VanDung Nguyen ◽  
Tran Trong Khanh ◽  
Tri D. T. Nguyen ◽  
Choong Seon Hong ◽  
Eui-Nam Huh

AbstractIn the Internet of Things (IoT) era, the capacity-limited Internet and uncontrollable service delays for various new applications, such as video streaming analysis and augmented reality, are challenges. Cloud computing systems, also known as a solution that offloads energy-consuming computation of IoT applications to a cloud server, cannot meet the delay-sensitive and context-aware service requirements. To address this issue, an edge computing system provides timely and context-aware services by bringing the computations and storage closer to the user. The dynamic flow of requests that can be efficiently processed is a significant challenge for edge and cloud computing systems. To improve the performance of IoT systems, the mobile edge orchestrator (MEO), which is an application placement controller, was designed by integrating end mobile devices with edge and cloud computing systems. In this paper, we propose a flexible computation offloading method in a fuzzy-based MEO for IoT applications in order to improve the efficiency in computational resource management. Considering the network, computation resources, and task requirements, a fuzzy-based MEO allows edge workload orchestration actions to decide whether to offload a mobile user to local edge, neighboring edge, or cloud servers. Additionally, increasing packet sizes will affect the failed-task ratio when the number of mobile devices increases. To reduce failed tasks because of transmission collisions and to improve service times for time-critical tasks, we define a new input crisp value, and a new output decision for a fuzzy-based MEO. Using the EdgeCloudSim simulator, we evaluate our proposal with four benchmark algorithms in augmented reality, healthcare, compute-intensive, and infotainment applications. Simulation results show that our proposal provides better results in terms of WLAN delay, service times, the number of failed tasks, and VM utilization.


Sign in / Sign up

Export Citation Format

Share Document