scholarly journals Communication Analysis of Distributed Programs

2006 ◽  
Vol 14 (2) ◽  
pp. 151-170 ◽  
Author(s):  
Sharon Simmons ◽  
Dennis Edwards ◽  
Phil Kearns

Capturing and examining the causal and concurrent relationships of a distributed system is essential to a wide range of distributed systems applications. Many approaches to gathering this information rely on trace files of executions. The information obtained through tracing is limited to those executions observed. We present a methodology that analyzes the source code of the distributed system. Our analysis considers each process's source code and produces a single comprehensive graph of the system's possible behaviors. The graph, termed the partial order graph (POG), uniquely represents each possible partial order of the system. Causal and concurrent relationships can be extracted relative either to a particular partial order, which is synonymous to a single execution, or to a collection of partial orders. The graph provides a means of reasoning about the system in terms of relationships that will definitely occur, may possible occur, and will never occur. Distributed assert statements provide a means to monitor distributed system executions. By constructing thePOGprior to system execution, the causality information provided by thePOGenables run-time evaluation of the assert statement without relying on traces or addition messages.

Author(s):  
Ciprian Dobre

The use of discrete-event simulators in the design and development of Large Scale Distributed Systems (LSDSs) is appealing due to their efficiency and scalability. Their core abstractions of process and event map neatly to the components and interactions of modern-day distributed systems and allow designing realistic simulation scenarios. MONARC 2, a multithreaded, process oriented simulation framework designed for modeling LSDSs, allows the realistic simulation of a wide-range of distributed system technologies, with respect to their specific components and characteristics. This chapter presents the design characteristics of the simulation model proposed in MONARC 2. It starts by first analyzing existing work, outlining the key decision points taken in the design of the MONARC’s simulation model. The model includes the necessary components to describe various actual distributed system technologies and provides the mechanisms to describe concurrent network traffic, evaluate different strategies in data replication, and analyze job scheduling procedures.


Author(s):  
Ahmad Shukri Mohd Noor ◽  
Nur Farhah Mat Zian ◽  
Fatin Nurhanani M. Shaiful Bahri

<p>Distributed systems mainly provide access to a large amount of data and computational resources through a wide range of interfaces. Besides its dynamic nature, which means that resources may enter and leave the environment at any time, many distributed systems applications will be running in an environment where faults are more likely to occur due to their ever-increasing scales and the complexity. Due to diverse faults and failures conditions, fault tolerance has become a critical element for distributed computing in order for the system to perform its function correctly even in the present of faults. Replication techniques primarily concentrate on the two fault tolerance manners precisely masking the failures as well as reconfigure the system in response. This paper presents a brief survey on different replication techniques such as Read One Write All (ROWA), Quorum Consensus (QC), Tree Quorum (TQ) Protocol, Grid Configuration (GC) Protocol, Two-Replica Distribution Techniques (TRDT), Neighbour Replica Triangular Grid (NRTG) and Neighbour Replication Distributed Techniques (NRDT). These techniques have its own redeeming features and shortcoming which forms the subject matter of this survey.</p>


2021 ◽  
Vol 40 (2) ◽  
pp. 65-69
Author(s):  
Richard Wai

Modern day cloud native applications have become broadly representative of distributed systems in the wild. However, unlike traditional distributed system models with conceptually static designs, cloud-native systems emphasize dynamic scaling and on-line iteration (CI/CD). Cloud-native systems tend to be architected around a networked collection of distinct programs ("microservices") that can be added, removed, and updated in real-time. Typically, distinct containerized programs constitute individual microservices that then communicate among the larger distributed application through heavy-weight protocols. Common communication stacks exchange JSON or XML objects over HTTP, via TCP/TLS, and incur significant overhead, particularly when using small size message sizes. Additionally, interpreted/JIT/VM-based languages such as Javascript (NodeJS/Deno), Java, and Python are dominant in modern microservice programs. These language technologies, along with the high-overhead messaging, can impose superlinear cost increases (hardware demands) on scale-out, particularly towards hyperscale and/or with latency-sensitive workloads.


2021 ◽  
Author(s):  
Johannes Keller ◽  
Johanna Fink ◽  
Norbert Klitzsch

&lt;p&gt;We present SHEMAT-Suite, a numerical code for simulating flow, heat, and mass transport in porous media that has been published as an open source code recently. The functionality of SHEMAT-Suite comprises pure forward computation, deterministic Bayesian inversion, and stochastic Monte Carlo&lt;br&gt;simulation and data assimilation. Additionally, SHEMAT-Suite features a multi-level OpenMP parallelization. Along with the source code of the software, extensive documentation and a suite of test models is provided.&lt;/p&gt;&lt;p&gt;SHEMAT-Suite has a modular structure that makes it easy for users to adapt the code to their needs. Most importantly, there is an interface for defining the functional relationship between dynamic variables and subsurface parameters. Additionally, user-defined input and output can be implemented without interfering with the core of the code. Finally, at a deeper level, linear solvers and preconditioners can be added to the code.&lt;/p&gt;&lt;p&gt;We present studies that have made use of the code's HPC capabilities. SHEMAT-Suite has been applied to large-scale groundwater models for a wide range of purposes, including studying the formation of convection cells, assessing geothermal potential below an office building, or modeling submarine groundwater discharge since the last ice age. The modular structure of SHEMAT-Suite has also led to diverse applications, such as glacier modeling, simulation of borehole heat exchangers, or Optimal Experimental Design applied to the placing of geothermal boreholes.&lt;/p&gt;&lt;p&gt;Further, we present ongoing developments for improving the performance of SHEMAT-Suite, both by refactoring the source code and by interfacing SHEMAT-Suite with up-to-date HPC software. Examples of this include interfacing SHEMAT-Suite with the Portable Data Interface (PDI) for improved data management, interfacing SHEMAT-Suite with PetSC for MPI-parallel solvers, and interfacing SHEMAT-Suite with PDAF for parallel EnKF algorithms.&lt;/p&gt;&lt;p&gt;The goal for the open source SHEMAT-Suite is to provide a rigorously tested core code for flow, heat and transport simulation, Bayesian and stochastic inversion, while at the same time enabling a wide range of scientific research through straightforward user interaction.&lt;/p&gt;


2019 ◽  
Vol 214 ◽  
pp. 05001 ◽  
Author(s):  
Stefan-Gabriel Chitic ◽  
Ben Couturier ◽  
Marco Clemencic ◽  
Joel Closier

A continuous integration system is crucial to maintain the quality of the 6 millions lines of C++ and Python source code of the LHCb software in order to ensure consistent builds of the software as well as to run the unit and integration tests. Jenkins automation server is used for this purpose. It builds and tests around 100 configurations and produces in the order of 1500 built artifacts per day which are installed on the CVMFS file system or potentially on the developers’ machines. Faced with a large and growing number of configurations built every day, and in order to ease inter-operation between the continuous integration system and the developers, we decided to put in place a flexible messaging system. As soon as the built artifacts have been produced, the distributed system allows their deployment based on the priority of the configurations. We will describe the architecture of the new system, which is based on RabbitMQ messaging system (and the pika Python client library), and uses priority queues to start the LHCb software integration tests and to drive the installation of the nightly builds on the CVMFS file system. We will also show how the introduction of an event based system can help with the communication of results to developers.


2020 ◽  
Vol 53 (4) ◽  
pp. 1060-1072 ◽  
Author(s):  
Edward L. Pang ◽  
Peter M. Larsen ◽  
Christopher A. Schuh

Resolving pseudosymmetry has long presented a challenge for electron backscatter diffraction and has been notoriously challenging in the case of tetragonal ZrO2 in particular. In this work, a method is proposed to resolve pseudosymmetry by building upon the dictionary indexing method and augmenting it with the application of global optimization to fit accurate pattern centers, clustering of the Hough-indexed orientations to focus the dictionary in orientation space and interpolation to improve the accuracy of the indexed solution. The proposed method is demonstrated to resolve pseudosymmetry with 100% accuracy in simulated patterns of tetragonal ZrO2, even with high degrees of binning and noise. The method is then used to index an experimental data set, which confirms its ability to efficiently and accurately resolve pseudosymmetry in these materials. The present method can be applied to resolve pseudosymmetry in a wide range of materials, possibly even some more challenging than tetragonal ZrO2. Source code for this implementation is available online.


SIMULATION ◽  
1970 ◽  
Vol 15 (6) ◽  
pp. 255-263 ◽  
Author(s):  
R.E. Goodson

Infinite product expansions for the transcendental terms in the transfer functions for linear distributed systems are developed. Simulation of the dynamic re sponse of such systems is indicated, using the product expansion. Comparisons are made between the classi cal eigenvalue and product expansion approximations. It is concluded that the product expansion is an ex tremum transient-amplitude-preserving approximation based on the correct characteristic roots.


2014 ◽  
Vol 25 (02) ◽  
pp. 125-139 ◽  
Author(s):  
JHENG-CHENG CHEN ◽  
CHIA-JUI LAI ◽  
CHANG-HSIUNG TSAI

Problem diagnosis in large distributed computer systems and networks is a challenging task that requires fast and accurate inferences from huge volumes of data. In this paper, the PMC diagnostic model is considered, based on the diagnostic approach of end-to-end probing technology. A probe is a test transaction whose outcome depends on some of the system's components; diagnosis is performed by selecting appropriate probes and analyzing the results. In the PMC model, every computer can execute a probe to test a dedicated system's components. Furthermore, any test result reported by a faulty probe station is unreliable and the test result reported by fault-free probe station is always correct. The aim of the diagnosis is to locate all faulty components in the system based on collection of the test results. A dual-cube DC(n) is an (n + 1)-regular spanning subgraph of a (2n + 1)-dimensional hypercube. It uses n-dimensional hypercubes as building blocks and returns the main desirable properties of the hypercube so that it is suitable as a topology for distributed systems. In this paper, we first show that the diagnosability of DC(n) is n + 1 and then show that adaptive diagnosis is possible using at most 22n+1 + n tests for a 22n+1-node distributed system modeled by dual-cubes DC(n) in which at most n + 1 processes are faulty. Furthermore, we propose an adaptive diagnostic algorithm for the DC(n) and show that it diagnoses the DC(n) in three testing rounds and at most 22n+1 + O(n3) tests, where each node is scheduled for at most one test in each round.


2010 ◽  
Vol 164 ◽  
pp. 183-188
Author(s):  
Cezary Orlikowski ◽  
Rafał Hein

This paper presents a uniform, port-based approach for modeling of both lumped and distributed parameter systems. Port-based model of the distributed system has been defined by application of bond graph methodology and distributed transfer function method (DTFM). The proposed approach combines versatility of port-based modeling and accuracy of distributed transfer function method. A concise representation of lumped-distributed systems has been obtained. The proposed method of modeling enables to formulate input data for computer analysis by application of DTFM.


Sign in / Sign up

Export Citation Format

Share Document