scholarly journals On the compatibility of topologies of parallel tasks and computing systems

2021 ◽  
Vol 1864 (1) ◽  
pp. 012103
Author(s):  
A.F. Zadorozhny ◽  
V.A. Melent’ev
Author(s):  
A. F. Zadorozhny ◽  
V. A. Melent’ev

The aspects of topological compatibility of parallel computing systems and tasks are investigated in the present contribution. Based on the original topological model of parallel computations and on the unconventional graph description by its projections, the introduction of appropriate indexes is proposed and elucidated. On the example of hypercubic computing system (CS) and tasks with ring and star information topologies, we demonstrate the determination of indexes and their use in a comparative analysis of the applicability of interconnect with a given topology to solve the tasks with the same and different types of information topologies.


2020 ◽  
Vol 10 (18) ◽  
pp. 6611
Author(s):  
Apolinar Velarde Martinez

The problem of scheduling parallel tasks graphs (PTGs) represented by directed acyclic graphs (DAGs) in heterogeneous distributed computing systems (HDCSs) is considered an nondeterministic polynomial time (NP) problem due to the diversity of characteristics and parameters, generally opposed, intended to be optimized. The PTGs are scheduled by a scheduler that determines the best location for the sub-tasks that constitute the PTGs and is responsible for allocating the resources of the HDCS to the sub-tasks of the PTGs. To optimize scheduling and allocations, the scheduler extracts characteristics from the internal structure of the PTGs. The prevailing characteristic in existing research is the critical path (CP), which is limited to providing execution paths of PTGs; considering this limitation, we extend the array method proposed in Velarde, which extracts two additional characteristics to the CP: the layering and the density of the graph for scheduling. These characteristics are represented as integer values of the PTGs to be scheduled; the values obtained from the characteristics are stored in arrays representing populations that are evaluated with the heuristic univariate marginal distribution algorithm (UMDA) and in terms of comparison with the genetic algorithm. With the best allocations produced by the algorithms, two performance parameters are evaluated: makespan and waiting time. The results indicate that when more PTGs characteristics are considered, resource allocations are optimized, and scheduling times are reduced. The results obtained with the heuristic algorithms show that UMDA provides shorter scheduling and allocation times compared with the genetic algorithm; UMDA widely distributes the sub-tasks in the clusters, whereas the genetic algorithm compacts the assignments of the PTGs in the clusters with a longer convergence time that translates into longer scheduling and allocation times. Extensive explanations of these conclusions are provided in this work, based on the conducted experiments.


2019 ◽  
Author(s):  
Jaime Freire de Souza ◽  
Hermes Senger ◽  
Fabricio A. B. Silva

Bag-of-Tasks (BoT) applications are parallel applications composed of independent (i.e., embarrassingly parallel) tasks, which do not communicate with each other, may depend upon one or more input files, and can be executed in any order. BoT applications are very frequent in several scientific areas, and it is the ideal application class for execution on large distributed computing systems composed of hundreds to many thousands of computational resources. This paper focusses on the scalability of BoT applications running on large heterogeneous distributed computing systems organized as a master-slave platform. The results demonstrate that heterogeneous master-slave platforms can achieve higher scalability than homogeneous platforms for the execution of BoT applications, when the computational power of individual nodes in the homogeneous platform is fixed. However, when individual nodes of the homogeneous platform can scale-up, experiments show that master-slave platforms can achieve near linear speedups.


2004 ◽  
Vol 14 (01) ◽  
pp. 23-32 ◽  
Author(s):  
Jacek Błażewicz ◽  
Adrian Moret-Salvador ◽  
Rafał Walkowiak

A tabu search based approach is studied as a method for solving in parallel the two-dimensional irregular cutting problem. We use and compare different, variants of the method and various parallel computing systems. Systems used are based on message passing or shared memory paradigm. Parallel algorithms using both methods of communication are proposed. The efficiency of computer system utilization is discussed in the context of unpredictable time requirements of parallel tasks. We present results for different variants of the method together with efficiency measures for parallel implementations, where IBM SP2 and CRAY T3E systems, respectively, have been used.


Author(s):  
Douglas L. Dorset ◽  
Barbara Moss

A number of computing systems devoted to the averaging of electron images of two-dimensional macromolecular crystalline arrays have facilitated the visualization of negatively-stained biological structures. Either by simulation of optical filtering techniques or, in more refined treatments, by cross-correlation averaging, an idealized representation of the repeating asymmetric structure unit is constructed, eliminating image distortions due to radiation damage, stain irregularities and, in the latter approach, imperfections and distortions in the unit cell repeat. In these analyses it is generally assumed that the electron scattering from the thin negativelystained object is well-approximated by a phase object model. Even when absorption effects are considered (i.e. “amplitude contrast“), the expansion of the transmission function, q(x,y)=exp (iσɸ (x,y)), does not exceed the first (kinematical) term. Furthermore, in reconstruction of electron images, kinematical phases are applied to diffraction amplitudes and obey the constraints of the plane group symmetry.


TAPPI Journal ◽  
2015 ◽  
Vol 14 (1) ◽  
pp. 51-60
Author(s):  
HONGHI TRAN ◽  
DANNY TANDRA

Sootblowing technology used in recovery boilers originated from that used in coal-fired boilers. It started with manual cleaning with hand lancing and hand blowing, and evolved slowly into online sootblowing using retractable sootblowers. Since 1991, intensive research and development has focused on sootblowing jet fundamentals and deposit removal in recovery boilers. The results have provided much insight into sootblower jet hydrodynamics, how a sootblower jet interacts with tubes and deposits, and factors influencing its deposit removal efficiency, and have led to two important innovations: fully-expanded sootblower nozzles that are used in virtually all recovery boilers today, and the low pressure sootblowing technology that has been implemented in several new recovery boilers. The availability of powerful computing systems, superfast microprocessors and data acquisition systems, and versatile computational fluid dynamics (CFD) modeling capability in the past two decades has also contributed greatly to the advancement of sootblowing technology. High quality infrared inspection cameras have enabled mills to inspect the deposit buildup conditions in the boiler during operation, and helped identify problems with sootblower lance swinging and superheater platens and boiler bank tube vibrations. As the recovery boiler firing capacity and steam parameters have increased markedly in recent years, sootblowers have become larger and longer, and this can present a challenge in terms of both sootblower design and operation.


Sign in / Sign up

Export Citation Format

Share Document