scholarly journals A Network Simulation Tool for Task Scheduling

10.14311/1658 ◽  
2012 ◽  
Vol 52 (5) ◽  
Author(s):  
Ondřej Votava

Distributed computing may be looked at from many points of view. Task scheduling is the viewpoint, where a distributed application can be described as a Directed Acyclic Graph and every node of the graph is executed independently. There are, however, data dependencies and the nodes have to be executed in a specified order. Hence the parallelism of the execution is limited. The scheduling problem is difficult and therefore heuristics are used. However, many inaccuracies are caused by the model used for the system, in which the heuristics are being tested. In this paper we present a tool for simulating the execution of the distributed application on a “real” computer network, and try to tell how the executionis influenced compared to the model.

Author(s):  
Jahwan Koo ◽  
Nawab Muhammad Faseeh Qureshi ◽  
Isma Farah Siddiqui ◽  
Asad Abbas ◽  
Ali Kashif Bashir

Abstract Real-time data streaming fetches live sensory segments of the dataset in the heterogeneous distributed computing environment. This process assembles data chunks at a rapid encapsulation rate through a streaming technique that bundles sensor segments into multiple micro-batches and extracts into a repository, respectively. Recently, the acquisition process is enhanced with an additional feature of exchanging IoT devices’ dataset comprised of two components: (i) sensory data and (ii) metadata. The body of sensory data includes record information, and the metadata part consists of logs, heterogeneous events, and routing path tables to transmit micro-batch streams into the repository. Real-time acquisition procedure uses the Directed Acyclic Graph (DAG) to extract live query outcomes from in-place micro-batches through MapReduce stages and returns a result set. However, few bottlenecks affect the performance during the execution process, such as (i) homogeneous micro-batches formation only, (ii) complexity of dataset diversification, (iii) heterogeneous data tuples processing, and (iv) linear DAG workflow only. As a result, it produces huge processing latency and the additional cost of extracting event-enabled IoT datasets. Thus, the Spark cluster that processes Resilient Distributed Dataset (RDD) in a fast-pace using Random access memory (RAM) defies expected robustness in processing IoT streams in the distributed computing environment. This paper presents an IoT-enabled Directed Acyclic Graph (I-DAG) technique that labels micro-batches at the stage of building a stream event and arranges stream elements with event labels. In the next step, heterogeneous stream events are processed through the I-DAG workflow, which has non-linear DAG operation for extracting queries’ results in a Spark cluster. The performance evaluation shows that I-DAG resolves homogeneous IoT-enabled stream event issues and provides an effective stream event heterogeneous solution for IoT-enabled datasets in spark clusters.


2012 ◽  
Vol 58 (4) ◽  
pp. 369-379 ◽  
Author(s):  
Dawid Król ◽  
Dawid Zydek ◽  
Leszek Koszałka

Abstract This paper concerns Directed Acyclic Graph task scheduling on parallel executors. The problem is solved using two new implementations of Tabu Search and genetic algorithm presented in the paper. A new approach to solution coding is also introduced and implemented in both metaheuristics algorithms. Results given by the algorithms are compared to those generated by greedy LPT and SS-FF algorithms; and HAR algorithm. The analysis of the obtained results of multistage simulation experiments confirms the conclusion that the proposed and implemented algorithms are characterized by very good performance and characteristics.


2019 ◽  
Vol 11 (6) ◽  
pp. 121 ◽  
Author(s):  
Ling Xu ◽  
Jianzhong Qiao ◽  
Shukuan Lin ◽  
Wanting Zhang

Volunteer computing (VC) is a distributed computing paradigm, which provides unlimited computing resources in the form of donated idle resources for many large-scale scientific computing applications. Task scheduling is one of the most challenging problems in VC. Although, dynamic scheduling problem with deadline constraint has been extensively studied in prior studies in the heterogeneous system, such as cloud computing and clusters, these algorithms can’t be fully applied to VC. This is because volunteer nodes can get offline whenever they want without taking any responsibility, which is different from other distributed computing. For this situation, this paper proposes a dynamic task scheduling algorithm for heterogeneous VC with deadline constraint, called deadline preference dispatch scheduling (DPDS). The DPDS algorithm selects tasks with the nearest deadline each time and assigns them to volunteer nodes (VN), which solves the dynamic task scheduling problem with deadline constraint. To make full use of resources and maximize the number of completed tasks before the deadline constraint, on the basis of the DPDS algorithm, improved dispatch constraint scheduling (IDCS) is further proposed. To verify our algorithms, we conducted experiments, and the results show that the proposed algorithms can effectively solve the dynamic task assignment problem with deadline constraint in VC.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Son Nguyen ◽  
Peggy Shu-Ling Chen ◽  
Yuquan Du

PurposeAlthough being considered for adoption by stakeholders in container shipping, application of blockchain is hindered by different factors. This paper investigates the potential operational risks of blockchain-integrated container shipping systems as one of such barriers.Design/methodology/approachLiterature review is employed as the method of risk identification. Scientific articles, special institutional reports and publications of blockchain solution providers were included in an inclusive qualitative analysis. A directed acyclic graph (DAG) was constructed and analyzed based on network topological metrics.FindingsTwenty-eight potential risks and 47 connections were identified in three groups of initiative, transitional and sequel. The DAG analysis results reflect a relatively well-connected network of identified hazardous events (HEs), suggesting the pervasiveness of information risks and various multiple-event risk scenarios. The criticality of the connected systems' security and information accuracy are also indicated.Originality/valueThis paper indicates the changes of container shipping operational risk in the process of blockchain integration by using updated data. It creates awareness of the emerging risks, provides their insights and establishes the basis for further research.


Sign in / Sign up

Export Citation Format

Share Document