scholarly journals Distributed Balanced Partitioning via Linear Embedding †

Algorithms ◽  
2019 ◽  
Vol 12 (8) ◽  
pp. 162 ◽  
Author(s):  
Kevin Aydin ◽  
MohammadHossein Bateni ◽  
Vahab Mirrokni

Balanced partitioning is often a crucial first step in solving large-scale graph optimization problems, for example, in some cases, a big graph can be chopped into pieces that fit on one machine to be processed independently before stitching the results together, leading to certain suboptimality from the interaction among different pieces. In other cases, links between different parts may show up in the running time and/or network communications cost, hence the desire to have small cut size. We study a distributed balanced-partitioning problem where the goal is to partition the vertices of a given graph into k pieces so as to minimize the total cut size. Our algorithm is composed of a few steps that are easily implementable in distributed computation frameworks such as MapReduce. The algorithm first embeds nodes of the graph onto a line, and then processes nodes in a distributed manner guided by the linear embedding order. We examine various ways to find the first embedding, for example, via a hierarchical clustering or Hilbert curves. Then we apply four different techniques including local swaps, and minimum cuts on the boundaries of partitions, as well as contraction and dynamic programming. As our empirical study, we compare the above techniques with each other, and also to previous work in distributed graph algorithms, for example, a label-propagation method, FENNEL and Spinner. We report our results both on a private map graph and several public social networks, and show that our results beat previous distributed algorithms: For instance, compared to the label-propagation algorithm, we report an improvement of 15–25% in the cut value. We also observe that our algorithms admit scalable distributed implementation for any number of partitions. Finally, we explain three applications of this work at Google: (1) Balanced partitioning is used to route multi-term queries to different replicas in Google Search backend in a way that reduces the cache miss rates by ≈ 0.5 % , which leads to a double-digit gain in throughput of production clusters. (2) Applied to the Google Maps Driving Directions, balanced partitioning minimizes the number of cross-shard queries with the goal of saving in CPU usage. This system achieves load balancing by dividing the world graph into several “shards”. Live experiments demonstrate an ≈ 40 % drop in the number of cross-shard queries when compared to a standard geography-based method. (3) In a job scheduling problem for our data centers, we use balanced partitioning to evenly distribute the work while minimizing the amount of communication across geographically distant servers. In fact, the hierarchical nature of our solution goes well with the layering of data center servers, where certain machines are closer to each other and have faster links to one another.

2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Junyi Huang ◽  
Yisheng Fang ◽  
Zhichao Ruan

AbstractRecently, spatial photonic Ising machines (SPIM) have been demonstrated to compute the minima of Hamiltonians for large-scale spin systems. Here we propose to implement an antiferromagnetic model through optoelectronic correlation computing with SPIM. Also we exploit the gauge transformation which enables encoding the spins and the interaction strengths in a single phase-only spatial light modulator. With a simple setup, we experimentally show the ground-state-search acceleration of an antiferromagnetic model with 40000 spins in number-partitioning problem. Thus such an optoelectronic computing exhibits great programmability and scalability for the practical applications of studying statistical systems and combinatorial optimization problems.


2016 ◽  
Vol 5 (3) ◽  
pp. 91-100
Author(s):  
Hanaa Abdelrahman ◽  
Mohammed Bakri Bashir ◽  
Adil Yousif

Grid computing presents a new trend to distribute and Internet computing to coordinate large scale heterogeneous resources providing sharing and problem solving in dynamic, multi- institutional virtual organizations. Scheduling is one of the most important problems in computational grid to increase the performance. Genetic Algorithm is adaptive method that can be used to solve optimization problems, based on the genetic process of biological organisms. The objective of this research is to develop a job scheduling algorithm using genetic algorithm with high exploration processes. To evaluate the proposed scheduling algorithm this study conducted a simulation using GridSim Simulator and a number of different workload. The research found that genetic algorithm get best results when increasing the mutation and these result directly proportional with the increase in the number of job. The paper concluded that, the mutation and exploration process has a good effect on the final execution time when we have large number of jobs. However, in small number of job mutation has no effects.


2021 ◽  
Vol 2 (1) ◽  
pp. 1-29
Author(s):  
Hayato Ushijima-Mwesigwa ◽  
Ruslan Shaydulin ◽  
Christian F. A. Negre ◽  
Susan M. Mniszewski ◽  
Yuri Alexeev ◽  
...  

Emerging quantum processors provide an opportunity to explore new approaches for solving traditional problems in the post Moore’s law supercomputing era. However, the limited number of qubits makes it infeasible to tackle massive real-world datasets directly in the near future, leading to new challenges in utilizing these quantum processors for practical purposes. Hybrid quantum-classical algorithms that leverage both quantum and classical types of devices are considered as one of the main strategies to apply quantum computing to large-scale problems. In this article, we advocate the use of multilevel frameworks for combinatorial optimization as a promising general paradigm for designing hybrid quantum-classical algorithms. To demonstrate this approach, we apply this method to two well-known combinatorial optimization problems, namely, the Graph Partitioning Problem, and the Community Detection Problem. We develop hybrid multilevel solvers with quantum local search on D-Wave’s quantum annealer and IBM’s gate-model based quantum processor. We carry out experiments on graphs that are orders of magnitude larger than the current quantum hardware size, and we observe results comparable to state-of-the-art solvers in terms of quality of the solution. Reproducibility : Our code and data are available at Reference [1].


2021 ◽  
Author(s):  
Zhichao Ruan ◽  
Huang Junyi ◽  
Yisheng Fang

Abstract Recently, spatial photonic Ising machines (SPIM) have been demonstrated to compute the minima of Hamiltonians for large-scale spin systems. Here we propose to implement an antiferromagnetic model through optoelectronic correlation computing with SPIM. Also we exploit the gauge transformation which enables encoding the spins and the interaction strengths in a single phase-only spatial light modulator. With a simple setup, we experimentally show the ground state search of an antiferromagnetic model with $40000$ spins in number-partitioning problem. Thus such an optoelectronic computing exhibits great programmability and scalability for the practical applications of studying statistical systems and combinatorial optimization problems.


In the present situation, it may be essential to build a simple data sharing environment to monitor and protect the unauthorized modification of data. In such case, mechanisms may be required to develop to focus on significant weakened networking with proper solutions. In some situations, block chain data management may be used considering the cloud environment. It is well understood that in virtual environment, allocating resources may have significant role towards evaluating the performance including utilization of resources linked to the data center. Accuracy towards allocation of virtual machines in cloud data centers may be more essential considering the optimization problems in cloud computing. In such cases, it may also be desirable to prioritize on virtual machines linked to cloud data centers. Consolidating the dynamic virtual machines may also permit the virtual server providers to optimize utilization of resources and to focus on energy consumption. In fact, tremendous rise in acquiring computational power driven by modern service applications may be linked towards establishment of large-scale virtualized data centers. Accordingly, the joint collaboration of smart connected devices with data analytics may also enable enormous applications towards different predictive maintenance systems. To obtain the near optimal as well as feasible results in this case, it may be desirable to simulate implementing the algorithms and focusing on application codes. Also, different approaches may also be needed to minimize development time and cost. In many cases, the experimental result proves that the simulation techniques may minimize the cache miss and improve the execution time. In this paper, it has been intended towards distribution of tasks along with implementation mechanisms linked to virtual machines.


2012 ◽  
Vol 2012 ◽  
pp. 1-18 ◽  
Author(s):  
Xiaocheng Liu ◽  
Bin Chen ◽  
Xiaogang Qiu ◽  
Ying Cai ◽  
Kedi Huang

An increasing number of high performance computing parallel applications leverages the power of the cloud for parallel processing. How to schedule the parallel applications to improve the quality of service is the key to the successful host of parallel applications in the cloud. The large scale of the cloud makes the parallel job scheduling more complicated as even simple parallel job scheduling problem is NP-complete. In this paper, we propose a parallel job scheduling algorithm named MEASY. MEASY adopts migration and consolidation to enhance the most popular EASY scheduling algorithm. Our extensive experiments on well-known workloads show that our algorithm takes very good care of the quality of service. For two common parallel job scheduling objectives, our algorithm produces an up to 41.1% and an average of 23.1% improvement on the average response time; an up to 82.9% and an average of 69.3% improvement on the average slowdown. Our algorithm is robust even in terms that it allows inaccurate CPU usage estimation and high migration cost. Our approach involves trivial modification on EASY and requires no additional technique; it is practical and effective in the cloud environment.


Author(s):  
Ruhul Sarker ◽  
Daryl Essam ◽  
S.M. Kamrul Hasan ◽  
A.N. Mustafizul Karim

AbstractThe job scheduling problem (JSP) is considered as one of the most complex combinatorial optimization problems. JSP is not an independent task, but is rather a part of a company business case. In this paper, we have studied JSPs under sudden machine breakdown scenarios that introduce a risk of not completing the jobs on time. We have first solved JSPs using an improved memetic algorithm and extended the algorithm to deal with the disruption situations, and then developed a simulation model to analyze the risk of using a job order and delivery scenario. This paper deals with job scheduling under ideal conditions and rescheduling under machine breakdown, and provides a risk analysis for a production business case. The extended algorithm provides better understanding and results than existing algorithms, the rescheduling shows a good way of recovering from disruptions, and the risk analysis shows an effective way of maximizing return under such situations.


Author(s):  
Paul Cronin ◽  
Harry Woerde ◽  
Rob Vasbinder

Sign in / Sign up

Export Citation Format

Share Document