scholarly journals Cardinality Encodings for Graph Optimization Problems

Author(s):  
Alexey Ignatiev ◽  
Antonio Morgado ◽  
Joao Marques-Silva

Different optimization problems defined on graphs find application in complex network analysis. Existing propositional encodings render impractical the use of propositional satisfiability (SAT) and maximum satisfiability (MaxSAT) solvers for solving a variety of these problems on large graphs. This paper has two main contributions. First, the paper identifies sources of inefficiency in existing encodings for different optimization problems in graphs. Second, for the concrete case of the maximum clique problem, the paper develops a novel encoding which is shown to be far more compact than existing encodings for large sparse graphs. More importantly, the experimental results show that the proposed encoding enables existing SAT solvers to compute a maximum clique for large sparse networks, often more efficiently than the state of the art.

Author(s):  
Jeremias Berg ◽  
Fahiem Bacchus ◽  
Alex Poole

Maximum satisfiability (MaxSat) solving is an active area of research motivated by numerous successful applications to solving NP-hard combinatorial optimization problems. One of the most successful approaches for solving MaxSat instances from real world domains are the so called implicit hitting set (IHS) solvers. IHS solvers decouple MaxSat solving into separate core-extraction (i.e. reasoning) and optimization steps which are tackled by a Boolean satisfiability (SAT) and an integer linear programming (IP) solver, respectively. While the approach shows state-of-the-art performance on many industrial instances, it is known that there exists instances on which IHS solvers need to extract an exponential number of cores before terminating. Motivated by the simplest of these problematic instances, we propose abstract cores, a compact representation for a potentially exponential number of regular cores. We demonstrate how to incorporate abstract core reasoning into the IHS algorithm and report on an empirical evaluation demonstrating, that including abstract cores into a state-of-the-art IHS solver improves its performance enough to surpass the best performing solvers of the 2019 MaxSat Evaluation.


Author(s):  
Fahiem Bacchus ◽  
Antti Hyttinen ◽  
Matti Järvisalo ◽  
Paul Saikko

Maximum satisfiability (MaxSAT) offers a competitive approach to solving NP-hard real-world optimization problems. While state-of-the-art MaxSAT solvers rely heavily on Boolean satisfiability (SAT) solvers, a recent trend, brought on by MaxSAT solvers implementing the so-called implicit hitting set (IHS) approach, is to integrate techniques from the realm of integer programming (IP) into the solving process. This allows for making use of additional IP solving techniques to further speed up MaxSAT solving. In this line of work, we investigate the integration of the technique of reduced cost fixing from the IP realm into IHS solvers, and empirically show that reduced cost fixing considerable speeds up a state-of-the-art MaxSAT solver implementing the IHS approach.


2021 ◽  
Vol 27_NS1 (1) ◽  
pp. 32-47
Author(s):  
Ákos Beke ◽  
Sándor Szabó ◽  
Bogdán Zavalnij

Many combinatorial optimization problems can be expressed in terms of zero-one linear programs. For the maximum clique problem the so-called edge reformulation is applied most commonly. Two less frequently used LP equivalents are the independent set and edge covering set reformulations. The number of the constraints (as a function of the number of vertices of the ground graph) is asymptotically quadratic in the edge and the edge covering set LP reformulations and it is exponential in the independent set reformulation, respectively. F. D. Croce and R. Tadei proposed an approach in which the number of the constraints is equal to the number of the vertices. In this paper we are looking for possible tighter variants of these linear programs.


Author(s):  
Bharath Pattabiraman ◽  
Md. Mostofa Ali Patwary ◽  
Assefaw H. Gebremedhin ◽  
Wei-keng Liao ◽  
Alok Choudhary

2020 ◽  
Vol 34 (02) ◽  
pp. 1544-1551
Author(s):  
Tuukka Korhonen ◽  
Matti J„ärvisalo

The reconstruction of the evolutionary tree of a set of species based on qualitative attributes is a central problem in phylogenetics. In the NP-hard perfect phylogeny problem the input is a set of taxa (species) and characters (attributes) on them, and the task is to find an evolutionary tree that describes the evolution of the taxa so that each character state evolves only once. However, in practical situations a perfect phylogeny rarely exists, motivating the maximum compatibility problem of finding the largest subset of characters admitting a perfect phylogeny. Various declarative approaches, based on applying integer programming (IP), answer set programming (ASP) and pseudo-Boolean optimization (PBO) solvers, have been proposed for maximum compatibility. In this work we develop a new hybrid approach to solving maximum compatibility for multi-state characters, making use of both declarative optimization techniques (specifically maximum satisfiability, MaxSAT) and an adaptation of the Bouchitt'e-Todinca approach to triangulation-based graph optimization problems. Empirically our approach outperforms in scalability the earlier proposed approaches w.r.t. various parameters underlying the problem.


2021 ◽  
Vol 72 ◽  
pp. 39-67
Author(s):  
Shaowei Cai ◽  
Jinkun Lin ◽  
Yiyuan Wang ◽  
Darren Strash

This paper explores techniques to quickly solve the maximum weight clique problem (MWCP) in very large scale sparse graphs. Due to their size, and the hardness of MWCP, it is infeasible to solve many of these graphs with exact algorithms. Although recent heuristic algorithms make progress in solving MWCP in large graphs, they still need considerable time to get a high-quality solution. In this work, we focus on solving MWCP for large sparse graphs within a short time limit. We propose a new method for MWCP which interleaves clique finding with data reduction rules. We propose novel ideas to make this process efficient, and develop an algorithm called FastWClq. Experiments on a broad range of large sparse graphs show that FastWClq finds better solutions than state-of-the-art algorithms while the running time of FastWClq is much shorter than the competitors for most instances. Further, FastWClq proves the optimality of its solutions for roughly half of the graphs, all with at least 105 vertices, with an average time of 21 seconds.


2018 ◽  
Vol 7 (4.3) ◽  
pp. 293
Author(s):  
S. V. Listrovoy ◽  
O. V. Golovko ◽  
V. M. Butenko ◽  
M. V. Ushakov ◽  
. .

In the train traffic organization, large amount of information should be processed in real time. Thereby in complicated dispatching systems, rational use of resources is needed. To perform this work a model of a management system is constructed and it is represented as a sparse graph. Further optimization of the model requires solving the Maximum Clique Problem (MCP) for the less time than the exponential time. This article contains two procedures for solving the task for subexponential time. Procedure B accurately estimates the size of the maximum clique in the graph and performs the sorting of the vertices, in such a way that the vertices that are not exactly in the maximum clique can be rejected. Procedure A, in fact, finds cliques of the maximum size in a given graph, using the procedure B. The main advantage of this method is that using it is expedient in real time for sufficiently large graphs, which in turn is important for the construction of control systems.  


Algorithms ◽  
2021 ◽  
Vol 14 (6) ◽  
pp. 187
Author(s):  
Aaron Barbosa ◽  
Elijah Pelofske ◽  
Georg Hahn ◽  
Hristo N. Djidjev

Quantum annealers, such as the device built by D-Wave Systems, Inc., offer a way to compute solutions of NP-hard problems that can be expressed in Ising or quadratic unconstrained binary optimization (QUBO) form. Although such solutions are typically of very high quality, problem instances are usually not solved to optimality due to imperfections of the current generations quantum annealers. In this contribution, we aim to understand some of the factors contributing to the hardness of a problem instance, and to use machine learning models to predict the accuracy of the D-Wave 2000Q annealer for solving specific problems. We focus on the maximum clique problem, a classic NP-hard problem with important applications in network analysis, bioinformatics, and computational chemistry. By training a machine learning classification model on basic problem characteristics such as the number of edges in the graph, or annealing parameters, such as the D-Wave’s chain strength, we are able to rank certain features in the order of their contribution to the solution hardness, and present a simple decision tree which allows to predict whether a problem will be solvable to optimality with the D-Wave 2000Q. We extend these results by training a machine learning regression model that predicts the clique size found by D-Wave.


Sign in / Sign up

Export Citation Format

Share Document