hardness of approximation
Recently Published Documents


TOTAL DOCUMENTS

51
(FIVE YEARS 13)

H-INDEX

9
(FIVE YEARS 1)

2021 ◽  
Vol 13 (4) ◽  
pp. 1-20
Author(s):  
Alex Brandts ◽  
Marcin Wrochna ◽  
Stanislav Živný

While 3-SAT is NP-hard, 2-SAT is solvable in polynomial time. Austrin et al. [SICOMP’17] proved a result known as “(2+ɛ)-SAT is NP-hard.” They showed that the problem of distinguishing k -CNF formulas that are g -satisfiable (i.e., some assignment satisfies at least g literals in every clause) from those that are not even 1-satisfiable is NP-hard if g/k < 1/2 and is in P otherwise. We study a generalisation of SAT on arbitrary finite domains, with clauses that are disjunctions of unary constraints, and establish analogous behaviour. Thus, we give a dichotomy for a natural fragment of promise constraint satisfaction problems ( PCSPs ) on arbitrary finite domains. The hardness side is proved using the algebraic approach via a new general NP-hardness criterion on polymorphisms, which is based on a gap version of the Layered Label Cover problem. We show that previously used criteria are insufficient—the problem hence gives an interesting benchmark of algebraic techniques for proving hardness of approximation in problems such as PCSPs.


2021 ◽  
Vol 64 (11) ◽  
pp. 131-138
Author(s):  
Zhengfeng Ji ◽  
Anand Natarajan ◽  
Thomas Vidick ◽  
John Wright ◽  
Henry Yuen

Note from the Research Highlights Co-Chairs: A Research Highlights paper appearing in Communications is usually peer-reviewed prior to publication. The following paper is unusual in that it is still under review. However, the result has generated enormous excitement in the research community, and came strongly nominated by SIGACT, a nomination seconded by external reviewers. The complexity class NP characterizes the collection of computational problems that have efficiently verifiable solutions. With the goal of classifying computational problems that seem to lie beyond NP, starting in the 1980s complexity theorists have considered extensions of the notion of efficient verification that allow for the use of randomness (the class MA), interaction (the class IP), and the possibility to interact with multiple proofs, or provers (the class MIP). The study of these extensions led to the celebrated PCP theorem and its applications to hardness of approximation and the design of cryptographic protocols. In this work, we study a fourth modification to the notion of efficient verification that originates in the study of quantum entanglement. We prove the surprising result that every problem that is recursively enumerable, including the Halting problem, can be efficiently verified by a classical probabilistic polynomial-time verifier interacting with two all-powerful but noncommunicating provers sharing entanglement. The result resolves long-standing open problems in the foundations of quantum mechanics (Tsirelson's problem) and operator algebras (Connes' embedding problem).


Author(s):  
Omar El Housni ◽  
Vineet Goyal

In this paper, we study the performance of affine policies for a two-stage, adjustable, robust optimization problem with a fixed recourse and an uncertain right-hand side belonging to a budgeted uncertainty set. This is an important class of uncertainty sets, widely used in practice, in which we can specify a budget on the adversarial deviations of the uncertain parameters from the nominal values to adjust the level of conservatism. The two-stage adjustable robust optimization problem is hard to approximate within a factor better than [Formula: see text] even for budget of uncertainty sets in which [Formula: see text] is the number of decision variables. Affine policies, in which the second-stage decisions are constrained to be an affine function of the uncertain parameters provide a tractable approximation for the problem and have been observed to exhibit good empirical performance. We show that affine policies give an [Formula: see text]-approximation for the two-stage, adjustable, robust problem with fixed nonnegative recourse for budgeted uncertainty sets. This matches the hardness of approximation, and therefore, surprisingly, affine policies provide an optimal approximation for the problem (up to a constant factor). We also show strong theoretical performance bounds for affine policy for a significantly more general class of intersection of budgeted sets, including disjoint constrained budgeted sets, permutation invariant sets, and general intersection of budgeted sets. Our analysis relies on showing the existence of a near-optimal, feasible affine policy that satisfies certain nice structural properties. Based on these structural properties, we also present an alternate algorithm to compute a near-optimal affine solution that is significantly faster than computing the optimal affine policy by solving a large linear program.


Algorithms ◽  
2020 ◽  
Vol 13 (12) ◽  
pp. 344
Author(s):  
Takeo Hagiwara ◽  
Tatsuie Tsukiji

Langton’s ant is a deterministic cellular automaton studied in many fields, artificial life, computational complexity, cryptography, emergent dynamics, Lorents lattice gas, and so forth, motivated by the hardness of predicting the ant’s macroscopic behavior from an initial microscopic configuration. Gajardo, Moreira, and Goles (2002) proved that Langton’s ant is PTIME -hard for reachability. On a twisted torus, we demonstrate that it is PSPACE hard to determine whether the ant will ever visit almost all vertices or nearly none of them.


2020 ◽  
Vol 26 (4) ◽  
Author(s):  
Zarathustra Brady ◽  
Larry Guth ◽  
Fedor Manin

2019 ◽  
Vol 29 (8) ◽  
pp. 1185-1210 ◽  
Author(s):  
Albert Atserias ◽  
Anuj Dawar

Abstract We consider the hardness of approximation of optimization problems from the point of view of definability. For many $\textrm{NP}$-hard optimization problems it is known that, unless $\textrm{P} = \textrm{NP} $, no polynomial-time algorithm can give an approximate solution guaranteed to be within a fixed constant factor of the optimum. We show, in several such instances and without any complexity theoretic assumption, that no algorithm that is expressible in fixed-point logic with counting (FPC) can compute an approximate solution. Since important algorithmic techniques for approximation algorithms (such as linear or semidefinite programming) are expressible in FPC, this yields lower bounds on what can be achieved by such methods. The results are established by showing lower bounds on the number of variables required in first-order logic with counting to separate instances with a high optimum from those with a low optimum for fixed-size instances.


Sign in / Sign up

Export Citation Format

Share Document