scholarly journals Sparse graphs and an augmentation problem

Author(s):  
Csaba Király ◽  
András Mihálykó

AbstractFor two integers $$k>0$$ k > 0 and $$\ell $$ ℓ , a graph $$G=(V,E)$$ G = ( V , E ) is called $$(k,\ell )$$ ( k , ℓ ) -tight if $$|E|=k|V|-\ell $$ | E | = k | V | - ℓ and $$i_G(X)\le k|X|-\ell $$ i G ( X ) ≤ k | X | - ℓ for each $$X\subseteq V$$ X ⊆ V for which $$i_G(X)\ge 1$$ i G ( X ) ≥ 1 , where $$i_G(X)$$ i G ( X ) denotes the number of induced edges by X. G is called $$(k,\ell )$$ ( k , ℓ ) -redundant if $$G-e$$ G - e has a spanning $$(k,\ell )$$ ( k , ℓ ) -tight subgraph for all $$e\in E$$ e ∈ E . We consider the following augmentation problem. Given a graph $$G=(V,E)$$ G = ( V , E ) that has a $$(k,\ell )$$ ( k , ℓ ) -tight spanning subgraph, find a graph $$H=(V,F)$$ H = ( V , F ) with the minimum number of edges, such that $$G\cup H$$ G ∪ H is $$(k,\ell )$$ ( k , ℓ ) -redundant. We give a polynomial algorithm and a min-max theorem for this augmentation problem when the input is $$(k,\ell )$$ ( k , ℓ ) -tight. For general inputs, we give a polynomial algorithm when $$k\ge \ell $$ k ≥ ℓ and show the NP-hardness of the problem when $$k<\ell $$ k < ℓ . Since $$(k,\ell )$$ ( k , ℓ ) -tight graphs play an important role in rigidity theory, these algorithms can be used to make several types of rigid frameworks redundantly rigid by adding a smallest set of new bars.

1998 ◽  
Vol Vol. 3 no. 1 ◽  
Author(s):  
Ulrik Brandes ◽  
Dagmar Handke

International audience For any fixed parameter t greater or equal to 1, a \emph t-spanner of a graph G is a spanning subgraph in which the distance between every pair of vertices is at most t times their distance in G. A \emph minimum t-spanner is a t-spanner with minimum total edge weight or, in unweighted graphs, minimum number of edges. In this paper, we prove the NP-hardness of finding minimum t-spanners for planar weighted graphs and digraphs if t greater or equal to 3, and for planar unweighted graphs and digraphs if t greater or equal to 5. We thus extend results on that problem to the interesting case where the instances are known to be planar. We also introduce the related problem of finding minimum \emphplanar t-spanners and establish its NP-hardness for similar fixed values of t.


2009 ◽  
Vol 20 (01) ◽  
pp. 135-149 ◽  
Author(s):  
JOACHIM GUDMUNDSSON ◽  
MICHIEL SMID

Given a connected geometric graph G, we consider the problem of constructing a t-spanner of G having the minimum number of edges. We prove that for every real number t with [Formula: see text], there exists a connected geometric graph G with n vertices, such that every t-spanner of G contains Ω(n1+1/t) edges. This bound almost matches the known upper bound, which states that every connected weighted graph with n vertices contains a t-spanner with O(n1+2/(t-1)) edges. We also prove that the problem of deciding whether a given geometric graph contains a t-spanner with at most K edges is NP-hard. Previously, this NP-hardness result was only known for non-geometric graphs.


2013 ◽  
Vol 2013 ◽  
pp. 1-4
Author(s):  
Jianping Li ◽  
Juanping Zhu

This paper considers the general capacity expansion path problem (GCEP) for the telecommunication operators. We investigate the polynomial equivalence between the GCEP problem and the constrained shortest path problem (CSP) and present a pseudopolynomial algorithm for the GCEP problem, no matter the graph is acyclic or not. Furthermore, we investigate two special versions of the GCEP problem. For the minimum number arc capacity expansion path problem (MN-CEP), we give a strongly polynomial algorithm based on the dynamic programming. For the minimum-cost capacity expansion shortest path problem (MCESP), we give a strongly polynomial algorithm by constructing a shortest paths network.


2005 ◽  
Vol 15 (06) ◽  
pp. 575-590 ◽  
Author(s):  
GRUIA CĂLINESCU ◽  
ADRIAN DUMITRESCU ◽  
HOWARD KARLOFF ◽  
PENG-JUN WAN

We study the problem of separating n points in the plane, no two of which have the same x- or y-coordinate, using a minimum number of vertical and horizontal lines avoiding the points, so that each cell of the subdivision contains at most one point. Extending previous NP-hardness results due to Freimer et al. we prove that this problem and some variants of it are APX-hard. We give a 2-approximation algorithm for this problem, and a d-approximation algorithm for the d-dimensional variant, in which the points are to be separated using axis-parallel hyperplanes. To this end, we reduce the point separation problem to the rectangle stabbing problem studied by Gaur et al. Their approximation algorithm uses LP-rounding. We present an alternative LP-rounding procedure which also works for the rectangle stabbing problem. We show that the integrality ratio of the LP is exactly 2.


1983 ◽  
Vol 6 (4) ◽  
pp. 715-726 ◽  
Author(s):  
E. J. Farrell

LetGbe a graph. With every pathαofGlet us associate a weightwαWith every spanning subgraphCofGconsisting of pathsα1,α2,…,αk, let us associate the weightw(C)=∏i=1kwαi. The path polynomial ofGis∑w(C), where the summation is taken over all the spanning subgraphs ofGwhose components are paths. Some basic properties of these polynomials are given. The polynomials are then used to obtain results about the minimum number of node disjoint path coverings in graphs.


2003 ◽  
Vol 01 (01) ◽  
pp. 71-94 ◽  
Author(s):  
MICHAL OZERY-FLATO ◽  
RON SHAMIR

A central problem in genome rearrangement is finding a most parsimonious rearrangement scenario using certain rearrangement operations. An important problem of this type is sorting a signed genome by reversals and translocations (SBRT). Hannenhalli and Pevzner presented a duality theorem for SBRT which leads to a polynomial time algorithm for sorting a multi-chromosomal genome using a minimum number of reversals and translocations. However, there is one case for which their theorem and algorithm fail. We describe that case and suggest a correction to the theorem and the polynomial algorithm. The solution of SBRT uses a reduction to the problem of sorting a signed permutation by reversals (SBR). The best extant algorithms for SBR require quadratic time. The common approach to solve SBR is by finding a safe reversal using the overlap graph or the interleaving graph of a permutation. We describe a family of signed permutations which proves a quadratic lower bound on the number of affected vertices in the overlap/interleaving graph during any optimal sorting scenario. This implies, in particular, an Ω(n3) lower bound for Bergeron's algorithm.


2016 ◽  
Vol 26 (1) ◽  
pp. 78-98 ◽  
Author(s):  
NATASHA MORRISON ◽  
JONATHAN A. NOEL ◽  
ALEX SCOTT

LetQddenote the hypercube of dimensiond. Givend⩾m, a spanning subgraphGofQdis said to be (Qd,Qm)-saturated if it does not containQmas a subgraph but adding any edge ofE(Qd)\E(G) creates a copy ofQminG. Answering a question of Johnson and Pinto [27], we show that for every fixedm⩾ 2 the minimum number of edges in a (Qd,Qm)-saturated graph is Θ(2d).We also study weak saturation, which is a form of bootstrap percolation. A spanning subgraph ofQdis said to beweakly(Qd,Qm)-saturated if the edges ofE(Qd)\E(G) can be added toGone at a time so that each added edge creates a new copy ofQm. Answering another question of Johnson and Pinto [27], we determine the minimum number of edges in a weakly (Qd,Qm)-saturated graph for alld⩾m⩾ 1. More generally, we determine the minimum number of edges in a subgraph of thed-dimensional gridPkdwhich is weakly saturated with respect to ‘axis aligned’ copies of a smaller gridPrm. We also study weak saturation of cycles in the grid.


2021 ◽  
pp. 120-128
Author(s):  
A. N. Rybalov ◽  

Generic-case approach to algorithmic problems was suggested by A. Miasnikov, V. Kapovich, P. Schupp, and V. Shpilrain in 2003. This approach studies behavior of an algorithm on typical (almost all) inputs and ignores the rest of inputs. In this paper, we study the generic complexity of the isomorphism problem for finite semigroups. In this problem, for any two semigroups of the same order, given by their multiplication tables, it is required to determine whether they are isomorphic. V. Zemlyachenko, N. Korneenko, and R. Tyshkevich in 1982 proved that the graph isomorphism problem polynomially reduces to this problem. The graph isomorphism problem is a well-known algorithmic problem that has been actively studied since the 1970s, and for which polynomial algorithms are still unknown. So from a computational point of view the studied problem is no simpler than the graph isomorphism problem. We present a generic polynomial algorithm for the isomorphism problem of finite semigroups. It is based on the characterization of almost all finite semigroups as 3-nilpotent semigroups of a special form, established by D. Kleitman, B. Rothschild, and J. Spencer, as well as the Bollobas polynomial algorithm, which solves the isomorphism problem for almost all strongly sparse graphs.


Author(s):  
D. C. Joy ◽  
R. D. Bunn

The information available from an SEM image is limited both by the inherent signal to noise ratio that characterizes the image and as a result of the transformations that it may undergo as it is passed through the amplifying circuits of the instrument. In applications such as Critical Dimension Metrology it is necessary to be able to quantify these limitations in order to be able to assess the likely precision of any measurement made with the microscope.The information capacity of an SEM signal, defined as the minimum number of bits needed to encode the output signal, depends on the signal to noise ratio of the image - which in turn depends on the probe size and source brightness and acquisition time per pixel - and on the efficiency of the specimen in producing the signal that is being observed. A detailed analysis of the secondary electron case shows that the information capacity C (bits/pixel) of the SEM signal channel could be written as :


2020 ◽  
Vol 63 (6) ◽  
pp. 1947-1957
Author(s):  
Alexandra Hollo ◽  
Johanna L. Staubitz ◽  
Jason C. Chow

Purpose Although sampling teachers' child-directed speech in school settings is needed to understand the influence of linguistic input on child outcomes, empirical guidance for measurement procedures needed to obtain representative samples is lacking. To optimize resources needed to transcribe, code, and analyze classroom samples, this exploratory study assessed the minimum number and duration of samples needed for a reliable analysis of conventional and researcher-developed measures of teacher talk in elementary classrooms. Method This study applied fully crossed, Person (teacher) × Session (samples obtained on 3 separate occasions) generalizability studies to analyze an extant data set of three 10-min language samples provided by 28 general and special education teachers recorded during large-group instruction across the school year. Subsequently, a series of decision studies estimated of the number and duration of sessions needed to obtain the criterion g coefficient ( g > .70). Results The most stable variables were total number of words and mazes, requiring only a single 10-min sample, two 6-min samples, or three 3-min samples to reach criterion. No measured variables related to content or complexity were adequately stable regardless of number and duration of samples. Conclusions Generalizability studies confirmed that a large proportion of variance was attributable to individuals rather than the sampling occasion when analyzing the amount and fluency of spontaneous teacher talk. In general, conventionally reported outcomes were more stable than researcher-developed codes, which suggests some categories of teacher talk are more context dependent than others and thus require more intensive data collection to measure reliably.


Sign in / Sign up

Export Citation Format

Share Document