scholarly journals A new 1.375-approximation algorithm for sorting by transpositions

2022 ◽  
Vol 17 (1) ◽  
Author(s):  
Luiz Augusto G. Silva ◽  
Luis Antonio B. Kowada ◽  
Noraí Romeu Rocco ◽  
Maria Emília M. T. Walter

Abstract Background sorting by transpositions (SBT) is a classical problem in genome rearrangements. In 2012, SBT was proven to be $$\mathcal {NP}$$ NP -hard and the best approximation algorithm with a 1.375 ratio was proposed in 2006 by Elias and Hartman (EH algorithm). Their algorithm employs simplification, a technique used to transform an input permutation $$\pi$$ π into a simple permutation$${\hat{\pi }}$$ π ^ , presumably easier to handle with. The permutation $${\hat{\pi }}$$ π ^ is obtained by inserting new symbols into $$\pi$$ π in a way that the lower bound of the transposition distance of $$\pi$$ π is kept on $${\hat{\pi }}$$ π ^ . The simplification is guaranteed to keep the lower bound, not the transposition distance. A sequence of operations sorting $${\hat{\pi }}$$ π ^ can be mimicked to sort $$\pi$$ π . Results and conclusions First, using an algebraic approach, we propose a new upper bound for the transposition distance, which holds for all $$S_n$$ S n . Next, motivated by a problem identified in the EH algorithm, which causes it, in scenarios involving how the input permutation is simplified, to require one extra transposition above the 1.375-approximation ratio, we propose a new approximation algorithm to solve SBT ensuring the 1.375-approximation ratio for all $$S_n$$ S n . We implemented our algorithm and EH’s. Regarding the implementation of the EH algorithm, two other issues were identified and needed to be fixed. We tested both algorithms against all permutations of size n, $$2\le n \le 12$$ 2 ≤ n ≤ 12 . The results show that the EH algorithm exceeds the approximation ratio of 1.375 for permutations with a size greater than 7. The percentage of computed distances that are equal to transposition distance, computed by the implemented algorithms are also compared with others available in the literature. Finally, we investigate the performance of both implementations on longer permutations of maximum length 500. From the experiments, we conclude that maximum and the average distances computed by our algorithm are a little better than the ones computed by the EH algorithm and the running times of both algorithms are similar, despite the time complexity of our algorithm being higher.

2021 ◽  
Author(s):  
Cristina G. Fernandes ◽  
Carla N. Lintzmayer ◽  
Mário César San Felice

We denote by Maximum Leaf Spanning k-Forest the problem of, given a positive integer k and a graph G with at most k components, finding a spanning forest in G with at most k components and the maximum number of leaves. A leaf in a forest is defined as a vertex of degree at most one. The case k = 1 for connected graphs is known to be NP-hard, and is well studied in the literature, with the best approximation algorithm proposed more than 20 years ago by Solis-Oba. The best known approximation algorithm for Maximum Leaf Spanning k-Forest with a slightly different leaf definition is a 3-approximation based on an approach by Lu and Ravi for the k = 1 case. We extend the algorithm of Solis-Oba to achieve a 2-approximation for Maximum Leaf Spanning k-Forest.


1979 ◽  
Vol 46 (3) ◽  
pp. 534-540 ◽  
Author(s):  
U. C. Luft ◽  
J. A. Loeppky ◽  
E. M. Mostyn

In view of uncertainties about the best way to estimate mean alveolar gases in patients with ventilation-perfusion inequalities, three different methods were evaluated on 54 patients. 1) O2 and CO2 were recorded by mass spectrometer on an O2 (x)-CO2 (y) diagram. The coordinates at the intersect of the expiratory record with the mixed expired R line (RE) ives the mean alveolar values (PAo2 and PAco2. 2)pa'co2 was calculated with the Bohr equation using a predicted anatomic dead space and PA'o2 was derived with the alveolar equation. 3) End-tidal (ET) P02 were averaged over 1 min at rest in steady state. Mean RET calculated from 3 was identical with RE. Mean values for PAco2, PA'CO2. and PETco2 differed by less that 1 Torr, but the variance was least with the end-tidal method. There was a highly significant correlation between delta aAPco2 using PETco2 and VD/VT, better than with either of the other methods. The end-tidal measurement appears to give the best approximation of mean alveolar gas in pulmonary patients.


2020 ◽  
Vol 117 (28) ◽  
pp. 16181-16186
Author(s):  
Rocco Martinazzo ◽  
Eli Pollak

The Ritz upper bound to eigenvalues of Hermitian operators is essential for many applications in science. It is a staple of quantum chemistry and physics computations. The lower bound devised by Temple in 1928 [G. Temple,Proc. R. Soc. A Math. Phys. Eng. Sci.119, 276–293 (1928)] is not, since it converges too slowly. The need for a good lower-bound theorem and algorithm cannot be overstated, since an upper bound alone is not sufficient for determining differences between eigenvalues such as tunneling splittings and spectral features. In this paper, after 90 y, we derive a generalization and improvement of Temple’s lower bound. Numerical examples based on implementation of the Lanczos tridiagonalization are provided for nontrivial lattice model Hamiltonians, exemplifying convergence over a range of 13 orders of magnitude. This lower bound is typically at least one order of magnitude better than Temple’s result. Its rate of convergence is comparable to that of the Ritz upper bound. It is not limited to ground states. These results complement Ritz’s upper bound and may turn the computation of lower bounds into a staple of eigenvalue and spectral problems in physics and chemistry.


Author(s):  
Chunying Ren ◽  
Dachuan Xu ◽  
Donglei Du ◽  
Min Li

Abstract In the k-means problem with penalties, we are given a data set ${\cal D} \subseteq \mathbb{R}^\ell $ of n points where each point $j \in {\cal D}$ is associated with a penalty cost p j and an integer k. The goal is to choose a set ${\rm{C}}S \subseteq {{\cal R}^\ell }$ with |CS| ≤ k and a penalized subset ${{\cal D}_p} \subseteq {\cal D}$ to minimize the sum of the total squared distance from the points in D / D p to CS and the total penalty cost of points in D p , namely $\sum\nolimits_{j \in {\cal D}\backslash {{\cal D}_p}} {d^2}(j,{\rm{C}}S) + \sum\nolimits_{j \in {{\cal D}_p}} {p_j}$ . We employ the primal-dual technique to give a pseudo-polynomial time algorithm with an approximation ratio of (6.357+ε) for the k-means problem with penalties, improving the previous best approximation ratio 19.849+∊ for this problem given by Feng et al. in Proceedings of FAW (2019).


2020 ◽  
Author(s):  
Yanze Huang ◽  
Limei Lin ◽  
Li Xu

Abstract As the size of a multiprocessor system grows, the probability that faults occur in this system increases. One measure of the reliability of a multiprocessor system is the probability that a fault-free subsystem of a certain size still exists with the presence of individual faults. In this paper, we use the probabilistic fault model to establish the subgraph reliability for $AG_n$, the $n$-dimensional alternating group graph. More precisely, we first analyze the probability $R_n^{n-1}(p)$ that at least one subgraph with dimension $n-1$ is fault-free in $AG_n$, when given a uniform probability of a single vertex being fault-free. Since subgraphs of $AG_n$ intersect in rather complicated manners, we resort to the principle of inclusion–exclusion by considering intersections of up to five subgraphs and obtain an upper bound of the probability. Then we consider the probabilistic fault model when the probability of a single vertex being fault-free is nonuniform, and we show that the upper bound under these two models is very close to the lower bound obtained in a previous result, and it is better than the upper bound deduced from that of the arrangement graph, which means that the upper bound we obtained is very tight.


1998 ◽  
Vol 6 ◽  
pp. 128
Author(s):  
O.V. Chernytska

We obtain upper bound of the best approximation of the classes $H^{\omega} [a, b]$ by piecewise-constant functions over uniform split in metrics of $L_{\varphi}[a, b]$ spaces, which are generated by continuous non-decreasing functions $\varphi$ that are equal to zero in zero. We study the classes of functions $\varphi$, for which the obtained bound is exact for all convex moduli of continuity.


Algorithms ◽  
2021 ◽  
Vol 14 (2) ◽  
pp. 65
Author(s):  
Danny Hucke ◽  
Carl Philipp Reh

A grammar-based compressor is an algorithm that receives a word and outputs a context-free grammar that only produces this word. The approximation ratio for a single input word is the size of the grammar produced for this word divided by the size of a smallest grammar for this word. The worst-case approximation ratio of a grammar-based compressor for a given word length is the largest approximation ratio over all input words of that length. In this work, we study the worst-case approximation ratio of the algorithms Greedy, RePair and LongestMatch on unary strings, i.e., strings that only make use of a single symbol. Our main contribution is to show the improved upper bound of O((logn)8·(loglogn)3) for the worst-case approximation ratio of Greedy. In addition, we also show the lower bound of 1.34847194⋯ for the worst-case approximation ratio of Greedy, and that RePair and LongestMatch have a worst-case approximation ratio of log2(3).


2007 ◽  
Vol 18-19 ◽  
pp. 347-352 ◽  
Author(s):  
Elkanah Oyetunji ◽  
Ayodeji E. Oluleye

This paper focuses on the problem of scheduling n jobs with release dates on a single machine in order to minimize the total completion time. Since the problem has been characterized as strongly NP-hard, two heuristics (HR1 and AEO) were proposed for solving the problem in polynomial time. The heuristics were compared with the best approximation algorithm for this problem to date (Best-alpha). Experimental results show that AEO performed better than the Bestalpha algorithm (selected from the literature) when the number of jobs (n) exceeds 5. This observation should prove useful in the operational dispatch of jobs in industrial production settings as well as the service sector.


1996 ◽  
Vol 06 (03) ◽  
pp. 331-344
Author(s):  
BIING-FENG WANG

In [2], a new problem called conditional cartesian product problem was introduced. The problem has a trivial time lower bound [Formula: see text], where A and B are two groups of N elements and CA,B is a subset of the cartesian product A×B satisfying some unknown condition C. Ben-Asher proposed an efficient algorithm for the problem and showed that the proposed algorithm can be performed in [Formula: see text] time [2]. As compared with the trivial time lower bound, the algorithm is not optimal. In this paper, with a better analysis, we show that the worst-case time complexity of Ben-Asher’s algorithm is [Formula: see text], which is much better than [Formula: see text]. Consequently, it can be concluded that for case the rate of growth of |CA,B| is not smaller than [Formula: see text] is a tight lower bound and Ben-Asher’s algorithm is indeed optimal.


Sign in / Sign up

Export Citation Format

Share Document