An O(n3) algorithm for sorting signed genomes by reversals, transpositions, transreversals and block-interchanges

2016 ◽  
Vol 14 (01) ◽  
pp. 1640002 ◽  
Author(s):  
Shuzhi Yu ◽  
Fanchang Hao ◽  
Hon Wai Leong

We consider the problem of sorting signed permutations by reversals, transpositions, transreversals, and block-interchanges. The problem arises in the study of species evolution via large-scale genome rearrangement operations. Recently, Hao et al. gave a 2-approximation scheme called genome sorting by bridges (GSB) for solving this problem. Their result extended and unified the results of (i) He and Chen — a 2-approximation algorithm allowing reversals, transpositions, and block-interchanges (by also allowing transversals) and (ii) Hartman and Sharan — a 1.5-approximation algorithm allowing reversals, transpositions, and transversals (by also allowing block-interchanges). The GSB result is based on introduction of three bridge structures in the breakpoint graph, the L-bridge, T-bridge, and X-bridge that models goodreversal, transposition/transreversal, and block-interchange, respectively. However, the paper by Hao et al. focused on proving the 2-approximation GSB scheme and only mention a straightforward [Formula: see text] algorithm. In this paper, we give an [Formula: see text] algorithm for implementing the GSB scheme. The key idea behind our faster GSB algorithm is to represent cycles in the breakpoint graph by their canonical sequences, which greatly simplifies the search for these bridge structures. We also give some comparison results (running time and computed distances) against the original GSB implementation.

2019 ◽  
Vol 35 (14) ◽  
pp. i417-i426 ◽  
Author(s):  
Erin K Molloy ◽  
Tandy Warnow

Abstract Motivation At RECOMB-CG 2018, we presented NJMerge and showed that it could be used within a divide-and-conquer framework to scale computationally intensive methods for species tree estimation to larger datasets. However, NJMerge has two significant limitations: it can fail to return a tree and, when used within the proposed divide-and-conquer framework, has O(n5) running time for datasets with n species. Results Here we present a new method called ‘TreeMerge’ that improves on NJMerge in two ways: it is guaranteed to return a tree and it has dramatically faster running time within the same divide-and-conquer framework—only O(n2) time. We use a simulation study to evaluate TreeMerge in the context of multi-locus species tree estimation with two leading methods, ASTRAL-III and RAxML. We find that the divide-and-conquer framework using TreeMerge has a minor impact on species tree accuracy, dramatically reduces running time, and enables both ASTRAL-III and RAxML to complete on datasets (that they would otherwise fail on), when given 64 GB of memory and 48 h maximum running time. Thus, TreeMerge is a step toward a larger vision of enabling researchers with limited computational resources to perform large-scale species tree estimation, which we call Phylogenomics for All. Availability and implementation TreeMerge is publicly available on Github (http://github.com/ekmolloy/treemerge). Supplementary information Supplementary data are available at Bioinformatics online.


2021 ◽  
Vol 9 (3) ◽  
pp. 1-31
Author(s):  
Khaled Elbassioni

We consider the problem of pricing edges of a line graph so as to maximize the profit made from selling intervals to single-minded customers. An instance is given by a set E of n edges with a limited supply for each edge, and a set of m clients, where each client specifies one interval of E she is interested in and a budget B j which is the maximum price she is willing to pay for that interval. An envy-free pricing is one in which every customer is allocated an (possibly empty) interval maximizing her utility. Grandoni and Rothvoss (SIAM J. Comput. 2016) proposed a polynomial-time approximation scheme ( PTAS ) for the unlimited supply case with running time ( nm ) O ((1/ɛ) 1/ɛ ) , which was extended to the limited supply case by Grandoni and Wiese (ESA 2019). By utilizing the known hierarchical decomposition of doubling metrics , we give a PTAS with running time ( nm ) O (1/ ɛ 2 ) for the unlimited supply case. We then consider the limited supply case, and the notion of ɛ-envy-free pricing in which a customer gets an allocation maximizing her utility within an additive error of ɛ. For this case, we develop an approximation scheme with running time ( nm ) O (log 5/2 max e H e /ɛ 3 ) , where H e = B max ( e )/ B min ( e ) is the maximum ratio of the budgets of any two customers demanding edge e . This yields a PTAS in the uniform budget case, and a quasi-PTAS for the general case. The best approximation known, in both cases, for the exact envy-free pricing version is O (log c max ), where c max is the maximum item supply. Our method is based on the known hierarchical decomposition of doubling metrics, and can be applied to other problems, such as the maximum feasible subsystem problem with interval matrices.


Author(s):  
Young Hyun Kim ◽  
Eun-Gyu Ha ◽  
Kug Jin Jeon ◽  
Chena Lee ◽  
Sang-Sun Han

Objectives: This study aimed to develop a fully automated human identification method based on a convolutional neural network (CNN) with a large-scale dental panoramic radiograph (DPR) dataset. Methods: In total, 2,760 DPRs from 746 subjects who had 2 to 17 DPRs with various changes in image characteristics due to various dental treatments (tooth extraction, oral surgery, prosthetics, orthodontics, or tooth development) were collected. The test dataset included the latest DPR of each subject (746 images) and the other DPRs (2,014 images) were used for model training. A modified VGG16 model with two fully connected layers was applied for human identification. The proposed model was evaluated with rank-1, –3, and −5 accuracies, running time, and gradient-weighted class activation mapping (Grad-CAM)–applied images. Results: This model had rank-1,–3, and −5 accuracies of 82.84%, 89.14%, and 92.23%, respectively. All rank-1 accuracy values of the proposed model were above 80% regardless of changes in image characteristics. The average running time to train the proposed model was 60.9 sec per epoch, and the prediction time for 746 test DPRs was short (3.2 sec/image). The Grad-CAM technique verified that the model automatically identified humans by focusing on identifiable dental information. Conclusion: The proposed model showed good performance in fully automatic human identification despite differing image characteristics of DPRs acquired from the same patients. Our model is expected to assist in the fast and accurate identification by experts by comparing large amounts of images and proposing identification candidates at high speed.


Author(s):  
Limu Chen ◽  
Ye Xia ◽  
Dexiong Pan ◽  
Chengbin Wang

<p>Deep-learning based navigational object detection is discussed with respect to active monitoring system for anti-collision between vessel and bridge. Motion based object detection method widely used in existing anti-collision monitoring systems is incompetent in dealing with complicated and changeable waterway for its limitations in accuracy, robustness and efficiency. The video surveillance system proposed contains six modules, including image acquisition, detection, tracking, prediction, risk evaluation and decision-making, and the detection module is discussed in detail. A vessel-exclusive dataset with tons of image samples is established for neural network training and a SSD (Single Shot MultiBox Detector) based object detection model with both universality and pertinence is generated attributing to tactics of sample filtering, data augmentation and large-scale optimization, which make it capable of stable and intelligent vessel detection. Comparison results with conventional methods indicate that the proposed deep-learning method shows remarkable advantages in robustness, accuracy, efficiency and intelligence. In-situ test is carried out at Songpu Bridge in Shanghai, and the results illustrate that the method is qualified for long-term monitoring and providing information support for further analysis and decision making.</p>


Author(s):  
Andrés Abarca ◽  
Ricardo Monteiro

In recent years, the use of large scale seismic risk assessment has become increasingly popular to evaluate the fragility of a specific region to an earthquake event, through the convolution of hazard, exposure and vulnerability. These studies tend to focus on the building stock of the region and sometimes neglect the evaluation of the infrastructure, which has great importance when determining the ability of a social group to attend to a disaster and to eventually resume normal activities. This study, developed within the scope of the EU-funded project ITERATE (Improved Tools for Disaster Risk Mitigation in Algeria), focuses on the proposal of an exposure model for bridge structures in Northern Algeria. The proposed model was developed using existing national data surveys, as well as satellite information and field observations. As a result, the location and detailed characterization of a significant share of the Algeria roadway bridge inventory was developed, as well as the definition of a taxonomy that is able to classify the most common structural systems used in Algerian bridge construction. The outcome of this study serves as input to estimate the fragility of the bridge infrastructure inventory and, furthermore, to the overall risk assessment of the Northern Algerian region. Such fragility model will, in turn, enable the evaluation of earthquake scenarios at a regional scale and provide valuable information to decision makers for the implementation of risk mitigation measures.


2017 ◽  
Vol 2017 ◽  
pp. 1-18 ◽  
Author(s):  
Ali Wagdy Mohamed ◽  
Abdulaziz S. Almazyad

This paper presents Differential Evolution algorithm for solving high-dimensional optimization problems over continuous space. The proposed algorithm, namely, ANDE, introduces a new triangular mutation rule based on the convex combination vector of the triplet defined by the three randomly chosen vectors and the difference vectors between the best, better, and the worst individuals among the three randomly selected vectors. The mutation rule is combined with the basic mutation strategy DE/rand/1/bin, where the new triangular mutation rule is applied with the probability of 2/3 since it has both exploration ability and exploitation tendency. Furthermore, we propose a novel self-adaptive scheme for gradual change of the values of the crossover rate that can excellently benefit from the past experience of the individuals in the search space during evolution process which in turn can considerably balance the common trade-off between the population diversity and convergence speed. The proposed algorithm has been evaluated on the 20 standard high-dimensional benchmark numerical optimization problems for the IEEE CEC-2010 Special Session and Competition on Large Scale Global Optimization. The comparison results between ANDE and its versions and the other seven state-of-the-art evolutionary algorithms that were all tested on this test suite indicate that the proposed algorithm and its two versions are highly competitive algorithms for solving large scale global optimization problems.


2020 ◽  
Vol 12 (4) ◽  
pp. 1-20
Author(s):  
Kuan-Chung Shih ◽  
Yan-Kwang Chen ◽  
Yi-Ming Li ◽  
Chih-Teng Chen

Integrated decisions on merchandise image display and inventory planning are closely related to operational performance of online stores. A visual-attention-dependent demand (VADD) model has been developed to support online stores make the decisions. In the face of evolving products, customer needs, and competitors in an e-commerce environment, the benefits of using VADD model depend on how fast the model runs on the computer. As a result, a discrete particle swarm optimization (DPSO) method is employed to solve the VADD model. To verify the usability and effectiveness of DPSO method, it was compared with the existing methods for large-scale, medium-scale, and small-scale problems. The comparison results show that both GA and DPSO method perform well in terms of the approximation rate, but the DPSO method takes less time than the GA method. A sensitivity is conducted to determine the model parameters that influence the above comparison result.


2020 ◽  
Vol 23 (1-4) ◽  
Author(s):  
Ruth Schöbel ◽  
Robert Speck

AbstractTo extend prevailing scaling limits when solving time-dependent partial differential equations, the parallel full approximation scheme in space and time (PFASST) has been shown to be a promising parallel-in-time integrator. Similar to space–time multigrid, PFASST is able to compute multiple time-steps simultaneously and is therefore in particular suitable for large-scale applications on high performance computing systems. In this work we couple PFASST with a parallel spectral deferred correction (SDC) method, forming an unprecedented doubly time-parallel integrator. While PFASST provides global, large-scale “parallelization across the step”, the inner parallel SDC method allows integrating each individual time-step “parallel across the method” using a diagonalized local Quasi-Newton solver. This new method, which we call “PFASST with Enhanced concuRrency” (PFASST-ER), therefore exposes even more temporal concurrency. For two challenging nonlinear reaction-diffusion problems, we show that PFASST-ER works more efficiently than the classical variants of PFASST and can use more processors than time-steps.


2016 ◽  
Vol 33 (1) ◽  
pp. 7-27
Author(s):  
Mahmoud Yazdani ◽  
Hamidreza Paseh ◽  
Mostafa Sharifzadeh

Purpose – The purpose of this paper is to find a convenient contact detection algorithm in order to apply in distinct element simulation. Design/methodology/approach – Taking the most computation effort, the performance of the contact detection algorithm highly affects the running time. The algorithms investigated in this study consist of Incremental Sort-and-Update (ISU) and Double-Ended Spatial Sorting (DESS). These algorithms are based on bounding boxes, which makes the algorithm independent of blocks shapes. ISU and DESS algorithms contain sorting and updating phases. To compare the algorithms, they were implemented in identical examples of rock engineering problems with varying parameters. Findings – The results show that the ISU algorithm gives lower running time and shows better performance when blocks are unevenly distributed in both axes. The conventional ISU merges the sorting and updating phases in its naïve implementation. In this paper, a new computational technique is proposed based on parallelization in order to effectively improve the ISU algorithm and decrease the running time of numerical analysis in large-scale rock mass projects. Originality/value – In this approach, the sorting and updating phases are separated by minor changes in the algorithm. This tends to a minimal overhead of running time and a little extra memory usage and then the parallelization of phases can be applied. On the other hand, the time consumed by the updating phase of ISU algorithm is about 30 percent of the total time, which makes the parallelization justifiable. Here, according to the results for the large-scale problems, this improved technique can increase the performance of the ISU algorithm up to 20 percent.


Sign in / Sign up

Export Citation Format

Share Document