Author(s):  
David Smith ◽  
Sara Rouhani ◽  
Vibhav Gogate

We consider the problem of computing r-th order statistics, namely finding an assignment having rank r in a probabilistic graphical model. We show that the problem is NP-hard even when the graphical model has no edges (zero-treewidth models) via a reduction from the partition problem. We use this reduction, specifically a pseudo-polynomial time algorithm for number partitioning to yield a pseudo-polynomial time approximation algorithm for solving the r-th order statistics problem in zero- treewidth models. We then extend this algorithm to arbitrary graphical models by generalizing it to tree decompositions, and demonstrate via experimental evaluation on various datasets that our proposed algorithm is more accurate than sampling algorithms.


2017 ◽  
Author(s):  
Ruiming Tang ◽  
Antoine Amarilli ◽  
Pierre Senellart ◽  
Stéphane Bressan

While price and data quality should define the major tradeoff for consumers in data markets, prices are usually prescribed by vendors and data quality is not negotiable. In this paper we study a model where data quality can be traded for a discount. We focus on the case of XML documents and consider completeness as the quality dimension. In our setting, the data provider offers an XML document, and sets both the price of the document and a weight to each node of the document, depending on its potential worth. The data consumer proposes a price. If the proposed price is lower than that of the entire document, then the data consumer receives a sample, i.e., a random rooted subtree of the document whose selection depends on the discounted price and the weight of nodes. By requesting several samples, the data consumer can iteratively explore the data in the document.We present a pseudo-polynomial time algorithm to select a rooted subtree with prescribed weight uniformly at random, but show that this problem is unfortunately intractable. Yet, we are able to identify several practical cases where our algorithm runs in polynomial time. The first case is uniform random sampling of a rooted subtree with prescribed size rather than weights; the second case restricts to binary weights.As a more challenging scenario for the sampling problem, we also study the uniform sampling of a rooted subtree of prescribed weight and prescribed height. We adapt our pseudo-polynomial time algorithm to this setting and identify tractable cases.


Author(s):  
Akiyoshi Shioura

In this paper, we consider a problem of minimizing an M-convex function under an L1-distance constraint (MML1); the constraint is given by an upper bound for L1-distance between a feasible solution and a given “center.” This is motivated by a nonlinear integer programming problem for reallocation of dock capacity in a bike-sharing system discussed by Freund et al. (2017). The main aim of this paper is to better understand the combinatorial structure of the dock reallocation problem through the connection with M-convexity and show its polynomial-time solvability using this connection. For this, we first show that the dock reallocation problem and its generalizations can be reformulated in the form of (MML1). We then present a pseudo-polynomial-time algorithm for (MML1) based on the steepest descent approach. We also propose two polynomial-time algorithms for (MML1) by replacing the L1-distance constraint with a simple linear constraint. Finally, we apply the results for (MML1) to the dock reallocation problem to obtain a pseudo-polynomial-time steepest descent algorithm and also polynomial-time algorithms for this problem. For this purpose, we develop a polynomial-time algorithm for a relaxation of the dock reallocation problem by using a proximity-scaling approach, which is of interest in its own right.


10.29007/v68w ◽  
2018 ◽  
Author(s):  
Ying Zhu ◽  
Mirek Truszczynski

We study the problem of learning the importance of preferences in preference profiles in two important cases: when individual preferences are aggregated by the ranked Pareto rule, and when they are aggregated by positional scoring rules. For the ranked Pareto rule, we provide a polynomial-time algorithm that finds a ranking of preferences such that the ranked profile correctly decides all the examples, whenever such a ranking exists. We also show that the problem to learn a ranking maximizing the number of correctly decided examples (also under the ranked Pareto rule) is NP-hard. We obtain similar results for the case of weighted profiles when positional scoring rules are used for aggregation.


Sign in / Sign up

Export Citation Format

Share Document