scholarly journals Testing methods for quantifying Monte Carlo variation for categorical variables in Probabilistic Genotyping

2021 ◽  
Author(s):  
Jo-Anne Bright ◽  
Duncan Alexander Taylor ◽  
James Michael Curran ◽  
JOHN BUCKLETON

Two methods for applying a lower bound to the variation induced by the Monte Carlo effect are trialled. One of these is implemented in the widely used probabilistic genotyping system, STRmix Neither approach is giving the desired 99% coverage. In some cases the coverage is much lower than the desired 99%. The discrepancy (i.e. the distance between the LR corresponding to the desired coverage and the LR observed coverage at 99%) is not large. For example, the discrepancy of 0.23 for approach 1 suggests the lower bounds should be moved downwards by a factor of 1.7 to achieve the desired 99% coverage. Although less effective than desired these methods provide a layer of conservatism that is additional to the other layers. These other layers are from factors such as the conservatism within the sub-population model, the choice of conservative measures of co-ancestry, the consideration of relatives within the population and the resampling method used for allele probabilities, all of which tend to understate the strength of the findings.

1970 ◽  
Vol 37 (2) ◽  
pp. 267-270 ◽  
Author(s):  
D. Pnueli

A method is presented to obtain both upper and lower bound to eigenvalues when a variational formulation of the problem exists. The method consists of a systematic shift in the weight function. A detailed procedure is offered for one-dimensional problems, which makes improvement of the bounds possible, and which involves the same order of detailed computation as the Rayleigh-Ritz method. The main contribution of this method is that it yields the “other bound;” i.e., the one which cannot be obtained by the Rayleigh-Ritz method.


2021 ◽  
Author(s):  
Jo-Anne Bright ◽  
Shan-I Lee ◽  
JOHN BUCKLETON ◽  
Duncan Alexander Taylor

In previously reported work a method for applying a lower bound to the variation induced by the Monte Carlo effect was trialled. This is implemented in the widely used probabilistic genotyping system, STRmix The approach did not give the desired 99% coverage. However, the method for assigning the lower bound to the MCMC variability is only one of a number of layers of conservativism applied in a typical application. We tested all but one of these sources of variability collectively and term the result the near global coverage. The near global coverage for all tested samples was greater than 99.5% for inclusionary average LRs of known donors. This suggests that when included in the probability interval method the other layers of conservativism are more than adequate to compensate for the intermittent underperformance of the MCMC variability component. Running for extended MCMC accepts was also shown to result in improved precision.


Author(s):  
V. BOKKA ◽  
H. GURLA ◽  
S. OLARIU ◽  
J.L. SCHWING ◽  
I. STOJMENOVIĆ

The main contribution of this work is to show that a number of digital geometry problems can be solved elegantly on meshes with multiple broadcasting by using a time-optimal solution to the leftmost one problem as a basic subroutine. Consider a binary image pretiled onto a mesh with multiple broadcasting of size [Formula: see text] one pixel per processor. Our first contribution is to prove an Ω(n1/6) time lower bound for the problem of deciding whether the image contains at least one black pixel. We then obtain time lower bounds for many other digital geometry problems by reducing this fundamental problem to all the other problems of interest. Specifically, the problems that we address are: detecting whether an image contains at least one black pixel, computing the convex hull of the image, computing the diameter of an image, deciding whether a set of digital points is a digital line, computing the minimum distance between two images, deciding whether two images are linearly separable, computing the perimeter, area and width of a given image. Our second contribution is to show that the time lower bounds obtained are tight by exhibiting simple O(n1/6) time algorithms for these problems. As previously mentioned, an interesting feature of these algorithms is that they use, directly or indirectly, an algorithm for the leftmost one problem recently developed by one of the authors.


Author(s):  
Jan Feliksiak

This paper presents research results, pertinent to the maximal prime gaps bounds. Four distinct bounds are presented: Upper bound, Infimum, Supremum and finally the Lower bound. Although the Upper and Lower bounds incur a relatively high estimation error cost, the functions representing them are quite simple. This ensures, that the computation of those bounds will be straightforward and efficient. The Lower bound is essential, to address the issue of the value of the lower bound implicit constant C, in the work of Ford et al (Ford, 2016). The concluding Corollary in this paper shows, that the value of the constant C does diverge, although very slowly. The constant C, will eventually take any arbitrary value, providing that a large enough N (for p <= N) is considered. The Infimum/Supremum bounds on the other hand are computationally very demanding. Their evaluation entails computations at an extreme level of precision. In return however, we obtain bounds, which provide an extremely close approximation of the maximal prime gaps. The Infimum/Supremum estimation error gradually increases over the range of p and attains at p = 18361375334787046697 approximately the value of 0.03.


Author(s):  
Markus Kiderlen ◽  
Florian Pausinger

AbstractWe extend the notion of jittered sampling to arbitrary partitions and study the discrepancy of the related point sets. Let $${\varvec{\Omega }}=(\Omega _1,\ldots ,\Omega _N)$$ Ω = ( Ω 1 , … , Ω N ) be a partition of $$[0,1]^d$$ [ 0 , 1 ] d and let the ith point in $${{\mathcal {P}}}$$ P be chosen uniformly in the ith set of the partition (and stochastically independent of the other points), $$i=1,\ldots ,N$$ i = 1 , … , N . For the study of such sets we introduce the concept of a uniformly distributed triangular array and compare this notion to related notions in the literature. We prove that the expected $${{{\mathcal {L}}}_p}$$ L p -discrepancy, $${{\mathbb {E}}}{{{\mathcal {L}}}_p}({{\mathcal {P}}}_{\varvec{\Omega }})^p$$ E L p ( P Ω ) p , of a point set $${{\mathcal {P}}}_{\varvec{\Omega }}$$ P Ω generated from any equivolume partition $${\varvec{\Omega }}$$ Ω is always strictly smaller than the expected $${{{\mathcal {L}}}_p}$$ L p -discrepancy of a set of N uniform random samples for $$p>1$$ p > 1 . For fixed N we consider classes of stratified samples based on equivolume partitions of the unit cube into convex sets or into sets with a uniform positive lower bound on their reach. It is shown that these classes contain at least one minimizer of the expected $${{{\mathcal {L}}}_p}$$ L p -discrepancy. We illustrate our results with explicit constructions for small N. In addition, we present a family of partitions that seems to improve the expected discrepancy of Monte Carlo sampling by a factor of 2 for every N.


Author(s):  
Christian Bruch

AbstractIn this paper, we propose a method that estimates the variance of an imputed estimator in a multistage sampling design. The method is based on the rescaling bootstrap for multistage sampling introduced by Preston (Surv Methodol 35(2):227–234, 2009). In his original version, this resampling method requires that the dataset includes only complete cases and no missing values. Thus, we propose two modifications for applying this method to nonresponse and imputation. These modifications are compared to other modifications in a Monte Carlo simulation study. The results of our simulation study show that our two proposed approaches are superior to the other modifications of the rescaling bootstrap and, in many situations, produce valid estimators for the variance of the imputed estimator in multistage sampling designs.


10.37236/1188 ◽  
1994 ◽  
Vol 1 (1) ◽  
Author(s):  
Geoffrey Exoo

For $k \geq 5$, we establish new lower bounds on the Schur numbers $S(k)$ and on the k-color Ramsey numbers of $K_3$.


2020 ◽  
Vol 26 (2) ◽  
pp. 131-161
Author(s):  
Florian Bourgey ◽  
Stefano De Marco ◽  
Emmanuel Gobet ◽  
Alexandre Zhou

AbstractThe multilevel Monte Carlo (MLMC) method developed by M. B. Giles [Multilevel Monte Carlo path simulation, Oper. Res. 56 2008, 3, 607–617] has a natural application to the evaluation of nested expectations {\mathbb{E}[g(\mathbb{E}[f(X,Y)|X])]}, where {f,g} are functions and {(X,Y)} a couple of independent random variables. Apart from the pricing of American-type derivatives, such computations arise in a large variety of risk valuations (VaR or CVaR of a portfolio, CVA), and in the assessment of margin costs for centrally cleared portfolios. In this work, we focus on the computation of initial margin. We analyze the properties of corresponding MLMC estimators, for which we provide results of asymptotic optimality; at the technical level, we have to deal with limited regularity of the outer function g (which might fail to be everywhere differentiable). Parallel to this, we investigate upper and lower bounds for nested expectations as above, in the spirit of primal-dual algorithms for stochastic control problems.


Algorithms ◽  
2021 ◽  
Vol 14 (6) ◽  
pp. 164
Author(s):  
Tobias Rupp ◽  
Stefan Funke

We prove a Ω(n) lower bound on the query time for contraction hierarchies (CH) as well as hub labels, two popular speed-up techniques for shortest path routing. Our construction is based on a graph family not too far from subgraphs that occur in real-world road networks, in particular, it is planar and has a bounded degree. Additionally, we borrow ideas from our lower bound proof to come up with instance-based lower bounds for concrete road network instances of moderate size, reaching up to 96% of an upper bound given by a constructed CH. For a variant of our instance-based schema applied to some special graph classes, we can even show matching upper and lower bounds.


2021 ◽  
Vol 15 (5) ◽  
pp. 1-32
Author(s):  
Quang-huy Duong ◽  
Heri Ramampiaro ◽  
Kjetil Nørvåg ◽  
Thu-lan Dam

Dense subregion (subgraph & subtensor) detection is a well-studied area, with a wide range of applications, and numerous efficient approaches and algorithms have been proposed. Approximation approaches are commonly used for detecting dense subregions due to the complexity of the exact methods. Existing algorithms are generally efficient for dense subtensor and subgraph detection, and can perform well in many applications. However, most of the existing works utilize the state-or-the-art greedy 2-approximation algorithm to capably provide solutions with a loose theoretical density guarantee. The main drawback of most of these algorithms is that they can estimate only one subtensor, or subgraph, at a time, with a low guarantee on its density. While some methods can, on the other hand, estimate multiple subtensors, they can give a guarantee on the density with respect to the input tensor for the first estimated subsensor only. We address these drawbacks by providing both theoretical and practical solution for estimating multiple dense subtensors in tensor data and giving a higher lower bound of the density. In particular, we guarantee and prove a higher bound of the lower-bound density of the estimated subgraph and subtensors. We also propose a novel approach to show that there are multiple dense subtensors with a guarantee on its density that is greater than the lower bound used in the state-of-the-art algorithms. We evaluate our approach with extensive experiments on several real-world datasets, which demonstrates its efficiency and feasibility.


Sign in / Sign up

Export Citation Format

Share Document