Bias-corrected Estimation of the Density of a Conditional Expectation in Nested Simulation Problems

2021 ◽  
Vol 31 (4) ◽  
pp. 1-36
Author(s):  
Ran Yang ◽  
David Kent ◽  
Daniel W. Apley ◽  
Jeremy Staum ◽  
David Ruppert

Many two-level nested simulation applications involve the conditional expectation of some response variable, where the expected response is the quantity of interest, and the expectation is with respect to the inner-level random variables, conditioned on the outer-level random variables. The latter typically represent random risk factors, and risk can be quantified by estimating the probability density function (pdf) or cumulative distribution function (cdf) of the conditional expectation. Much prior work has considered a naïve estimator that uses the empirical distribution of the sample averages across the inner-level replicates. This results in a biased estimator, because the distribution of the sample averages is over-dispersed relative to the distribution of the conditional expectation when the number of inner-level replicates is finite. Whereas most prior work has focused on allocating the numbers of outer- and inner-level replicates to balance the bias/variance tradeoff, we develop a bias-corrected pdf estimator. Our approach is based on the concept of density deconvolution, which is widely used to estimate densities with noisy observations but has not previously been considered for nested simulation problems. For a fixed computational budget, the bias-corrected deconvolution estimator allows more outer-level and fewer inner-level replicates to be used, which substantially improves the efficiency of the nested simulation.

2012 ◽  
Vol 2012 ◽  
pp. 1-10 ◽  
Author(s):  
Edis Mekić ◽  
Mihajlo Stefanović ◽  
Petar Spalević ◽  
Nikola Sekulović ◽  
Ana Stanković

The distributions of random variables are of interest in many areas of science. In this paper, the probability density function (PDF) and cumulative distribution function (CDF) of ratio of products of two random variables and random variable are derived. Random variables are described with Rayleigh, Nakagami-m, Weibull, andα-μdistributions. An application of obtained results in performance analysis of multihop wireless communication systems in different transmission environments described in detail. The proposed mathematical analysis is also complemented by various graphically presented numerical results.


Author(s):  
RONALD R. YAGER

We look at the issue of obtaining a variance like measure associated with probability distributions over ordinal sets. We call these dissonance measures. We specify some general properties desired in these dissonance measures. The centrality of the cumulative distribution function in formulating the concept of dissonance is pointed out. We introduce some specific examples of measures of dissonance.


2017 ◽  
Vol 20 (5) ◽  
pp. 939-951
Author(s):  
Amal Almarwani ◽  
Bashair Aljohani ◽  
Rasha Almutairi ◽  
Nada Albalawi ◽  
Alya O. Al Mutairi

2018 ◽  
Vol 47 (2) ◽  
pp. 53-67 ◽  
Author(s):  
Jalal Chachi

In this paper, rst a new notion of fuzzy random variables is introduced. Then, usingclassical techniques in Probability Theory, some aspects and results associated to a randomvariable (including expectation, variance, covariance, correlation coecient, etc.) will beextended to this new environment. Furthermore, within this framework, we can use thetools of general Probability Theory to dene fuzzy cumulative distribution function of afuzzy random variable.


Author(s):  
M. D. Edge

Nonparametric and semiparametric statistical methods assume models whose properties cannot be described by a finite number of parameters. For example, a linear regression model that assumes that the disturbances are independent draws from an unknown distribution is semiparametric—it includes the intercept and slope as regression parameters but has a nonparametric part, the unknown distribution of the disturbances. Nonparametric and semiparametric methods focus on the empirical distribution function, which, assuming that the data are really independent observations from the same distribution, is a consistent estimator of the true cumulative distribution function. In this chapter, with plug-in estimation and the method of moments, functionals or parameters are estimated by treating the empirical distribution function as if it were the true cumulative distribution function. Such estimators are consistent. To understand the variation of point estimates, bootstrapping is used to resample from the empirical distribution function. For hypothesis testing, one can either use a bootstrap-based confidence interval or conduct a permutation test, which can be designed to test null hypotheses of independence or exchangeability. Resampling methods—including bootstrapping and permutation testing—are flexible and easy to implement with a little programming expertise.


2017 ◽  
Author(s):  
Nash Rochman

AbstractIt is often challenging to find the right bin size when constructing a histogram to represent a noisy experimental data set. This problem is frequently faced when assessing whether a cell synchronization experiment was successful or not. In this case the goal is to determine whether the DNA content is best represented by a unimodal, indicating successful synchronization, or bimodal, indicating unsuccessful synchronization, distribution. This choice of bin size can greatly affect the interpretation of the results; however, it can be avoided by fitting the data to a cumulative distribution function (CDF). Fitting data to a CDF removes the need for bin size selection. The sorted data can also be used to reconstruct an approximate probability density function (PDF) without selecting a bin size. A simple CDF-based approach is presented and the benefits and drawbacks relative to usual methods are discussed.


2021 ◽  
Author(s):  
Nefeli Moridis ◽  
W. John Lee ◽  
Wayne Sim ◽  
Thomas Blasingame

Abstract The objective of this work is to numerically estimate the fraction of Reserves assigned to each Reserves category of the PRMS matrix through a cumulative distribution function. We selected 38 wells from a Permian Basin dataset available to Texas A&M University. Previous work has shown that Swanson's Mean, which relates the Reserves categories through a cdf of a normal distribution, is an inaccurate method to determine the relationship of the Reserves categories with asymmetric distributions. Production data are lognormally distributed, regardless of basin type, thus cannot follow the SM concept. The Gaussian Quadrature (GQ) provides a methodology to accurately estimate the fraction of Reserves that lie in 1P, 2P, and 3P categories – known as the weights. Gaussian Quadrature is a numerical integration method that uses discrete random variables and a distribution that matches the original data. For this work, we associate the lognormal cumulative distribution function (CDF) with a set of discrete random variables that replace the production data, and determine the associated probabilities. The production data for both conventional and unconventional fields are lognormally distributed, thus we expect that this methodology can be implemented in any field. To do this, we performed probabilistic decline curve analysis (DCA) using Arps’ Hyperbolic model and Monte Carlo simulation to obtain the 1P, 2P, and 3P volumes, and calculated the relative weights of each Reserves category. We performed probabilistic rate transient analysis (RTA) using a commercial software to obtain the 1P, 2P, and 3P volumes, and calculated the relative weights of each Reserves category. We implemented the 3-, 5-, and 10-point GQ to obtain the weight and percentiles for each well. Once this was completed, we validated the GQ results by calculating the percent-difference between the probabilistic DCA, RTA, and GQ results. We increase the standard deviation to account for the uncertainty of Contingent and Prospective resources and implemented 3-, 5-, and 10-point GQ to obtain the weight and percentiles for each well. This allows us to also approximate the weights of these volumes to track them through the life of a given project. The probabilistic DCA, RTA and Reserves results indicate that the SM is an inaccurate method for estimating the relative weights of each Reserves category. The 1C, 2C, 3C, and 1U, 2U, and 3U Contingent and Prospective Resources, respectively, are distributed in a similar way but with greater variance, incorporated in the standard deviation. The results show that the GQ is able to capture an accurate representation of the Reserves weights through a lognormal CDF. Based on the proposed results, we believe that the GQ is accurate and can be used to approximate the relationship between the PRMS categories. This relationship will aid in booking Reserves to the SEC because it can be recreated for any field. These distributions of Reserves and resources other than Reserves (ROTR) are important for planning and for resource inventorying. The GQ provides a measure of confidence on the prediction of the Reserves weights because of the low percent difference between the probabilistic DCA, RTA, and GQ weights. This methodology can be implemented in both conventional and unconventional fields.


2011 ◽  
Vol 62 (4) ◽  
pp. 244-248 ◽  
Author(s):  
Aleksandra Cvetković ◽  
Jelena Anastasov ◽  
Stefan Panić ◽  
Mihajlo Stefanović ◽  
Dejan Milić

Performance of Dual-Hop Relaying Over Shadowed Ricean Fading Channels In this paper, an analytical approach for evaluating performance of dual-hop cooperative link over shadowed Ricean fading channels is presented. New lower bound expressions for the probability density function (PDF), cumulative distribution function (CDF) and average bit error probability (ABEP) for system with channel state information (CSI) relay are derived. Some numerical results are presented to show behavior of performance gain for the proposed system. Analytical exact and lower bound expression for the outage probability (OP) of CSI assisted relay are obtained and required numerical results are compared.


Sign in / Sign up

Export Citation Format

Share Document