Effect of the Number of (Multiple) Imputations on Parameter Estimates and P-Values

2010 ◽  
Author(s):  
Kevin A. Kupzyk
2014 ◽  
Author(s):  
Andrew Anand Brown

This is an implementation of the R statistical software qvalue package (Alan Dabney, John D. Storey and with assistance from Gregory R. Warnes (). qvalue: Q-value estimation for false discovery rate control. R package version 1.34.0.), designed for use with large datasets where memory or computation time is limiting. In addition to estimating p values adjusted for multiple testing, the software outputs a script which can be pasted into R to produce diagnostic plots and report parameter estimates. This program runs almost 30 times faster and requests substantially less memory than the qvalue package when analysing 10 million p values on a high performance cluster. The software has been used to control for the multiple testing of 390 million tests when analysing a full cis scan of RNA-seq exon level gene expression from the Eurobats project. The source code and links to executable files for linux and Mac OSX can be found here: https://github.com/abrown25/qvalue. Help for the package can be found by running ./largeQvalue --help.


Author(s):  
Roger Newson ◽  

multproc carries out multiple-test procedures, taking as input a list of p-values and an uncorrected critical p-value, and calculating a corrected overall critical p-value for rejection of null hypotheses. These procedures define a confidence region for a set-valued parameter, namely the set of null hypotheses that are true. They aim to control either the family-wise error rate (FWER) or the false discovery rate (FDR) at a level no greater than the uncorrected critical p-value. smileplot calls multproc and then creates a smile plot, with data points corresponding to estimated parameters, the p-values (on a reverse log scale) on the y-axis, and the parameter estimates (or another variable) on the x-axis. There are y-axis reference lines at the uncorrected and corrected overall critical p-values. The reference line for the corrected overall critical p-value, known as the parapet line, is an informal “upper confidence limit” for the set of null hypotheses that are true and defines a boundary between data mining and data dredging. A smile plot summarizes a set of multiple analyses just as a Cochrane forest plot summarizes a meta-analysis.


2018 ◽  
Vol 28 (5) ◽  
pp. 1399-1411 ◽  
Author(s):  
Susan K Mikulich-Gilbertson ◽  
Brandie D Wagner ◽  
Gary K Grunwald ◽  
Paula D Riggs ◽  
Gary O Zerbe

Medical research is often designed to investigate changes in a collection of response variables that are measured repeatedly on the same subjects. The multivariate generalized linear mixed model (MGLMM) can be used to evaluate random coefficient associations (e.g. simple correlations, partial regression coefficients) among outcomes that may be non-normal and differently distributed by specifying a multivariate normal distribution for their random effects and then evaluating the latent relationship between them. Empirical Bayes predictors are readily available for each subject from any mixed model and are observable and hence, plotable. Here, we evaluate whether second-stage association analyses of empirical Bayes predictors from a MGLMM, provide a good approximation and visual representation of these latent association analyses using medical examples and simulations. Additionally, we compare these results with association analyses of empirical Bayes predictors generated from separate mixed models for each outcome, a procedure that could circumvent computational problems that arise when the dimension of the joint covariance matrix of random effects is large and prohibits estimation of latent associations. As has been shown in other analytic contexts, the p-values for all second-stage coefficients that were determined by naively assuming normality of empirical Bayes predictors provide a good approximation to p-values determined via permutation analysis. Analyzing outcomes that are interrelated with separate models in the first stage and then associating the resulting empirical Bayes predictors in a second stage results in different mean and covariance parameter estimates from the maximum likelihood estimates generated by a MGLMM. The potential for erroneous inference from using results from these separate models increases as the magnitude of the association among the outcomes increases. Thus if computable, scatterplots of the conditionally independent empirical Bayes predictors from a MGLMM are always preferable to scatterplots of empirical Bayes predictors generated by separate models, unless the true association between outcomes is zero.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Chien-Ching Lee ◽  
Chia-Chun Chuang ◽  
Chin-Li Lu ◽  
Bo-Cheng Lai ◽  
Edmund Cheung So ◽  
...  

AbstractThe sensitivity of pneumothorax diagnosis via handheld ultrasound is low, and there is no equipment suitable for use with life-threatening tension pneumothorax in a prehospital setting. This study proposes a novel technology involving optical fibers and near-infrared spectroscopy to assist in needle thoracostomy decompression. The proposed system via the optical fibers emitted dual wavelengths of 690 and 850 nm, allowing distinction among different layers of tissue in vivo. The fundamental principle is the modified Beer–Lambert law (MBLL) which is the basis of near-infrared tissue spectroscopy. Changes in optical density corresponding to different wavelengths (690 and 850 nm) and hemoglobin parameters (levels of Hb and HbO2) were examined. The Kruskal–Wallis H test was used to compare the differences in parameter estimates among tissue layers; all p-values were < 0.001 relevant to 690 nm and 850 nm. In comparisons of Hb and HbO2 levels relative to those observed in the vein and artery, all p-values were also < 0.001. This study proposes a new optical probe to assist needle thoracostomy in a swine model. Different types of tissue can be identified by changes in optical density and hemoglobin parameters. The aid of the proposed system may yield fewer complications and a higher success rate in needle thoracostomy procedures.


Author(s):  
Michel Jacques Counotte ◽  
Shannon Axiak Flammer ◽  
Sonja Hartnack
Keyword(s):  

1999 ◽  
Vol 15 (2) ◽  
pp. 91-98 ◽  
Author(s):  
Lutz F. Hornke

Summary: Item parameters for several hundreds of items were estimated based on empirical data from several thousands of subjects. The logistic one-parameter (1PL) and two-parameter (2PL) model estimates were evaluated. However, model fit showed that only a subset of items complied sufficiently, so that the remaining ones were assembled in well-fitting item banks. In several simulation studies 5000 simulated responses were generated in accordance with a computerized adaptive test procedure along with person parameters. A general reliability of .80 or a standard error of measurement of .44 was used as a stopping rule to end CAT testing. We also recorded how often each item was used by all simulees. Person-parameter estimates based on CAT correlated higher than .90 with true values simulated. For all 1PL fitting item banks most simulees used more than 20 items but less than 30 items to reach the pre-set level of measurement error. However, testing based on item banks that complied to the 2PL revealed that, on average, only 10 items were sufficient to end testing at the same measurement error level. Both clearly demonstrate the precision and economy of computerized adaptive testing. Empirical evaluations from everyday uses will show whether these trends will hold up in practice. If so, CAT will become possible and reasonable with some 150 well-calibrated 2PL items.


Methodology ◽  
2005 ◽  
Vol 1 (2) ◽  
pp. 81-85 ◽  
Author(s):  
Stefan C. Schmukle ◽  
Jochen Hardt

Abstract. Incremental fit indices (IFIs) are regularly used when assessing the fit of structural equation models. IFIs are based on the comparison of the fit of a target model with that of a null model. For maximum-likelihood estimation, IFIs are usually computed by using the χ2 statistics of the maximum-likelihood fitting function (ML-χ2). However, LISREL recently changed the computation of IFIs. Since version 8.52, IFIs reported by LISREL are based on the χ2 statistics of the reweighted least squares fitting function (RLS-χ2). Although both functions lead to the same maximum-likelihood parameter estimates, the two χ2 statistics reach different values. Because these differences are especially large for null models, IFIs are affected in particular. Consequently, RLS-χ2 based IFIs in combination with conventional cut-off values explored for ML-χ2 based IFIs may lead to a wrong acceptance of models. We demonstrate this point by a confirmatory factor analysis in a sample of 2449 subjects.


Methodology ◽  
2015 ◽  
Vol 11 (3) ◽  
pp. 89-99 ◽  
Author(s):  
Leslie Rutkowski ◽  
Yan Zhou

Abstract. Given a consistent interest in comparing achievement across sub-populations in international assessments such as TIMSS, PIRLS, and PISA, it is critical that sub-population achievement is estimated reliably and with sufficient precision. As such, we systematically examine the limitations to current estimation methods used by these programs. Using a simulation study along with empirical results from the 2007 cycle of TIMSS, we show that a combination of missing and misclassified data in the conditioning model induces biases in sub-population achievement estimates, the magnitude and degree to which can be readily explained by data quality. Importantly, estimated biases in sub-population achievement are limited to the conditioning variable with poor-quality data while other sub-population achievement estimates are unaffected. Findings are generally in line with theory on missing and error-prone covariates. The current research adds to a small body of literature that has noted some of the limitations to sub-population estimation.


2008 ◽  
Author(s):  
Geoff Cumming ◽  
Jerry Lai ◽  
Fiona Fidler
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document