scholarly journals Efficiency of Bayesian Approaches in Quantile Regression with Small Sample Size

Author(s):  
Neveen Sayed-Ahmed

Quantile regression is a statistical technique intended to estimate, and conduct inference about the conditional quantile functions. Just as the classical linear regression methods estimate model for the conditional mean function, quantile regression offers a mechanism for estimating models for the conditional median function, and the full range of other conditional quantile functions. In the Bayesian approach to variable selection prior distributions representing the subjective beliefs about the parameters are assigned to the regression coefficients. The estimation of parameters and the selection of the best subset of variables is accomplished by using adaptive lasso quantile regression. In this paper we describe, compare, and apply the two suggested Bayesian approaches. The two suggested Bayesian suggested approaches are used to select the best subset of variables and estimate the parameters of the quantile regression equation when small sample sizes are used.  Simulations show that the proposed approaches are very competitive in terms of variable selection, estimation accuracy and efficient when small sample sizes are used.   

PLoS ONE ◽  
2018 ◽  
Vol 13 (6) ◽  
pp. e0197910 ◽  
Author(s):  
Alexander Kirpich ◽  
Elizabeth A. Ainsworth ◽  
Jessica M. Wedow ◽  
Jeremy R. B. Newman ◽  
George Michailidis ◽  
...  

2021 ◽  
Vol 37 (3) ◽  
pp. 771-789
Author(s):  
Jairo Fúquene-Patiño ◽  
César Cristancho ◽  
Mariana Ospina ◽  
Domingo Morales Gonzalez

Abstract This article proposes a new methodology for estimating the proportions of households that had experience of international migration at the municipal level in Colombia. The Colombian National Statistical Office usually produces estimations of internal migration based on the results of population censuses, but there is a lack of disaggregated information about the main small areas of origin of the population that emigrates from Colombia. The proposed methodology uses frequentist and Bayesian approaches based on a Fay-Herriot model and is illustrated by one example with a dependent variable from the Demographic and Health Survey 2015 and covariables available from the population census 2005. The proposed alternative produces proportion estimates that are consistent with sample sizes and the main internal immigration trends in Colombia. Additionally, the estimated coefficients of variation are lower than 20% for municipalities for both frequentist and Bayesian approaches and large demographically-relevant capital cities and therefore estimates may be considered to be reliable. Finally, we illustrate how the proposed alternative leads to important reductions of the estimated coefficients of variations for the areas with very small sample sizes.


2018 ◽  
Author(s):  
Christopher Chabris ◽  
Patrick Ryan Heck ◽  
Jaclyn Mandart ◽  
Daniel Jacob Benjamin ◽  
Daniel J. Simons

Williams and Bargh (2008) reported that holding a hot cup of coffee caused participants to judge a person’s personality as warmer, and that holding a therapeutic heat pad caused participants to choose rewards for other people rather than for themselves. These experiments featured large effects (r = .28 and .31), small sample sizes (41 and 53 participants), and barely statistically significant results. We attempted to replicate both experiments in field settings with more than triple the sample sizes (128 and 177) and double-blind procedures, but found near-zero effects (r = –.03 and .02). In both cases, Bayesian analyses suggest there is substantially more evidence for the null hypothesis of no effect than for the original physical warmth priming hypothesis.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Florent Le Borgne ◽  
Arthur Chatton ◽  
Maxime Léger ◽  
Rémi Lenain ◽  
Yohann Foucher

AbstractIn clinical research, there is a growing interest in the use of propensity score-based methods to estimate causal effects. G-computation is an alternative because of its high statistical power. Machine learning is also increasingly used because of its possible robustness to model misspecification. In this paper, we aimed to propose an approach that combines machine learning and G-computation when both the outcome and the exposure status are binary and is able to deal with small samples. We evaluated the performances of several methods, including penalized logistic regressions, a neural network, a support vector machine, boosted classification and regression trees, and a super learner through simulations. We proposed six different scenarios characterised by various sample sizes, numbers of covariates and relationships between covariates, exposure statuses, and outcomes. We have also illustrated the application of these methods, in which they were used to estimate the efficacy of barbiturates prescribed during the first 24 h of an episode of intracranial hypertension. In the context of GC, for estimating the individual outcome probabilities in two counterfactual worlds, we reported that the super learner tended to outperform the other approaches in terms of both bias and variance, especially for small sample sizes. The support vector machine performed well, but its mean bias was slightly higher than that of the super learner. In the investigated scenarios, G-computation associated with the super learner was a performant method for drawing causal inferences, even from small sample sizes.


2013 ◽  
Vol 113 (1) ◽  
pp. 221-224 ◽  
Author(s):  
David R. Johnson ◽  
Lauren K. Bachan

In a recent article, Regan, Lakhanpal, and Anguiano (2012) highlighted the lack of evidence for different relationship outcomes between arranged and love-based marriages. Yet the sample size ( n = 58) used in the study is insufficient for making such inferences. This reply discusses and demonstrates how small sample sizes reduce the utility of this research.


2016 ◽  
Vol 41 (5) ◽  
pp. 472-505 ◽  
Author(s):  
Elizabeth Tipton ◽  
Kelly Hallberg ◽  
Larry V. Hedges ◽  
Wendy Chan

Background: Policy makers and researchers are frequently interested in understanding how effective a particular intervention may be for a specific population. One approach is to assess the degree of similarity between the sample in an experiment and the population. Another approach is to combine information from the experiment and the population to estimate the population average treatment effect (PATE). Method: Several methods for assessing the similarity between a sample and population currently exist as well as methods estimating the PATE. In this article, we investigate properties of six of these methods and statistics in the small sample sizes common in education research (i.e., 10–70 sites), evaluating the utility of rules of thumb developed from observational studies in the generalization case. Result: In small random samples, large differences between the sample and population can arise simply by chance and many of the statistics commonly used in generalization are a function of both sample size and the number of covariates being compared. The rules of thumb developed in observational studies (which are commonly applied in generalization) are much too conservative given the small sample sizes found in generalization. Conclusion: This article implies that sharp inferences to large populations from small experiments are difficult even with probability sampling. Features of random samples should be kept in mind when evaluating the extent to which results from experiments conducted on nonrandom samples might generalize.


2018 ◽  
Vol 15 ◽  
pp. 1-5 ◽  
Author(s):  
M.G.M. Kok ◽  
M.W.J. de Ronde ◽  
P.D. Moerland ◽  
J.M. Ruijter ◽  
E.E. Creemers ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document