effective sample size
Recently Published Documents


TOTAL DOCUMENTS

165
(FIVE YEARS 75)

H-INDEX

24
(FIVE YEARS 3)

2022 ◽  
Vol 22 (1) ◽  
Author(s):  
Prem Shankar Mishra ◽  
Debashree Sinha ◽  
Pradeep Kumar ◽  
Shobhit Srivastava

Abstract Background Despite a significant increase in the skilled birth assisted (SBA) deliveries in India, there are huge gaps in availing maternity care services across social gradients - particularly across states and regions. Therefore, this study applies the spatial-regression model to examine the spatial distribution of SBA across districts of India. Furthermore, the study tries to understand the spatially associated population characteristics that influence the low coverage of SBA across districts of India and its regions. Methods The study used national representative cross-sectional survey data obtained from the fourth round of National Family Health Survey, conducted in 2015-16. The effective sample size was 259,469 for the analysis. Moran’s I statistics and bivariate Local Indicator for Spatial Association maps were used to understand spatial dependence and clustering of deliveries conducted by SBA coverage in districts of India. Ordinary least square, spatial lag and spatial error models were used to examine the correlates of deliveries conducted by SBA. Results Moran’s I value for SBA among women was 0.54, which represents a high spatial auto-correlation of deliveries conducted by SBA over 640 districts of India. There were 145 hotspots for deliveries conducted by SBA among women in India, which includes almost the entire southern part of India. The spatial error model revealed that with a 10% increase in exposure to mass media in a particular district, the deliveries conducted by SBA increased significantly by 2.5%. Interestingly, also with the 10% increase in the four or more antenatal care (ANC) in a particular district, the deliveries conducted by SBA increased significantly by 2.5%. Again, if there was a 10% increase of women with first birth order in a particular district, then the deliveries conducted by SBA significantly increased by 6.1%. If the district experienced an increase of 10% household as female-headed, then the deliveries conducted by SBA significantly increased by 1.4%. Conclusion The present study highlights the important role of ANC visits, mass media exposure, education, female household headship that augment the use of an SBA for delivery. Attention should be given in promoting regular ANC visits and strengthening women’s education.


2021 ◽  
Author(s):  
Quan Sun ◽  
Weifang Liu ◽  
Jonathan D Rosen ◽  
Le Huang ◽  
Rhonda G Pace ◽  
...  

Cystic fibrosis (CF) is a severe genetic disorder that can cause multiple comorbidities affecting the lungs, the pancreas, the luminal digestive system and beyond. In our previous genome-wide association studies (GWAS), we genotyped ~8,000 CF samples using a mixture of different genotyping platforms. More recently, the Cystic Fibrosis Genome Project (CFGP) performed deep (~30x) whole genome sequencing (WGS) of 5,095 samples to better understand the genetic mechanisms underlying clinical heterogeneity among CF patients. For mixtures of GWAS array and WGS data, genotype imputation has proven effective in increasing effective sample size. Therefore, we first performed imputation for the ~8,000 CF samples with GWAS array genotype using the TOPMed freeze 8 reference panel. Our results demonstrate that TOPMed can provide high-quality imputation for CF patients, boosting genomic coverage from ~0.3 - 4.2 million genotyped markers to ~11 - 43 million well-imputed markers, and significantly improving Polygenic Risk Score (PRS) prediction accuracy. Furthermore, we built a CF-specific CFGP reference panel based on WGS data of CF patients. We demonstrate that despite having ~3% the sample size of TOPMed, our CFGP reference panel can still outperform TOPMed when imputing some CF disease-causing variants, likely due to allele and haplotype differences between CF patients and general populations. We anticipate our imputed data for 4,656 samples without WGS data will benefit our subsequent genetic association studies, and the CFGP reference panel built from CF WGS samples will benefit other investigators studying CF.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Shobhit Srivastava ◽  
KM Sulaiman ◽  
Drishti Drishti ◽  
T. Muhammad

AbstractSince untreated or undertreated late-life mental disorders is associated with grave consequences including poor quality of life and increased mortality rates, this study investigates the associated factors of psychiatric disorders and its treatment seeking among older adults in India. Data for this study were derived from the Longitudinal Ageing Study in India (LASI) conducted during 2017–2018. The effective sample size was 31,464 older adults aged 60 years and above. Descriptive statistics and bivariate analysis were used to present the preliminary results. Chi-square test was used to find the significance level for bivariate association. Additionally, the Heckprobit selection model was employed to fulfil the objectives. It was revealed that about 2.8% of older adults had psychiatric disorder and of those who were suffering from psychiatric disorder, 41.3% (out of 2.8%) sought medical treatment. It was found that older adults who ever worked but currently not working, who had low level of life satisfaction, had poor self-rated health, had difficulty in activities of daily living (ADL) and instrumental ADL and had symptoms of psychological distress had higher probability of suffering from psychiatric disorder in reference to their counterparts. Older adults from oldest-old age group, who were females, from poorest wealth quintile, from Scheduled Tribe and from eastern region had lower probability of seeking treatment for psychiatric disorder in reference to their counterparts. The findings of the present study urge that greater attention be devoted at detecting and preventing late-life psychiatric disorder particularly among those who are at greater risk vis., male gender, working status as “ever worked but currently not working”, having low life satisfaction, poor SRH, ADL and IADL difficulties, higher psychological distress, belonging to higher wealth quintile and rural place of residence.


2021 ◽  
Vol 58 (2) ◽  
pp. 133-147
Author(s):  
Rownak Jahan Tamanna ◽  
M. Iftakhar Alam ◽  
Ahmed Hossain ◽  
Md Hasinur Rahaman Khan

Summary Sample size calculation is an integral part of any clinical trial design, and determining the optimal sample size for a study ensures adequate power to detect statistical significance. It is a critical step in designing a planned research protocol, since using too many participants in a study is expensive, exposing more subjects to the procedure. If a study is underpowered, it will be statistically inconclusive and may cause the whole protocol to fail. Amidst the attempt to maximize power and the underlying effort to minimize the budget, the optimization of both has become a significant issue in the determination of sample size for clinical trials in recent decades. Although it is hard to generalize a single method for sample size calculation, this study is an attempt to offer something that might be a basis for finding a permanent answer to the contradictions of sample size determination, by the use of simulation studies under simple random and cluster sampling schemes, with different sizes of power and type I error. The effective sample size is much higher when the design effect of the sampling method is smaller, particularly less than 1. Sample size increases for cluster sampling when the number of clusters increases.


Algorithms ◽  
2021 ◽  
Vol 14 (12) ◽  
pp. 351
Author(s):  
Wilson Tsakane Mongwe ◽  
Rendani Mbuvha ◽  
Tshilidzi Marwala

Markov chain Monte Carlo (MCMC) techniques are usually used to infer model parameters when closed-form inference is not feasible, with one of the simplest MCMC methods being the random walk Metropolis–Hastings (MH) algorithm. The MH algorithm suffers from random walk behaviour, which results in inefficient exploration of the target posterior distribution. This method has been improved upon, with algorithms such as Metropolis Adjusted Langevin Monte Carlo (MALA) and Hamiltonian Monte Carlo being examples of popular modifications to MH. In this work, we revisit the MH algorithm to reduce the autocorrelations in the generated samples without adding significant computational time. We present the: (1) Stochastic Volatility Metropolis–Hastings (SVMH) algorithm, which is based on using a random scaling matrix in the MH algorithm, and (2) Locally Scaled Metropolis–Hastings (LSMH) algorithm, in which the scaled matrix depends on the local geometry of the target distribution. For both these algorithms, the proposal distribution is still Gaussian centred at the current state. The empirical results show that these minor additions to the MH algorithm significantly improve the effective sample rates and predictive performance over the vanilla MH method. The SVMH algorithm produces similar effective sample sizes to the LSMH method, with SVMH outperforming LSMH on an execution time normalised effective sample size basis. The performance of the proposed methods is also compared to the MALA and the current state-of-art method being the No-U-Turn sampler (NUTS). The analysis is performed using a simulation study based on Neal’s funnel and multivariate Gaussian distributions and using real world data modeled using jump diffusion processes and Bayesian logistic regression. Although both MALA and NUTS outperform the proposed algorithms on an effective sample size basis, the SVMH algorithm has similar or better predictive performance when compared to MALA and NUTS across the various targets. In addition, the SVMH algorithm outperforms the other MCMC algorithms on a normalised effective sample size basis on the jump diffusion processes datasets. These results indicate the overall usefulness of the proposed algorithms.


Psych ◽  
2021 ◽  
Vol 3 (4) ◽  
pp. 751-779
Author(s):  
Martin Hecht ◽  
Sebastian Weirich ◽  
Steffen Zitzmann

Bayesian MCMC is a widely used model estimation technique, and software from the BUGS family, such as JAGS, have been popular for over two decades. Recently, Stan entered the market with promises of higher efficiency fueled by advanced and more sophisticated algorithms. With this study, we want to contribute empirical results to the discussion about the sampling efficiency of JAGS and Stan. We conducted three simulation studies in which we varied the number of warmup iterations, the prior informativeness, and sample sizes and employed the multi-level intercept-only model in the covariance- and mean-based and in the classic parametrization. The target outcome was MCMC efficiency measured as effective sample size per second (ESS/s). Based on our specific (and limited) study setup, we found that (1) MCMC efficiency is much higher for the covariance- and mean-based parametrization than for the classic parametrization, (2) Stan clearly outperforms JAGS when the covariance- and mean-based parametrization is used, and that (3) JAGS clearly outperforms Stan when the classic parametrization is used.


2021 ◽  
Vol 32 (1) ◽  
Author(s):  
L. Mihaela Paun ◽  
Dirk Husmeier

AbstractWe propose to accelerate Hamiltonian and Lagrangian Monte Carlo algorithms by coupling them with Gaussian processes for emulation of the log unnormalised posterior distribution. We provide proofs of detailed balance with respect to the exact posterior distribution for these algorithms, and validate the correctness of the samplers’ implementation by Geweke consistency tests. We implement these algorithms in a delayed acceptance (DA) framework, and investigate whether the DA scheme can offer computational gains over the standard algorithms. A comparative evaluation study is carried out to assess the performance of the methods on a series of models described by differential equations, including a real-world application of a 1D fluid-dynamics model of the pulmonary blood circulation. The aim is to identify the algorithm which gives the best trade-off between accuracy and computational efficiency, to be used in nonlinear DE models, which are computationally onerous due to repeated numerical integrations in a Bayesian analysis. Results showed no advantage of the DA scheme over the standard algorithms with respect to several efficiency measures based on the effective sample size for most methods and DE models considered. These gradient-driven algorithms register a high acceptance rate, thus the number of expensive forward model evaluations is not significantly reduced by the first emulator-based stage of DA. Additionally, the Lagrangian Dynamical Monte Carlo and Riemann Manifold Hamiltonian Monte Carlo tended to register the highest efficiency (in terms of effective sample size normalised by the number of forward model evaluations), followed by the Hamiltonian Monte Carlo, and the No U-turn sampler tended to be the least efficient.


2021 ◽  
Author(s):  
Akira Endo

Abstract In a paper recently published in Nature Medicine, Fukumoto et al. tried to assess the government-led school closure policy during the early phase of the COVID-19 pandemic in Japan. They compared the reported incidence rates between municipalities that had and had not implemented school closure in selected periods from March–May 2020, where they rigorously matched for potential confounders, and claimed that they found no causal effect on the incidence rates of COVID-19. However, the effective sample size (ESS) of their dataset had been substantially reduced in the process of matching due to imbalanced covariates between the treatment (i.e. with closure) and control (without) municipalities, which led to the wide uncertainty in the estimates. That said, the study title "No causal effect…" is a rather strong statement because the results are also consistent with a strong mitigating effect of school closure on incidence of COVID-19.


Sign in / Sign up

Export Citation Format

Share Document