scholarly journals Software to calculate sample size for different designs, including targeted and non‐targeted surveys, as well as actual allocation and selection of sample units

2016 ◽  
Vol 13 (7) ◽  
Author(s):  
Tobias Verbeke ◽  
Machteld Varewyck

2021 ◽  
Author(s):  
Herdiantri Sufriyana ◽  
Yu Wei Wu ◽  
Emily Chia-Yu Su

Abstract We aimed to provide a resampling protocol for dimensional reduction resulting a few latent variables. The applicability focuses on but not limited for developing a machine learning prediction model in order to improve the number of sample size in relative to the number of candidate predictors. By this feature representation technique, one can improve generalization by preventing latent variables to overfit data used to conduct the dimensional reduction. However, this technique may warrant more computational capacity and time to conduct the procedure. The key stages consisted of derivation of latent variables from multiple resampling subsets, parameter estimation of latent variables in population, and selection of latent variables transformed by the estimated parameters.


2013 ◽  
Vol 5 (3) ◽  
pp. 235 ◽  
Author(s):  
Jeehyoung Kim ◽  
Bong Soo Seo

1995 ◽  
Vol 22 (1) ◽  
pp. 105-119 ◽  
Author(s):  
S. Devaraj Arumainayagam ◽  
V. Soundararajan

2013 ◽  
pp. 193-204
Author(s):  
Milena Stefanovic ◽  
Slobodanka Mitrovic ◽  
Dragica Obratov-Petkovic ◽  
Vera Vidakovic ◽  
Zorica Popovic ◽  
...  

In studies of population variability, particular attention has to be paid to the selection of a representative sample. The aim of this study was to assess the size of the new representative sample on the basis of the variability of chemical content of the initial sample on the example of a whitebark pine population. Statistical analysis included the content of 19 characteristics (terpene hydrocarbons and their derivates) of the initial sample of 10 elements (trees). It was determined that the new sample should contain 20 trees so that the mean value calculated from it represents a basic set with a probability higher than 95 %. Determination of the lower limit of the representative sample size that guarantees a satisfactory reliability of generalization proved to be very important in order to achieve cost efficiency of the research.


2021 ◽  
Vol 53 (2) ◽  
pp. 1-10
Author(s):  
Aparecido De Moraes ◽  
Matheus Henrique Silveira Mendes ◽  
Mauro Sérgio de Oliveira Leite ◽  
Regis De Castro Carvalho ◽  
Flávia Maria Avelar Gonçalves

The purpose of this study was to identify the ideal sample size representing a family in its potential, to identify superior families and, in parallel, determine in which spatial arrangement they may have a better accuracy in the selection of new varieties of sugarcane. For such purpose, five families of full-sibs were evaluated, each with 360 individuals, in the randomized blocks design, with three replications in three different spacing among plants in the row (50 cm, 75 cm, and 100 cm) and 150 cm between the rows. To determine the ideal sample size, as well as the better spacing for evaluation, the bootstrap method was adopted. It was observed that 100 cm spacings provided the best average for the stalk numbers, stalk diameter and for estimated weight of stalks in the stool. The spacing of 75 cm between the plants allowed a better power of discrimination among the families for all characters evaluated. At this 75 cm spacing  was also possible to identify superior families with a sample of 30 plants each plot and 3 reps in the trial. Highlights The bootstrap method was efficient to determine the ideal sample size, as well as the best spacing for evaluation. The 75-cm spacing had the highest power of discrimination among families, indicating that this spacing is the most efficient in evaluating sugarcane families for selection purposes. From all the results and considering selective accuracy as the guiding parameter for decision making, the highest values obtained considering the number of stalks and weight of stalks in the stools were found at the 75-cm spacing.


Methodology ◽  
2016 ◽  
Vol 12 (2) ◽  
pp. 61-71 ◽  
Author(s):  
Antoine Poncet ◽  
Delphine S. Courvoisier ◽  
Christophe Combescure ◽  
Thomas V. Perneger

Abstract. Many applied researchers are taught to use the t-test when distributions appear normal and/or sample sizes are large and non-parametric tests otherwise, and fear inflated error rates if the “wrong” test is used. In a simulation study (four tests: t-test, Mann-Whitney test, Robust t-test, Permutation test; seven sample sizes between 2 × 10 and 2 × 500; four distributions: normal, uniform, log-normal, bimodal; under the null and alternate hypotheses), we show that type 1 errors are well controlled in all conditions. The t-test is most powerful under the normal and the uniform distributions, the Mann-Whitney test under the lognormal distribution, and the robust t-test under the bimodal distribution. Importantly, even the t-test was more powerful under asymmetric distributions than under the normal distribution for the same effect size. It appears that normality and sample size do not matter for the selection of a test to compare two groups of same size and variance. The researcher can opt for the test that fits the scientific hypothesis the best, without fear of poor test performance.


Sign in / Sign up

Export Citation Format

Share Document