scholarly journals Sampling of cashew nuts as an aid to research for the genetic improvement of cashew tree

Author(s):  
Adroaldo Guimarães Rossetti ◽  
Francisco das Chagas Vidal Neto ◽  
Levi de Moura Barros

Abstract: The objective of this work was to estimate sample sizes of cashew (Anacardium occidentale) nuts as an aid to the genetic improvement of cashew tree. Nuts were separated by size: nuts < 17 mm were classified as size 1 (S1); 17 mm ≤ nuts < 19 mm, size 2 (S2); 19 mm ≤ nuts < 23 mm, size 3 (S3); 23 mm ≤ nuts < 25 mm, size 4 (S4); and nuts ≥ 25 mm, size 5 (S5). Sizing the sample for each stratum depends on the variance of nut size and on the error level B allowed for either the estimates or the desired precision in the results. The sample size will be larger the greater the variance of the stratum, the lower the error level allowed for the estimates, or the greater the precision desired in the results. For an error B = 0.2 g, the sample sizes of the S5, S4, and S3 strata were n5 = 42 nuts, n4 = 30 nuts, and n3 = 19 nuts, respectively. In the S5 and S4 strata, with better nut classifications, the average weight was 12.71 and 9.76 g, respectively. The Sm stratum - formed by the mixture of nuts of several sizes - should not be used as a parameter in this context due to its great variability, which is far larger than that of the other strata. Stratified sampling composed of six strata (S1, S2, ..., S5, Sm) is effective for estimating different sizes of nut samples.

2020 ◽  
Vol 42 (1) ◽  
Author(s):  
Adroaldo Guimarães Rossetti ◽  
Francisco das Chagas Vidal Neto ◽  
Levi de Moura Barros

Abstract The aim of this work was to estimate sample sizes to assist the genetic improvement of the cashew tree (Anacardium occidentale L.). Stratified sampling, comprising five strata (S5, S4, S3, S2, and S1) of five cashew clones (BRS 274, BRS 275, BRS 226, BRS 189 and CCP 76), was effective for estimating the different sample sizes of the nut. Sample size for each clone depends on the weight-nut variance, the margin of error B permitted in the estimates and the desired precision of the results. The increases in sample size with clone variance, lowered the permitted margin of error B, and increased the desired precision of the results. These clones required different sample sizes for a morphological study of the nuts. Larger nuts require larger samples for the same margin of error B. For an error B of 0.2g, the sample size for clones S5, S4 and S3 were n5 = 84, n4 = 49 and n3 = 37 nuts. For clones BRS 274 (S5) and BRS 275 (S4), with better nut classification, the mean weights were respectively 16.79 and 12.78g. Clones BRS 189 (S2) and CCP 76 (S1), with smaller nuts, have a smaller variances, s22 = 0.7638 and s12 = 1.0712, where the mean weight was 8.29 and 7.81g respectively.


2018 ◽  
Vol 7 (6) ◽  
pp. 68
Author(s):  
Karl Schweizer ◽  
Siegbert Reiß ◽  
Stefan Troche

An investigation of the suitability of threshold-based and threshold-free approaches for structural investigations of binary data is reported. Both approaches implicitly establish a relationship between binary data following the binomial distribution on one hand and continuous random variables assuming a normal distribution on the other hand. In two simulation studies we investigated: whether the fit results confirm the establishment of such a relationship, whether the differences between correct and incorrect models are retained and to what degree the sample size influences the results. Both approaches proved to establish the relationship. Using the threshold-free approach it was achieved by customary ML estimation whereas robust ML estimation was necessary in the threshold-based approach. Discrimination between correct and incorrect models was observed for both approaches. Larger CFI differences were found for the threshold-free approach than for the threshold-based approach. Dependency on sample size characterized the threshold-based approach but not the threshold-free approach. The threshold-based approach tended to perform better in large sample sizes, while the threshold-free approach performed better in smaller sample sizes.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Morteza Bitaraf Sani ◽  
Javad Zare Harofte ◽  
Mohammad Hossein Banabazi ◽  
Saeid Esmaeilkhanian ◽  
Ali Shafei Naderi ◽  
...  

AbstractFor thousands of years, camels have produced meat, milk, and fiber in harsh desert conditions. For a sustainable development to provide protein resources from desert areas, it is necessary to pay attention to genetic improvement in camel breeding. By using genotyping-by-sequencing (GBS) method we produced over 14,500 genome wide markers to conduct a genome- wide association study (GWAS) for investigating the birth weight, daily gain, and body weight of 96 dromedaries in the Iranian central desert. A total of 99 SNPs were associated with birth weight, daily gain, and body weight (p-value < 0.002). Genomic breeding values (GEBVs) were estimated with the BGLR package using (i) all 14,522 SNPs and (ii) the 99 SNPs by GWAS. Twenty-eight SNPs were associated with birth weight, daily gain, and body weight (p-value < 0.001). Annotation of the genomic region (s) within ± 100 kb of the associated SNPs facilitated prediction of 36 candidate genes. The accuracy of GEBVs was more than 0.65 based on all 14,522 SNPs, but the regression coefficients for birth weight, daily gain, and body weight were 0.39, 0.20, and 0.23, respectively. Because of low sample size, the GEBVs were predicted using the associated SNPs from GWAS. The accuracy of GEBVs based on the 99 associated SNPs was 0.62, 0.82, and 0.57 for birth weight, daily gain, and body weight. This report is the first GWAS using GBS on dromedary camels and identifies markers associated with growth traits that could help to plan breeding program to genetic improvement. Further researches using larger sample size and collaboration of the camel farmers and more profound understanding will permit verification of the associated SNPs identified in this project. The preliminary results of study show that genomic selection could be the appropriate way to genetic improvement of body weight in dromedary camels, which is challenging due to a long generation interval, seasonal reproduction, and lack of records and pedigrees.


2021 ◽  
Vol 13 (3) ◽  
pp. 368
Author(s):  
Christopher A. Ramezan ◽  
Timothy A. Warner ◽  
Aaron E. Maxwell ◽  
Bradley S. Price

The size of the training data set is a major determinant of classification accuracy. Nevertheless, the collection of a large training data set for supervised classifiers can be a challenge, especially for studies covering a large area, which may be typical of many real-world applied projects. This work investigates how variations in training set size, ranging from a large sample size (n = 10,000) to a very small sample size (n = 40), affect the performance of six supervised machine-learning algorithms applied to classify large-area high-spatial-resolution (HR) (1–5 m) remotely sensed data within the context of a geographic object-based image analysis (GEOBIA) approach. GEOBIA, in which adjacent similar pixels are grouped into image-objects that form the unit of the classification, offers the potential benefit of allowing multiple additional variables, such as measures of object geometry and texture, thus increasing the dimensionality of the classification input data. The six supervised machine-learning algorithms are support vector machines (SVM), random forests (RF), k-nearest neighbors (k-NN), single-layer perceptron neural networks (NEU), learning vector quantization (LVQ), and gradient-boosted trees (GBM). RF, the algorithm with the highest overall accuracy, was notable for its negligible decrease in overall accuracy, 1.0%, when training sample size decreased from 10,000 to 315 samples. GBM provided similar overall accuracy to RF; however, the algorithm was very expensive in terms of training time and computational resources, especially with large training sets. In contrast to RF and GBM, NEU, and SVM were particularly sensitive to decreasing sample size, with NEU classifications generally producing overall accuracies that were on average slightly higher than SVM classifications for larger sample sizes, but lower than SVM for the smallest sample sizes. NEU however required a longer processing time. The k-NN classifier saw less of a drop in overall accuracy than NEU and SVM as training set size decreased; however, the overall accuracies of k-NN were typically less than RF, NEU, and SVM classifiers. LVQ generally had the lowest overall accuracy of all six methods, but was relatively insensitive to sample size, down to the smallest sample sizes. Overall, due to its relatively high accuracy with small training sample sets, and minimal variations in overall accuracy between very large and small sample sets, as well as relatively short processing time, RF was a good classifier for large-area land-cover classifications of HR remotely sensed data, especially when training data are scarce. However, as performance of different supervised classifiers varies in response to training set size, investigating multiple classification algorithms is recommended to achieve optimal accuracy for a project.


2013 ◽  
Vol 113 (1) ◽  
pp. 221-224 ◽  
Author(s):  
David R. Johnson ◽  
Lauren K. Bachan

In a recent article, Regan, Lakhanpal, and Anguiano (2012) highlighted the lack of evidence for different relationship outcomes between arranged and love-based marriages. Yet the sample size ( n = 58) used in the study is insufficient for making such inferences. This reply discusses and demonstrates how small sample sizes reduce the utility of this research.


2012 ◽  
Vol 2012 ◽  
pp. 1-8 ◽  
Author(s):  
Louis M. Houston

We derive a general equation for the probability that a measurement falls within a range of n standard deviations from an estimate of the mean. So, we provide a format that is compatible with a confidence interval centered about the mean that is naturally independent of the sample size. The equation is derived by interpolating theoretical results for extreme sample sizes. The intermediate value of the equation is confirmed with a computational test.


2008 ◽  
Vol 32 (4) ◽  
pp. 364
Author(s):  
Leonard Tedong ◽  
Louis C. Martineau ◽  
Ali Benhaddou-Andaloussi ◽  
Hoda M. Eid ◽  
Pierre S. Haddad

Blood ◽  
2016 ◽  
Vol 128 (22) ◽  
pp. 3620-3620
Author(s):  
Sule Unal ◽  
Neslihan Kalkan ◽  
Mualla Cetin ◽  
Fatma Gumruk

Abstract Introduction: Iron overload is one of themajor complicationsof transfusion treatment in patient with thalassemia major. Deferasirox is a once-daily orally active iron chelator and long-term efficacy and safety data are being published. Herein we report the long-term follow-up data of thalassemia major patients in a single center. Methods: Of the 67 patients with thalassemia major who were under follow-up in a single center, 42 who were on deferasirox chelation for at least three years were included in the study. Patients' initial serum ferritin, ALT, creatinine, cardiac T2* and hepatic T2* values were recorded at the time of deferasirox initiation and at last visit. Deferasirox was not initiated as an iron chelator to none of the patients with a cardiac T2* value below 8 ms. All of the patients had creatinine clearance above 40 ml/minute and had serum creatinine levels within age appropriate normals at deferasirox initiation. None of the patients received any other chelations during the follow-up period. Results: Mean age of the patients were 16±9.4 years (2-33.4 years) at initiation of deferasirox and 22 (52%) were females. Eighteen (43%) of the patients were splenectomized. Median follow-up time of deferasirox chelation was 7.9 years (3-10). The median deferasirox doses at initiation of chelation and at last visit were 20.5 mg/kg/day and 30.7 mg/kg/day (7-40), respectively. Serum ferritin levels decreased significantly with deferasirox chelation (median 1969 ng/ml (516-5404) vs 1113 ng/ml (339-4003), p<0,001). We did not find statistically significant difference between the inital cardiac T2* values and the values at the last visit (median 25 .3 ms((8.7-42) vs 32 ms (6.6-42), p=0.607), despite a dramatic increase. On the other hand, hepatic T2* values did not significantly change compared to initial values, as well (median 3.7 ms (1-13.6) vs 3.3 (1-16), p=0.865). However of the patients who had cardiac T2* value between 10-20 ms, 67% was found to have T2* value above 20 ms by the end of the follow-up duration. On the other hand 53% of the patients with hepatic T2* value below 3.5 ms, had T2* values above 3.5 ms by the end of the follow-up, indicating improvement in iron stores. None of the patients exibited an adverse event that requires cessation of the drug totally, but patients exibited transient hypertransaminasemia that required transient cessation and/or dose decrement. The changes in serum ALT and serum creatinine levels at the initiation and at last visit were not significant. Conclusions: This is a a study that includes patients with a relatively long duration of follow-up. Although the cardiac T2* values improved by the end of the follow-up, this change was not found statistically significant. This can be attributed to the sample size and in a larger sample size, the change might be found significant. Additionally, the patients included in the study were composed of not only naive patients to chelation but also of the patients who were imcomplant to previous chelation and who were highly iron loaded before initiation of deferasirox. Disclosures No relevant conflicts of interest to declare.


2019 ◽  
Author(s):  
Peter E Clayson ◽  
Kaylie Amanda Carbine ◽  
Scott Baldwin ◽  
Michael J. Larson

Methodological reporting guidelines for studies of event-related potentials (ERPs) were updated in Psychophysiology in 2014. These guidelines facilitate the communication of key methodological parameters (e.g., preprocessing steps). Failing to report key parameters represents a barrier to replication efforts, and difficultly with replicability increases in the presence of small sample sizes and low statistical power. We assessed whether guidelines are followed and estimated the average sample size and power in recent research. Reporting behavior, sample sizes, and statistical designs were coded for 150 randomly-sampled articles from five high-impact journals that frequently publish ERP research from 2011 to 2017. An average of 63% of guidelines were reported, and reporting behavior was similar across journals, suggesting that gaps in reporting is a shortcoming of the field rather than any specific journal. Publication of the guidelines paper had no impact on reporting behavior, suggesting that editors and peer reviewers are not enforcing these recommendations. The average sample size per group was 21. Statistical power was conservatively estimated as .72-.98 for a large effect size, .35-.73 for a medium effect, and .10-.18 for a small effect. These findings indicate that failing to report key guidelines is ubiquitous and that ERP studies are primarily powered to detect large effects. Such low power and insufficient following of reporting guidelines represent substantial barriers to replication efforts. The methodological transparency and replicability of studies can be improved by the open sharing of processing code and experimental tasks and by a priori sample size calculations to ensure adequately powered studies.


Sign in / Sign up

Export Citation Format

Share Document