scholarly journals Cross-Validation Selection of Regularisation Parameter(s) for Semiparametric Transformation Models

Author(s):  
Senay Sokullu ◽  
Sami Stouli
2002 ◽  
Vol 14 (10) ◽  
pp. 2439-2468 ◽  
Author(s):  
Aki Vehtari ◽  
Jouko Lampinen

In this work, we discuss practical methods for the assessment, comparison, and selection of complex hierarchical Bayesian models. A natural way to assess the goodness of the model is to estimate its future predictive capability by estimating expected utilities. Instead of just making a point estimate, it is important to obtain the distribution of the expected utility estimate because it describes the uncertainty in the estimate. The distributions of the expected utility estimates can also be used to compare models, for example, by computing the probability of one model having a better expected utility than some other model. We propose an approach using cross-validation predictive densities to obtain expected utility estimates and Bayesian bootstrap to obtain samples from their distributions. We also discuss the probabilistic assumptions made and properties of two practical cross-validation methods, importance sampling and k-fold cross-validation. As illustrative examples, we use multilayer perceptron neural networks and gaussian processes with Markov chain Monte Carlo sampling in one toy problem and two challenging real-world problems.


2019 ◽  
Vol 17 (03) ◽  
pp. 1950017 ◽  
Author(s):  
Matthew Stephenson ◽  
Gerarda A. Darlington ◽  
Flavio S. Schenkel ◽  
E. James Squires ◽  
R. Ayesha Ali

Genetic selection of farm animals plays an important role in genetic improvement programs. Regularized regression methods on single nucleotide polymorphism (SNP) data from a set of candidate genes can help to identify genes that are associated with the trait of interest. This complex task must also consider the relative effect sizes on the desired trait and account for the relationships among the candidate SNPs so that selection of a SNP does not promote other undesirable traits through breeding. We present the Doubly Sparse Regression Incorporating Graphical structure (DSRIG), a novel regularized method for genetic selection that exploits the relationships among candidate SNPs to improve prediction. DSRIG was applied in the prediction of skatole and androstenone levels, two compounds known to be associated with boar taint. DSRIG was shown to provide a predictive benefit when compared to ordinary least squares (OLS) and the least absolute shrinkage and selection operator (LASSO) in a cross-validation procedure. The relative sizes of the coefficient estimates over the cross-validation procedure were compared to determine which SNPs may have the greatest impact on expression of the boar taint compounds and a consensus graph was used to infer the relationships among SNPs.


1988 ◽  
Vol 110 (1) ◽  
pp. 37-41 ◽  
Author(s):  
C. R. Dohrmann ◽  
H. R. Busby ◽  
D. M. Trujillo

Smoothing and differentiation of noisy data using spline functions requires the selection of an unknown smoothing parameter. The method of generalized cross-validation provides an excellent estimate of the smoothing parameter from the data itself even when the amount of noise associated with the data is unknown. In the present model only a single smoothing parameter must be obtained, but in a more general context the number may be larger. In an earlier work, smoothing of the data was accomplished by solving a minimization problem using the technique of dynamic programming. This paper shows how the computations required by generalized cross-validation can be performed as a simple extension of the dynamic programming formulas. The results of numerical experiments are also included.


2015 ◽  
Vol 57 (5) ◽  
pp. 808-833 ◽  
Author(s):  
Eunhee Kim ◽  
Donglin Zeng ◽  
Xiao-Hua Zhou

Sign in / Sign up

Export Citation Format

Share Document