scholarly journals Observation error model selection by information criteria vs. normality testing

2015 ◽  
Vol 59 (4) ◽  
pp. 489-504 ◽  
Author(s):  
Rüdiger Lehmann
2021 ◽  
pp. 283-294
Author(s):  
Timothy E. Essington

The chapter “Putting It Together: Fitting a Dynamic Model” provides a synthesis of the material presented in the book, by presenting a worked example that embraces concepts of density dependence, complex model dynamics, parameter estimation, and model selection using the Akaike information criterion. The example chosen is the recovery of gray wolf (Canis lupus) population in Washington State since 2008. The chapter begins by explaining how to fit an observation error model. Next, it examines how to fit a process error model. It then discusses parameter estimates and model selection. The chapter concludes with discussion of how to determine whether the population given as an example exhibits complex population dynamics.


Economies ◽  
2020 ◽  
Vol 8 (2) ◽  
pp. 49 ◽  
Author(s):  
Waqar Badshah ◽  
Mehmet Bulut

Only unstructured single-path model selection techniques, i.e., Information Criteria, are used by Bounds test of cointegration for model selection. The aim of this paper was twofold; one was to evaluate the performance of these five routinely used information criteria {Akaike Information Criterion (AIC), Akaike Information Criterion Corrected (AICC), Schwarz/Bayesian Information Criterion (SIC/BIC), Schwarz/Bayesian Information Criterion Corrected (SICC/BICC), and Hannan and Quinn Information Criterion (HQC)} and three structured approaches (Forward Selection, Backward Elimination, and Stepwise) by assessing their size and power properties at different sample sizes based on Monte Carlo simulations, and second was the assessment of the same based on real economic data. The second aim was achieved by the evaluation of the long-run relationship between three pairs of macroeconomic variables, i.e., Energy Consumption and GDP, Oil Price and GDP, and Broad Money and GDP for BRICS (Brazil, Russia, India, China and South Africa) countries using Bounds cointegration test. It was found that information criteria and structured procedures have the same powers for a sample size of 50 or greater. However, BICC and Stepwise are better at small sample sizes. In the light of simulation and real data results, a modified Bounds test with Stepwise model selection procedure may be used as it is strongly theoretically supported and avoids noise in the model selection process.


2021 ◽  
Vol 20 (3) ◽  
pp. 450-461
Author(s):  
Stanley L. Sclove

AbstractThe use of information criteria, especially AIC (Akaike’s information criterion) and BIC (Bayesian information criterion), for choosing an adequate number of principal components is illustrated.


2007 ◽  
Vol 16 (06) ◽  
pp. 1093-1113 ◽  
Author(s):  
N. S. THOMAIDIS ◽  
V. S. TZASTOUDIS ◽  
G. D. DOUNIAS

This paper compares a number of neural network model selection approaches on the basis of pricing S&P 500 stock index options. For the choice of the optimal architecture of the neural network, we experiment with a “top-down” pruning technique as well as two “bottom-up” strategies that start with simple models and gradually complicate the architecture if data indicate so. We adopt methods that base model selection on statistical hypothesis testing and information criteria and we compare their performance to a simple heuristic pruning technique. In the first set of experiments, neural network models are employed to fit the entire options surface and in the second they are used as parts of a hybrid intelligence scheme that combines a neural network model with theoretical option-pricing hints.


Entropy ◽  
2019 ◽  
Vol 21 (6) ◽  
pp. 561
Author(s):  
Miki Aoyagi

In recent years, selecting appropriate learning models has become more important with the increased need to analyze learning systems, and many model selection methods have been developed. The learning coefficient in Bayesian estimation, which serves to measure the learning efficiency in singular learning models, has an important role in several information criteria. The learning coefficient in regular models is known as the dimension of the parameter space over two, while that in singular models is smaller and varies in learning models. The learning coefficient is known mathematically as the log canonical threshold. In this paper, we provide a new rational blowing-up method for obtaining these coefficients. In the application to Vandermonde matrix-type singularities, we show the efficiency of such methods.


2020 ◽  
Vol 69 (6) ◽  
pp. 1163-1179 ◽  
Author(s):  
Kris V Parag ◽  
Christl A Donnelly

Abstract Estimating temporal changes in a target population from phylogenetic or count data is an important problem in ecology and epidemiology. Reliable estimates can provide key insights into the climatic and biological drivers influencing the diversity or structure of that population and evidence hypotheses concerning its future growth or decline. In infectious disease applications, the individuals infected across an epidemic form the target population. The renewal model estimates the effective reproduction number, R, of the epidemic from counts of observed incident cases. The skyline model infers the effective population size, N, underlying a phylogeny of sequences sampled from that epidemic. Practically, R measures ongoing epidemic growth while N informs on historical caseload. While both models solve distinct problems, the reliability of their estimates depends on p-dimensional piecewise-constant functions. If p is misspecified, the model might underfit significant changes or overfit noise and promote a spurious understanding of the epidemic, which might misguide intervention policies or misinform forecasts. Surprisingly, no transparent yet principled approach for optimizing p exists. Usually, p is heuristically set, or obscurely controlled via complex algorithms. We present a computable and interpretable p-selection method based on the minimum description length (MDL) formalism of information theory. Unlike many standard model selection techniques, MDL accounts for the additional statistical complexity induced by how parameters interact. As a result, our method optimizes p so that R and N estimates properly and meaningfully adapt to available data. It also outperforms comparable Akaike and Bayesian information criteria on several classification problems, given minimal knowledge of the parameter space, and exposes statistical similarities among renewal, skyline, and other models in biology. Rigorous and interpretable model selection is necessary if trustworthy and justifiable conclusions are to be drawn from piecewise models. [Coalescent processes; epidemiology; information theory; model selection; phylodynamics; renewal models; skyline plots]


Sign in / Sign up

Export Citation Format

Share Document