Global- and Item-Level Model Fit Indices

Author(s):  
Zhuangzhuang Han ◽  
Matthew S. Johnson
2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Mahdieh Shafipoor ◽  
Hamdollah Ravand ◽  
Parviz Maftoon

AbstractThe current study compared the model fit indices, skill mastery probabilities, and classification accuracy of six Diagnostic Classification Models (DCMs): a general model (G-DINA) against five specific models (LLM, RRUM, ACDM, DINA, and DINO). To do so, the response data to the grammar and vocabulary sections of a General English Achievement Test, designed specifically for cognitive diagnostic purposes from scratch, was analyzed. The results of the test-level-model fit values obtained strong evidence in supporting the G-DINA and LLM models possessing the best model fit. In addition, the ACDM and RRUM were almost very identical to that of the G-DINA. The value indices of the DINO and DINA models were very close to each other but larger than those of the G-DINA and LLM. The model fit was also investigated at the item level, and the results revealed that model selection should be performed at the item level rather than the test level, and most of the specific models might perform well for the test. The findings of this study suggested that the relationships among the attributes of grammar and vocabulary are not ‘either-or’ compensatory or non-compensatory but a combination of both.


2018 ◽  
Vol 15 (4) ◽  
pp. 2407
Author(s):  
Yeşim Bayrakdaroglu ◽  
Dursun Katkat

The purpose of this study is to research how marketing activities of international sports organizations are performed and to develop a scale determining the effects of image management on public. The audiences of interuniversity World Winter Olympic sheld in Erzurum in 2011 participated in the research. Explanatory and Confirmatory Factor Analysis, reliability analysis were performed over the data obtained. All model fit indices of 25-item and four-factor structure of quality-image scale perceived in sports organizations applied were found to be at good level. In line with the findings obtained from the explanatory and confirmatory factor analyses and reliability analysis, it can be uttered that the scale is a valid and reliable measurement tool that can be used in field researches.


2018 ◽  
Vol 18 (3) ◽  
Author(s):  
Pablo Ezequiel Flores-Kanter ◽  
Sergio Dominguez-Lara ◽  
Mario Alberto Trógolo ◽  
Leonardo Adrián Medrano

<p>Bifactor models have gained increasing popularity in the literature concerned with personality, psychopathology and assessment. Empirical studies using bifactor analysis generally judge the estimated model using SEM model fit indices, which may lead to erroneous interpretations and conclusions. To address this problem, several researchers have proposed multiple criteria to assess bifactor models, such as a) conceptual grounds, b) overall model fit indices, and c) specific bifactor model indicators. In this article, we provide a brief summary of these criteria. An example using data gathered from a recently published research article is also provided to show how taking into account all criteria, rather than solely SEM model fit indices, may prevent researchers from drawing wrong conclusions.</p>


2020 ◽  
pp. 073428292093092 ◽  
Author(s):  
Patrícia Silva Lúcio ◽  
Joachim Vandekerckhove ◽  
Guilherme V. Polanczyk ◽  
Hugo Cogo-Moreira

The present study compares the fit of two- and three-parameter logistic (2PL and 3PL) models of item response theory in the performance of preschool children on the Raven’s Colored Progressive Matrices. The test of Raven is widely used for evaluating nonverbal intelligence of factor g. Studies comparing models with real data are scarce on the literature and this is the first to compare models of two and three parameters for the test of Raven, evaluating the informational gain of considering guessing probability. Participants were 582 Brazilian’s preschool children ( Mage = 57 months; SD = 7 months; 46% female) who responded individually to the instrument. The model fit indices suggested that the 2PL fit better to the data. The difficulty and ability parameters were similar between the models, with almost perfect correlations. Differences were observed in terms of discrimination and test information. The principle of parsimony must be called for comparing models.


2020 ◽  
pp. 001316442094289
Author(s):  
Amanda K. Montoya ◽  
Michael C. Edwards

Model fit indices are being increasingly recommended and used to select the number of factors in an exploratory factor analysis. Growing evidence suggests that the recommended cutoff values for common model fit indices are not appropriate for use in an exploratory factor analysis context. A particularly prominent problem in scale evaluation is the ubiquity of correlated residuals and imperfect model specification. Our research focuses on a scale evaluation context and the performance of four standard model fit indices: root mean square error of approximate (RMSEA), standardized root mean square residual (SRMR), comparative fit index (CFI), and Tucker–Lewis index (TLI), and two equivalence test-based model fit indices: RMSEAt and CFIt. We use Monte Carlo simulation to generate and analyze data based on a substantive example using the positive and negative affective schedule ( N = 1,000). We systematically vary the number and magnitude of correlated residuals as well as nonspecific misspecification, to evaluate the impact on model fit indices in fitting a two-factor exploratory factor analysis. Our results show that all fit indices, except SRMR, are overly sensitive to correlated residuals and nonspecific error, resulting in solutions that are overfactored. SRMR performed well, consistently selecting the correct number of factors; however, previous research suggests it does not perform well with categorical data. In general, we do not recommend using model fit indices to select number of factors in a scale evaluation framework.


2020 ◽  
Author(s):  
Amanda Kay Montoya ◽  
Michael C. Edwards

Model fit indices are being increasingly recommended and used to select the number of factors in an exploratory factor analysis. Growing evidence suggests that the recommended cutoff values for common model fit indices are not appropriate for use in an exploratory factor analysis context. A particularly prominent problem in scale evaluation is the ubiquity of correlated residuals and imperfect model specification. Our research focuses on a scale evaluation context and the performance of four standard model fit indices: root mean squared error of approximate (RMSEA), standardized root mean squared residual (SRMR), comparative fit index (CFI), and Tucker-Lewis index (TLI), and two equivalence test-based model fit indices: RMSEAt and CFIt. We use Monte Carlo simulation to generate and analyze data based on a substantive example using the positive and negative affective schedule (N = 1000). We systematically vary the number and magnitude of correlated residuals as well as nonspecific misspecification, to evaluate the impact on model fit indices in fitting a two-factor EFA. Our results show that all fit indices, except SRMR, are overly sensitive to correlated residuals and nonspecific error, resulting in solutions which are over-factored. SRMR performed well, consistently selecting the correct number of factors; however, previous research suggests it does not perform well with categorical data. In general, we do not recommend using model fit indices to select number of factors in a scale evaluation framework.


2017 ◽  
Vol 48 (1) ◽  
pp. 21-31 ◽  
Author(s):  
Angelina Wilson ◽  
Marié P Wissing ◽  
Lusilda Schutte

Although there has been extensive research on the phenomenon of stress, there is still a lack of assessment tools, especially in the South African context, that have strong theoretical underpinnings, tapping into both internal depletion of resources and the excessive external demands from the environment in the measurement of stress. The aim of this study was to validate the Setswana version of the original 30-item long form of the Stress Overload Scale as well as the 10-item short form (Stress Overload Scale–Short Form), both evaluating experienced personal vulnerability and external event load. A sample of N = 376 adults living in a rural community in the Northern Cape Province of South Africa were randomly selected to partake in the study. Emerging model fit indices of confirmatory factor analysis testing the hypothesized two-factor structure of the original Stress Overload Scale were not convincingly good. However, we found a remarkable improvement in model fit indices in the case of the Stress Overload Scale–Short Form. Concurrent validity was shown for the Stress Overload Scale–Short Form in significant correlations with depression and emotional well-being. We conclude that the Setswana version of the Stress Overload Scale–Short Form is a psychometrically sound instrument for measuring stress in the present context; however, further validation of the original Stress Overload Scale in diverse samples is necessary to provide stronger support for the hypothesized two-factor structure.


Sign in / Sign up

Export Citation Format

Share Document