A method for exploring attitude systems by combining Belief Network Analysis and Item Response Theory (ResIN)

2021 ◽  
Author(s):  
Dino Carpentras ◽  
Adrian Lueders ◽  
Michael Quayle

Belief network analysis (BNA) is a new class of methods with strong potential to research the organization and development of abstract meaning systems. By mapping the attitude system, this method provides a more profound understanding of often “fuzzy” concepts such as ideologies, worldviews, and norm systems. BNA therefore holds potential implications for a plethora of socially relevant issues. For example, by informing the architecture of extreme belief sets or lines of conflict underlying partisan polarization. Despite the huge potential of this approach, it has some major limitations. Indeed, BNA methods start from the simplistic assumption that opposing groups should be perfectly symmetric in their attitudes (e.g. the more democrats are positive, the more republicans should be negative about each topic). Another important aspect of BNA methods is that they are often grounded on new, instead of well-established theories. This sometimes results in problems of interpretation and reliability of the results.In this article, we introduce a new method by combining BNA with item response theory (IRT). We refer to it as the Response-Item Network (or ResIN) method. This method has the advantage of being grounded in the well-developed psychometrics literature. Furthermore, it allows us to analyze attitudes from different groups without assuming symmetric behavior. This allows us to explore more deeply relationships and differences in the attitude system.Besides validating ResIN using IRT, we also test this method on real data, showing that it produces new insights compared to both classical BNA and IRT. Indeed, we are able to easily distinguish attitudes which belong to the republican and to the democrat side, even in counter-intuitive situations. We furthermore validated the reliability of these results by relying on additional data, such as self-identification measurements.

2021 ◽  
Vol 10 (3) ◽  
pp. 388
Author(s):  
Melissa Alves Braga de Oliveira ◽  
Euclides de Mendonça Filho ◽  
Alicia Carissimi ◽  
Luciene Lima dos Santos Garay ◽  
Marina Scop ◽  
...  

Background: Recent studies with the mood rhythm instrument (MRhI) have shown that the presence of recurrent daily peaks in specific mood symptoms are significantly associated with increased risk of psychiatric disorders. Using a large sample collected in Brazil, Spain, and Canada, we aimed to analyze which MRhI items maintained good psychometric properties across cultures. As a secondary aim, we used network analysis to visualize the strength of the association between the MRhI items. Methods: Adults (n = 1275) between 18–60 years old from Spain (n = 458), Brazil (n = 415), and Canada (n = 401) completed the MRhI and the self-reporting questionnaire (SRQ-20). Psychometric analyses followed three steps: Factor analysis, item response theory, and network analysis. Results: The factor analysis indicated the retention of three factors that grouped the MRhI items into cognitive, somatic, and affective domains. The item response theory analysis suggested the exclusion of items that displayed a significant divergence in difficulty measures between countries. Finally, the network analysis revealed a structure where sleepiness plays a central role in connecting the three domains. These psychometric analyses enabled a psychometric-based refinement of the MRhI, where the 11 items with good properties across cultures were kept in a shorter, revised MRhI version (MRhI-r). Limitations: Participants were mainly university students and, as we did not conduct a formal clinical assessment, any potential correlations (beyond the validated SRQ) cannot be ascertained. Conclusions: The MRhI-r is a novel tool to investigate self-perceived rhythmicity of mood-related symptoms and behaviors, with good psychometric properties across multiple cultures.


2020 ◽  
Vol 44 (5) ◽  
pp. 362-375
Author(s):  
Tyler Strachan ◽  
Edward Ip ◽  
Yanyan Fu ◽  
Terry Ackerman ◽  
Shyh-Huei Chen ◽  
...  

As a method to derive a “purified” measure along a dimension of interest from response data that are potentially multidimensional in nature, the projective item response theory (PIRT) approach requires first fitting a multidimensional item response theory (MIRT) model to the data before projecting onto a dimension of interest. This study aims to explore how accurate the PIRT results are when the estimated MIRT model is misspecified. Specifically, we focus on using a (potentially misspecified) two-dimensional (2D)-MIRT for projection because of its advantages, including interpretability, identifiability, and computational stability, over higher dimensional models. Two large simulation studies (I and II) were conducted. Both studies examined whether the fitting of a 2D-MIRT is sufficient to recover the PIRT parameters when multiple nuisance dimensions exist in the test items, which were generated, respectively, under compensatory MIRT and bifactor models. Various factors were manipulated, including sample size, test length, latent factor correlation, and number of nuisance dimensions. The results from simulation studies I and II showed that the PIRT was overall robust to a misspecified 2D-MIRT. Smaller third and fourth simulation studies were done to evaluate recovery of the PIRT model parameters when the correctly specified higher dimensional MIRT or bifactor model was fitted with the response data. In addition, a real data set was used to illustrate the robustness of PIRT.


2019 ◽  
Vol 45 (3) ◽  
pp. 274-296
Author(s):  
Yang Liu ◽  
Xiaojing Wang

Parametric methods, such as autoregressive models or latent growth modeling, are usually inflexible to model the dependence and nonlinear effects among the changes of latent traits whenever the time gap is irregular and the recorded time points are individually varying. Often in practice, the growth trend of latent traits is subject to certain monotone and smooth conditions. To incorporate such conditions and to alleviate the strong parametric assumption on regressing latent trajectories, a flexible nonparametric prior has been introduced to model the dynamic changes of latent traits for item response theory models over the study period. Suitable Bayesian computation schemes are developed for such analysis of the longitudinal and dichotomous item responses. Simulation studies and a real data example from educational testing have been used to illustrate our proposed methods.


2019 ◽  
Vol 45 (3) ◽  
pp. 339-368 ◽  
Author(s):  
Chun Wang ◽  
Steven W. Nydick

Recent work on measuring growth with categorical outcome variables has combined the item response theory (IRT) measurement model with the latent growth curve model and extended the assessment of growth to multidimensional IRT models and higher order IRT models. However, there is a lack of synthetic studies that clearly evaluate the strength and limitations of different multilevel IRT models for measuring growth. This study aims to introduce the various longitudinal IRT models, including the longitudinal unidimensional IRT model, longitudinal multidimensional IRT model, and longitudinal higher order IRT model, which cover a broad range of applications in education and social science. Following a comparison of the parameterizations, identification constraints, strengths, and weaknesses of the different models, a real data example is provided to illustrate the application of different longitudinal IRT models to model students’ growth trajectories on multiple latent abilities.


2016 ◽  
Vol 50 (6) ◽  
pp. 165
Author(s):  
Tetiana V. Lisova

The necessary condition for the presence of biased assessment by some test is differential item functioning in different groups of test takers. The ideas of some statistical methods for detecting Differential Item Functioning are described in the given article. They were developed in the framework of the main approaches to modeling test results: using contingency tables, regression models, multidimensional models and models of Item Response Theory. The Mantel-Haenszel procedure, logistic regression method, SIBTEST and Item Response Theory Likelihood Ratio Test are considered. The characteristics of each method and conditions of their application are specified. Overview of existing free software tools implementing these methods is carried out. Comparisons of these methods are conducted on the example of real data. Also notes that it is appropriate to use several methods simultaneously to reduce the risk of false conclusions.


2019 ◽  
Vol 45 (4) ◽  
pp. 383-402
Author(s):  
Paul A. Jewsbury ◽  
Peter W. van Rijn

In large-scale educational assessment data consistent with a simple-structure multidimensional item response theory (MIRT) model, where every item measures only one latent variable, separate unidimensional item response theory (UIRT) models for each latent variable are often calibrated for practical reasons. While this approach can be valid for data from a linear test, unacceptable item parameter estimates are obtained when data arise from a multistage test (MST). We explore this situation from a missing data perspective and show mathematically that MST data will be problematic for calibrating multiple UIRT models but not MIRT models. This occurs because some items that were used in the routing decision are excluded from the separate UIRT models, due to measuring a different latent variable. Both simulated and real data from the National Assessment of Educational Progress are used to further confirm and explore the unacceptable item parameter estimates. The theoretical and empirical results confirm that only MIRT models are valid for item calibration of multidimensional MST data.


2020 ◽  
Vol 80 (4) ◽  
pp. 665-694
Author(s):  
Ken A. Fujimoto ◽  
Sabina R. Neugebauer

Although item response theory (IRT) models such as the bifactor, two-tier, and between-item-dimensionality IRT models have been devised to confirm complex dimensional structures in educational and psychological data, they can be challenging to use in practice. The reason is that these models are multidimensional IRT (MIRT) models and thus are highly parameterized, making them only suitable for data provided by large samples. Unfortunately, many educational and psychological studies are conducted on a small scale, leaving the researchers without the necessary MIRT models to confirm the hypothesized structures in their data. To address the lack of modeling options for these researchers, we present a general Bayesian MIRT model based on adaptive informative priors. Simulations demonstrated that our MIRT model could be used to confirm a two-tier structure (with two general and six specific dimensions), a bifactor structure (with one general and six specific dimensions), and a between-item six-dimensional structure in rating scale data representing sample sizes as small as 100. Although our goal was to provide a general MIRT model suitable for smaller samples, the simulations further revealed that our model was applicable to larger samples. We also analyzed real data from 121 individuals to illustrate that the findings of our simulations are relevant to real situations.


2016 ◽  
Vol 59 (2) ◽  
pp. 281-289 ◽  
Author(s):  
Guido Makransky ◽  
Philip S. Dale ◽  
Philip Havmose ◽  
Dorthe Bleses

Purpose This study investigated the feasibility and potential validity of an item response theory (IRT)–based computerized adaptive testing (CAT) version of the MacArthur–Bates Communicative Development Inventory: Words & Sentences (CDI:WS; Fenson et al., 2007) vocabulary checklist, with the objective of reducing length while maintaining measurement precision. Method Parent-reported vocabulary for the American CDI:WS norming sample consisting of 1,461 children between the ages of 16 and 30 months was used to investigate the fit of the items to the 2-parameter logistic IRT model and to simulate CDI-CAT versions with 400, 200, 100, 50, 25, 10, and 5 items. Results All but 14 items fit the 2-parameter logistic IRT model, and real data simulations of CDI-CATs with at least 50 items recovered full CDI scores with correlations over .95. Furthermore, the CDI-CATs with at least 50 items had similar correlations with age and socioeconomic status as the full CDI:WS. Conclusion These results provide strong evidence that a CAT version of the CDI:WS has the potential to reduce length while maintaining the accuracy and precision of the full instrument.


Sign in / Sign up

Export Citation Format

Share Document