Stability of Scores for the Slosson Full-Range Intelligence Test

2007 ◽  
Vol 101 (1) ◽  
pp. 135-140 ◽  
Author(s):  
Thomas O. Williams ◽  
Ronald C. Eaves ◽  
Suzanne Woods-Groves ◽  
Gina Mariano

The test-retest stability of the Slosson Full-Range Intelligence Test by Algozzine, Eaves, Mann, and Vance was investigated with test scores from a sample of 103 students. With a mean interval of 13.7 mo. and different examiners for each of the two test administrations, the test-retest reliability coefficients for the Full-Range IQ, Verbal Reasoning, Abstract Reasoning, Quantitative Reasoning, and Memory were .93, .85, .80, .80, and .83, respectively. Mean differences from the test-retest scores were not statistically significantly different for any of the scales. Results suggest that Slosson scores are stable over time even when different examiners administer the test.

1984 ◽  
Vol 54 (3) ◽  
pp. 873-874 ◽  
Author(s):  
Norris D. Vestre

The Idea Inventory proposed as a measure of irrational thinking as defined by rational-emotive theory, was administered to two independent samples of college students on two occasions. Sample 1 ( n = 135) provided a test-retest interval of 4 wk.; Sample 2 ( n = 114), an interval of 4 to 6 wk. Indices of temporal stability, test-retest reliability coefficients (product-moment) and group changes over time, indicated satisfactory reliability for the Idea Inventory.


1997 ◽  
Vol 39 (2) ◽  
pp. 1-18 ◽  
Author(s):  
Ka Shiing Woo ◽  
Simon S K Lam

The SERVQUAL scale has been widely used by both academics and practising managers to measure service quality, but recent research has raised concerns about the reliability of SERVQUAL. Using the test-retest correlation method, this paper assesses short and long range stability of SERVQUAL. The results indicate that the SERVQUAL scale is not stable over time as revealed by the insignificant correlation between the test scores and retest scores. Although items in the expectation battery of the SERVQUAL scale remain fairly stable over time, the performance items are subject to instability even in a one-week test-retest interval. These results seem to cast doubt on the usefulness of using SERVQUAL performance items to measure service quality.


2020 ◽  
Vol 4 (1) ◽  
Author(s):  
Claudia Haberland ◽  
Anna Filonenko ◽  
Christian Seitz ◽  
Matthias Börner ◽  
Christoph Gerlinger ◽  
...  

Abstract Background To evaluate the psychometric and measurement properties of two patient-reported outcome instruments, the menstrual pictogram superabsorbent polymer-containing version 3 (MP SAP-c v3) and Uterine Fibroid Daily Bleeding Diary (UF-DBD). Test-retest reliability, criterion, construct validity, responsiveness, missingness and comparability of the MP SAP-c v3 and UF-DBD versus the alkaline hematin (AH) method and a patient global impression of severity (PGI-S) were analyzed in post hoc trial analyses. Results Analyses were based on data from up to 756 patients. The full range of MP SAP-c v3 and UF-DBD response options were used, with score distributions reflecting the cyclic character of the disease. Test-retest reliability of MP SAP-c v3 and UF-DBD scores was supported by acceptable intraclass correlation coefficients when stability was defined by the AH method and Patient Global Impression of Severity (PGI-S) scores (0.80–0.96 and 0.42–0.94, respectively). MP SAP-c v3 and UF-DBD scores demonstrated strong and moderate-to-strong correlations with menstrual blood loss assessed by the AH method. Scores increased in monotonic fashion, with greater disease severities, defined by the AH method and PGI-S scores; differences between groups were mostly statistically significant (P < 0.05). MP SAP-c v3 and UF-DBD were sensitive to changes in disease severity, defined by the AH method and PGI-S. MP SAP-c v3 and UF-DBD showed a lower frequency of missing patient data versus the AH method, and good agreement with the AH method. Conclusions This evidence supports the use of the MP SAP-c v3 and UF-DBD to assess clinical efficacy endpoints in UF phase III studies replacing the AH method.


1967 ◽  
Vol 20 (3_suppl) ◽  
pp. 1085-1086 ◽  
Author(s):  
R. A. Ruiz ◽  
H. H. Krauss

Upon re-testing with Shipley S-ILS (3 mo., 56 Ss), reliability coefficients are: Verbal, .77; Abstract, .63; Conceptual Quotient, .57; and Intelligence Quotient, .74 (all ps < .01). These coefficients and the pattern of mean-score increases suggest that the S-ILS is a “weak” measure of intellectual deterioration but a “better” measure of intellectual functioning.


1978 ◽  
Vol 42 (3_suppl) ◽  
pp. 1117-1118 ◽  
Author(s):  
Anthony F. Rotatori ◽  
Michael Epstein

The study investigated the test-retest reliability of the Slosson Intelligence Test when administered and scored by 9 special education teachers who had no experience in psychological testing. Subjects were 53 profoundly and severely retarded children, ranging in age from 5 to 16 yr. Pearson product-moment coefficients were .92 for girls, .96 for boys, and .94 for all subjects for scores on two occasions. The results support Slosson's contention that the test can be reliably administered by personnel with little psychometric knowledge.


1988 ◽  
Vol 66 (2) ◽  
pp. 503-506 ◽  
Author(s):  
John R. Reddon ◽  
David M. Gill ◽  
Stephen E. Gauk ◽  
Marita D. Maerz

26 normal, self-reported dextral subjects (12 men, 14 women) were assessed with a Purdue Pegboard 5 times at weekly intervals to evaluate temporal stability and efficacy of lateralization with this test. There was a statistically significant increase in performance over time for men on the right- and left-hand placing subtests and for women on the assemblies subtest. For men/women the test-retest reliability over the 5 sessions averaged .63/.76 for the right-hand, .64/.79 for the left-hand, .67/.81 for both-hands, .81/.83 for assemblies, and .33/.22 for the right/left-hand ratio.


Pain Medicine ◽  
2020 ◽  
Vol 21 (12) ◽  
pp. 3645-3654
Author(s):  
Louisa Picco ◽  
Melissa Middleton ◽  
Raimondo Bruno ◽  
Michala Kowalski ◽  
Suzanne Nielsen

Abstract Objective The Routine Opioid Outcome Monitoring (ROOM) tool measures outcomes with opioids using an established framework which includes domains such as pain, mood, opioid use disorder, alcohol use, and constipation. This study aims to validate and establish the test-retest reliability of the computer-administered ROOM tool. Design and Setting Cross-sectional analysis of an online sample. Subjects Participants comprised those with chronic noncancer pain who regularly used prescription opioids. Methods Participants self-completed the online ROOM tool along with other validated measures (validation questionnaire), and those who were agreeable also completed the online test-retest questionnaire approximately two weeks later. Subcomponents of the ROOM tool (i.e., pain, mood, alcohol use, opioid use disorder, and constipation) were validated against longer measures of the same construct using Pearson correlation coefficients. Intraclass correlation coefficients were used to assess the stability of the ROOM tool over time. Results A total of 324 participants completed the validation questionnaire, of whom 260 also completed the test-retest questionnaire. The opioid use disorder domain showed good sensitivity (73.6) and specificity (75.8) against the Diagnostic and Statistical Manual of Mental Disorders, 5th Edition, any opioid use disorder. All ROOM components showed moderate correlation (r = 0.55–0.73) with their longer counterparts. Test-retest reliability was fair (0.58–0.75), indicating that responses were relatively stable over time. Reliability did vary, however, based on the components being measured and how certain tools were scored. Conclusion The computer-administered ROOM tool is a valid approach for brief monitoring of outcomes with prescribed opioids in primary care settings and appears to be acceptable to people who are using prescribed opioids for chronic pain.


2020 ◽  
pp. 174702182092919 ◽  
Author(s):  
Alasdair DF Clarke ◽  
Jessica L Irons ◽  
Warren James ◽  
Andrew B Leber ◽  
Amelia R Hunt

A striking range of individual differences has recently been reported in three different visual search tasks. These differences in performance can be attributed to strategy, that is, the efficiency with which participants control their search to complete the task quickly and accurately. Here, we ask whether an individual’s strategy and performance in one search task is correlated with how they perform in the other two. We tested 64 observers and found that even though the test–retest reliability of the tasks was high, an observer’s performance and strategy in one task was not predictive of their behaviour in the other two. These results suggest search strategies are stable over time, but context-specific. To understand visual search, we therefore need to account not only for differences between individuals but also how individuals interact with the search task and context.


1982 ◽  
Vol 54 (2) ◽  
pp. 369-370 ◽  
Author(s):  
David E. Suddick ◽  
Charles L. Bowen

A 2. 5-yr. longitudinal study of the reading scales of the Stanford Achievement Test confirmed the redundancy in its subtests. Strong and stable test-retest reliability coefficients were evident and supported the use of the Stanford total reading scale for longitudinal evaluations.


Sign in / Sign up

Export Citation Format

Share Document