Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 1 |
Descriptor
Comparative Analysis | 5 |
Reliability | 5 |
True Scores | 5 |
Statistical Analysis | 4 |
Equated Scores | 3 |
Measurement Techniques | 3 |
Correlation | 2 |
Error of Measurement | 2 |
Estimation (Mathematics) | 2 |
Ability | 1 |
Classification | 1 |
More ▼ |
Author
Brennan, Robert L. | 1 |
Hanson, Bradley A. | 1 |
Kim, Sooyeon | 1 |
Lord, Frederic M. | 1 |
Moses, Tim | 1 |
O'Connor, Edward F., Jr. | 1 |
Sullins, Walter L. | 1 |
Wingersky, Marilyn S. | 1 |
Publication Type
Journal Articles | 2 |
Reports - Research | 2 |
Reports - Evaluative | 1 |
Education Level
Audience
Researchers | 1 |
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Moses, Tim; Kim, Sooyeon – ETS Research Report Series, 2007
This study evaluated the impact of unequal reliability on test equating methods in the nonequivalent groups with anchor test (NEAT) design. Classical true score-based models were compared in terms of their assumptions about how reliability impacts test scores. These models were related to treatment of population ability differences by different…
Descriptors: Reliability, Equated Scores, Test Items, Statistical Analysis

Hanson, Bradley A.; Brennan, Robert L. – Journal of Educational Measurement, 1990
Using several data sets, the relative performance of the beta binomial model and two more general strong true score models in estimating several indices of classification consistency is examined. It appears that the beta binomial model can provide inadequate fits to raw score distributions compared to more general models. (TJH)
Descriptors: Classification, Comparative Analysis, Equations (Mathematics), Estimation (Mathematics)
Lord, Frederic M.; Wingersky, Marilyn S. – 1983
Two methods of 'equating' tests using item response theory (IRT) are compared, one using true scores, the other using the estimated distribution of observed scores. On the data studied, they yield almost indistinguishable results. This is a reassuring result for users of IRT equating methods. (Author)
Descriptors: Comparative Analysis, Equated Scores, Estimation (Mathematics), Latent Trait Theory
Sullins, Walter L. – 1971
Five-hundred dichotomously scored response patterns were generated with sequentially independent (SI) items and 500 with dependent (SD) items for each of thirty-six combinations of sampling parameters (i.e., three test lengths, three sample sizes, and four item difficulty distributions). KR-20, KR-21, and Split-Half (S-H) reliabilities were…
Descriptors: Comparative Analysis, Correlation, Error of Measurement, Item Analysis
O'Connor, Edward F., Jr. – 1970
The problem of the comparability of change scores is investigated. Change quotients and residual change scores are evaluated as alternative approaches and methods for estimating the true change and true score residual, the reliability of change scores and residuals, and procedures for constructing confidence intervals for residuals are explored.…
Descriptors: Comparative Analysis, Correlation, Equated Scores, Evaluation Methods