NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 12 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Joo, Seang-Hwane; Lee, Philseok; Stark, Stephen – Journal of Educational Measurement, 2018
This research derived information functions and proposed new scalar information indices to examine the quality of multidimensional forced choice (MFC) items based on the RANK model. We also explored how GGUM-RANK information, latent trait recovery, and reliability varied across three MFC formats: pairs (two response alternatives), triplets (three…
Descriptors: Item Response Theory, Models, Item Analysis, Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Shu, Lianghua; Schwarz, Richard D. – Journal of Educational Measurement, 2014
As a global measure of precision, item response theory (IRT) estimated reliability is derived for four coefficients (Cronbach's a, Feldt-Raju, stratified a, and marginal reliability). Models with different underlying assumptions concerning test-part similarity are discussed. A detailed computational example is presented for the targeted…
Descriptors: Item Response Theory, Reliability, Models, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Won-Chan – Journal of Educational Measurement, 2010
In this article, procedures are described for estimating single-administration classification consistency and accuracy indices for complex assessments using item response theory (IRT). This IRT approach was applied to real test data comprising dichotomous and polytomous items. Several different IRT model combinations were considered. Comparisons…
Descriptors: Classification, Item Response Theory, Comparative Analysis, Models
Peer reviewed Peer reviewed
Smith, Philip L. – Journal of Educational Measurement, 1981
This study explores a strategy for improving the stability of variance component estimates when only small samples are available, using a series of small, less complex generalizability (G) study designs as a surrogate for a single large design. (Author/BW)
Descriptors: Models, Reliability, Research Design, Sampling
Peer reviewed Peer reviewed
Lee, Guemin – Journal of Educational Measurement, 2002
Studied the effects of items, passages, contents, themes, and types of passages on the reliability and standard errors of measurement for complex reading comprehension tests using seven different generalizability theory models. Results suggest that passages and themes should be taken into account when evaluating the reliability of test scores for…
Descriptors: Error of Measurement, Generalizability Theory, Models, Reading Comprehension
Peer reviewed Peer reviewed
Direct linkDirect link
DeMars, Christine E. – Journal of Educational Measurement, 2006
Four item response theory (IRT) models were compared using data from tests where multiple items were grouped into testlets focused on a common stimulus. In the bi-factor model each item was treated as a function of a primary trait plus a nuisance trait due to the testlet; in the testlet-effects model the slopes in the direction of the testlet…
Descriptors: Item Response Theory, Reliability, Item Analysis, Factor Analysis
Peer reviewed Peer reviewed
Hanson, Bradley A.; Brennan, Robert L. – Journal of Educational Measurement, 1990
Using several data sets, the relative performance of the beta binomial model and two more general strong true score models in estimating several indices of classification consistency is examined. It appears that the beta binomial model can provide inadequate fits to raw score distributions compared to more general models. (TJH)
Descriptors: Classification, Comparative Analysis, Equations (Mathematics), Estimation (Mathematics)
Peer reviewed Peer reviewed
Sirotnik, Kenneth; Wellington, Roger – Journal of Educational Measurement, 1977
A single conceptual and theoretical framework for sampling any configuration of data from one or more population matrices is presented, integrating past designs and discussing implications for more general designs. The theory is based upon a generalization of the generalized symmetric mean approach for single matrix samples. (Author/CTM)
Descriptors: Analysis of Variance, Data Analysis, Item Sampling, Mathematical Models
Peer reviewed Peer reviewed
Van der Linden, Wim J. – Journal of Educational Measurement, 1982
An ignored aspect of standard setting, namely the possibility that Angoff or Nedelsky judges specify inconsistent probabilities (e.g., low probabilities for easy items but large probabilities for hard items) is explored. A latent trait method is proposed to estimate such misspecifications, and an index of consistency is defined. (Author/PN)
Descriptors: Cutting Scores, Latent Trait Theory, Mastery Tests, Mathematical Models
Peer reviewed Peer reviewed
Sykes, Robert C.; Fitzpatrick, Anne R. – Journal of Educational Measurement, 1992
Explanations for an observed change in Rasch item parameters ("b" values) from consecutive administrations of a professional licensing examination were investigated. Analysis of covariance indicated that the change was not related to item position or type. It is hypothesized that the change is attributable to shifts in curriculum…
Descriptors: Analysis of Covariance, Change, Curriculum, Higher Education
Peer reviewed Peer reviewed
Kolen, Michael J.; And Others – Journal of Educational Measurement, 1992
A procedure is described for estimating the reliability and conditional standard errors of measurement of scale scores incorporating the discrete transformation of raw scores to scale scores. The method is illustrated using a strong true score model, and practical applications are described. (SLD)
Descriptors: College Entrance Examinations, Equations (Mathematics), Error of Measurement, Estimation (Mathematics)
Peer reviewed Peer reviewed
Wolfle, Lee M.; Robertshaw, Dianne – Journal of Educational Measurement, 1983
Racial differences in the reporting accuracy of parental status characteristics by White and Black high school seniors were investigated using Joreskog's general framework for simultaneous covariance structure analyses of multiple populations. Reliability estimates for Whites were significantly higher than for Blacks due to differences in true…
Descriptors: Academic Achievement, Black Students, Educational Research, Error of Measurement