NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 15 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; DiStefano, Christine; Calvocoressi, Lisa; Volker, Martin – Educational and Psychological Measurement, 2022
A class of effect size indices are discussed that evaluate the degree to which two nested confirmatory factor analysis models differ from each other in terms of fit to a set of observed variables. These descriptive effect measures can be used to quantify the impact of parameter restrictions imposed in an initially considered model and are free…
Descriptors: Effect Size, Models, Measurement Techniques, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A.; Li, Tenglong – Educational and Psychological Measurement, 2018
This note extends the results in the 2016 article by Raykov, Marcoulides, and Li to the case of correlated errors in a set of observed measures subjected to principal component analysis. It is shown that when at least two measures are fallible, the probability is zero for any principal component--and in particular for the first principal…
Descriptors: Factor Analysis, Error of Measurement, Correlation, Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Nicewander, W. Alan – Educational and Psychological Measurement, 2019
This inquiry is focused on three indicators of the precision of measurement--conditional on fixed values of ?, the latent variable of item response theory (IRT). The indicators that are compared are (1) The traditional, conditional standard errors, s(eX|?) = CSEM; (2) the IRT-based conditional standard errors, s[subscript irt](eX|?)=C[subscript…
Descriptors: Measurement, Accuracy, Scores, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
McNeish, Daniel – Educational and Psychological Measurement, 2017
In behavioral sciences broadly, estimating growth models with Bayesian methods is becoming increasingly common, especially to combat small samples common with longitudinal data. Although Mplus is becoming an increasingly common program for applied research employing Bayesian methods, the limited selection of prior distributions for the elements of…
Descriptors: Models, Bayesian Statistics, Statistical Analysis, Computer Software
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko; Marcoulides, George A. – Educational and Psychological Measurement, 2018
This article outlines a procedure for examining the degree to which a common factor may be dominating additional factors in a multicomponent measuring instrument consisting of binary items. The procedure rests on an application of the latent variable modeling methodology and accounts for the discrete nature of the manifest indicators. The method…
Descriptors: Measurement Techniques, Factor Analysis, Item Response Theory, Likert Scales
Peer reviewed Peer reviewed
Direct linkDirect link
Raykov, Tenko – Educational and Psychological Measurement, 2012
A latent variable modeling approach that permits estimation of propensity scores in observational studies containing fallible independent variables is outlined, with subsequent examination of treatment effect. When at least one covariate is measured with error, it is indicated that the conventional propensity score need not possess the desirable…
Descriptors: Computation, Probability, Error of Measurement, Observation
Peer reviewed Peer reviewed
Direct linkDirect link
DeMars, Christine E. – Educational and Psychological Measurement, 2008
The graded response (GR) and generalized partial credit (GPC) models do not imply that examinees ordered by raw observed score will necessarily be ordered on the expected value of the latent trait (OEL). Factors were manipulated to assess whether increased violations of OEL also produced increased Type I error rates in differential item…
Descriptors: Test Items, Raw Scores, Test Theory, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Schumacker, Randall E.; Smith, Everett V., Jr. – Educational and Psychological Measurement, 2007
Measurement error is a common theme in classical measurement models used in testing and assessment. In classical measurement models, the definition of measurement error and the subsequent reliability coefficients differ on the basis of the test administration design. Internal consistency reliability specifies error due primarily to poor item…
Descriptors: Measurement Techniques, Error of Measurement, Item Sampling, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Wilcox, Rand R. – Educational and Psychological Measurement, 2005
It is known that nonnormality, a heteroscedastic error term, or a nonlinear association can create serious practical problems when using the conventional analysis of covariance (ANCOVA) method. This article describes a simple ANCOVA method that allows heteroscedasticity, nonnormality, nonlinearity, and multiple covariates. When standard…
Descriptors: Statistical Analysis, Error of Measurement, Measurement Techniques
Peer reviewed Peer reviewed
Whitney, Douglas R.; And Others – Educational and Psychological Measurement, 1986
This paper summarizes much of the available information concerning the reliability and validity of the Tests of General Educational Development (GED Tests). The data suggest that the results are sufficiently reliable for continued use and that the validity evidence generally supports the intended uses of the tests. (Author/LMO)
Descriptors: Correlation, Equivalency Tests, Error of Measurement, Predictive Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Doran, Harold C. – Educational and Psychological Measurement, 2005
The information function is an important statistic in item response theory (IRT) applications. Although the information function is often described as the IRT version of reliability, it differs from the classical notion of reliability from a critical perspective: replication. This article first explores the information function for the…
Descriptors: Item Response Theory, Error of Measurement, Evaluation Methods, Reliability
Peer reviewed Peer reviewed
Kingma, Johannes; Reuvekamp, Johan – Educational and Psychological Measurement, 1987
This paper describes a PASCAL program that computes both different types of transitions and learning statistics suitable for learning experiments in which a two-stage Markov model is used. The frequency counts of the different transitions are used for estimating the parameters of the two-stage Markov model. (Author/LMO)
Descriptors: Computer Software Reviews, Error of Measurement, Goodness of Fit, Input Output
Peer reviewed Peer reviewed
Direct linkDirect link
Kristjansson, Elizabeth; Aylesworth, Richard; Mcdowell, Ian; Zumbo, Bruno D. – Educational and Psychological Measurement, 2005
Item bias is a major threat to measurement validity. Methods for detecting differential item functioning (DIF) are now commonly used to identify potentially biased items. DIF detection methods for dichotomous items are well developed, but those for ordinal items are less well developed. In this article, the authors compare four methods for…
Descriptors: Discriminant Analysis, Test Bias, Multivariate Analysis, Regression (Statistics)
Peer reviewed Peer reviewed
Direct linkDirect link
Chang, Lei; Van Der Linden, Wim J.; Vos, Hans J. – Educational and Psychological Measurement, 2004
This article introduces a new test-centered standard-setting method as well as a procedure to detect intrajudge inconsistency of the method. The standard-setting method that is based on interdependent evaluations of alternative responses has judges closely evaluate the process that examinees use to solve multiple-choice items. The new method is…
Descriptors: Standard Setting (Scoring), Interrater Reliability, Foreign Countries, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Wen-Chung; Chen, Hsueh-Chu – Educational and Psychological Measurement, 2004
As item response theory (IRT) becomes popular in educational and psychological testing, there is a need of reporting IRT-based effect size measures. In this study, we show how the standardized mean difference can be generalized into such a measure. A disattenuation procedure based on the IRT test reliability is proposed to correct the attenuation…
Descriptors: Test Reliability, Rating Scales, Sample Size, Error of Measurement