Publication Date
In 2025 | 39 |
Since 2024 | 192 |
Since 2021 (last 5 years) | 495 |
Since 2016 (last 10 years) | 996 |
Since 2006 (last 20 years) | 2028 |
Descriptor
Source
Author
Publication Type
Education Level
Audience
Researchers | 93 |
Practitioners | 23 |
Teachers | 22 |
Policymakers | 10 |
Administrators | 5 |
Students | 4 |
Counselors | 2 |
Parents | 2 |
Community | 1 |
Location
United States | 47 |
Germany | 42 |
Australia | 34 |
Canada | 27 |
Turkey | 27 |
California | 22 |
United Kingdom (England) | 20 |
Netherlands | 18 |
China | 16 |
New York | 15 |
United Kingdom | 15 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Does not meet standards | 1 |

Zwick, Rebecca – Journal of Educational Statistics, 1990
Use of the Mantel-Haenszel procedure as a test for differential item functioning under the Rasch model of item-response theory is examined. Results of the procedure cannot be generalized to the class of items for which item-response functions are monotonic and local independence holds. (TJH)
Descriptors: Demography, Equations (Mathematics), Error of Measurement, Item Bias

Wilcox, Rand R. – Journal of Educational Statistics, 1990
Recently, C. E. McCulloch (1987) suggested a modification of the Morgan-Pitman test for comparing the variances of two dependent groups. This paper demonstrates that there are situations where the procedure is not robust. A subsample approach, similar to the Box-Scheffe test, and the Sandvik-Olsson procedure are also assessed. (TJH)
Descriptors: Comparative Analysis, Equations (Mathematics), Error of Measurement, Mathematical Models

Norcini, John J.; And Others – Evaluation and the Health Professions, 1990
Aggregate scoring was applied to a recertifying examination for medical professionals to generate an answer key and allow comparison of peer examinees. Results for 1,927 candidates for recertification indicate considerable agreement between the traditional answer key and the aggregate answer key. (TJH)
Descriptors: Answer Keys, Criterion Referenced Tests, Error of Measurement, Generalizability Theory

Cahan, Sorel; Cohen, Nora – Educational and Psychological Measurement, 1990
A solution is offered to problems associated with the inequality in the manipulability of probabilities of classification errors of masters versus nonmasters, based on competency test results. Eschewing the typical arbitrary establishment of observed-score standards below 100 percent, the solution incorporates a self-correction of wrong answers.…
Descriptors: Classification, Error of Measurement, Mastery Tests, Minimum Competency Testing

Algina, James; Tang, Kezhen L. – Journal of Educational Statistics, 1988
For Y. Yao's and G. S. James' tests, Type I error rates were estimated for various combinations of the number of variables, sample-size and sample-size-to-variables ratios, and heteroscedasticity. These tests are alternatives to Hotelling's T(sup 2) and are intended for use when variance-covariance matrices are unequal for two independent samples.…
Descriptors: Analysis of Covariance, Analysis of Variance, Equations (Mathematics), Error of Measurement

Hough, Susan L.; Hall, Bruce W. – Journal of Educational Research, 1994
Compares results of Hunter-Schmidt meta-analytic technique with results of Glass meta-analytic technique on three meta-analytic data sets chosen from the literature, hypothesizing that the Hunter-Schmidt mean effect size would be significantly larger than the Glass mean effect size because of correlation for measurement error. Results confirmed…
Descriptors: Comparative Analysis, Educational Research, Effect Size, Error of Measurement

Liou, Michelle; Cheng, Philip E. – Journal of Educational and Behavioral Statistics, 1995
Simplified formulas are proposed for computing the standard errors of equipercentile equating for continuous and discrete test scores. These formulas are easily extended to more complicated equating designs. Results from a study of 719 subjects taking an English test indicated that the formulas work reasonably well for moderate-size samples. (SLD)
Descriptors: College Students, Equated Scores, Equations (Mathematics), Error of Measurement

Lund, Thorleif – Scandinavian Journal of Educational Research, 1995
Four general criteria are proposed for the choice of a metrical solution for a causal effect: (1) compatibility with the effect; (2) ease of communication; (3) lack of measurement error bias; and (4) stability across subjects and situations. These criteria are illustrated for randomized and nonrandomized designs. (SLD)
Descriptors: Causal Models, Communication (Thought Transfer), Criteria, Error of Measurement

Evans, Brian – Canadian Journal of Program Evaluation/La Revue canadienne d'evaluation de programme, 1995
The distinction between two models of reliability is clarified. Reliability may be conceived of and estimated from a true score model or from the perspective of sampling precision. Basic models are developed and illustrated for each approach using data from the author's work on measuring organizational climate. (SLD)
Descriptors: Data Analysis, Error of Measurement, Evaluators, Models

Bamezai, Anil – Evaluation Review, 1995
Some of the threats to internal validity that arise when evaluating the impact of water conservation programs during a drought are illustrated. These include differential response to the drought, self-selection bias, and measurement error. How to deal with these problems when high-quality disaggregate data are available is discussed. (SLD)
Descriptors: Conservation (Environment), Drought, Error of Measurement, Evaluation Methods

Allison, David B.; And Others – Journal of Experimental Education, 1992
Effects of response guided experimentation in applied behavior analysis on Type I error rates are explored. Data from T. A. Matyas and K. M. Greenwood (1990) suggest that, when visual inspection is combined with response guided experimentation, Type I error rates can be as high as 25%. (SLD)
Descriptors: Behavioral Science Research, Error of Measurement, Evaluation Methods, Experiments

Sheard, Christine; And Others – Journal of Speech and Hearing Research, 1991
The study calculated indices of interjudge reliability and interjudge and intrajudge agreement on ratings made by 15 experienced speech clinicians on 5 deviant speech dimensions of 15 speakers with ataxic dysarthria and a wide range of speech intelligibility. Judges were reliable in tracking imprecise consonants, excess and equal stress, and harsh…
Descriptors: Adults, Error of Measurement, Evaluation Methods, Interrater Reliability

Pohlmann, John T. – Mid-Western Educational Researcher, 1993
Nonlinear relationships and latent variable assumptions can lead to serious specification errors in structural models. A quadratic relationship, described by a linear structural model with a latent variable, is shown to have less predictive validity than a simple manifest variable regression model. Advocates the use of simpler preliminary…
Descriptors: Causal Models, Error of Measurement, Predictor Variables, Research Methodology

McDonald, Roderick P.; And Others – Psychometrika, 1993
A reparameterization is formulated that yields estimates of scale-invariant parameters in recursive path models with latent variables, and (asymptotically) correct standard errors, without the use of constrained optimization. The method is based on the logical structure of the reticular action model. (Author)
Descriptors: Correlation, Equations (Mathematics), Error of Measurement, Estimation (Mathematics)

Kurokawa, Nancy K. S.; Weed, Nathan C. – Assessment, 1998
The relationship between self- and peer report on the Coping Inventory for Stressful Situations (CISS) (N. Endler and J. Parker, 1990) was studied with 163 pairs of friends (college students). Positive but modest correlations were found between peer and self-report for three types of coping. Findings that attenuate these correlations are…
Descriptors: College Students, Coping, Error of Measurement, Interrater Reliability