Publication Date
| In 2026 | 0 |
| Since 2025 | 2 |
| Since 2022 (last 5 years) | 12 |
| Since 2017 (last 10 years) | 26 |
| Since 2007 (last 20 years) | 90 |
Descriptor
| True Scores | 416 |
| Error of Measurement | 121 |
| Test Reliability | 110 |
| Statistical Analysis | 107 |
| Mathematical Models | 97 |
| Item Response Theory | 87 |
| Correlation | 76 |
| Equated Scores | 76 |
| Reliability | 64 |
| Test Theory | 52 |
| Test Items | 51 |
| More ▼ | |
Source
Author
Publication Type
Education Level
Audience
| Researchers | 12 |
| Practitioners | 2 |
| Administrators | 1 |
| Teachers | 1 |
Location
| Australia | 1 |
| Canada | 1 |
| China | 1 |
| Colorado | 1 |
| Illinois | 1 |
| Israel | 1 |
| New York | 1 |
| Oregon | 1 |
| Taiwan | 1 |
| Texas | 1 |
| United Kingdom (England) | 1 |
| More ▼ | |
Laws, Policies, & Programs
| Elementary and Secondary… | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Magidson, Jay – Evaluation Quarterly, 1977
Path analysis was used to reevaluate the analysis of covariance quasiexperimental study of the effectiveness of the Head Start program. Contrary to the original analysis, the alternative approach yields small positive estimates of effect. (Author/CTM)
Descriptors: Analysis of Covariance, Factor Analysis, Mathematical Models, Path Analysis
Peer reviewedWerts, C. E.; And Others – Educational and Psychological Measurement, 1976
A procedure is presented for the analysis of rating data with correlated intrajudge and uncorrelated interjudge measurement errors. Correlations between true scores on different rating dimensions, reliabilities for each judge on each dimension and correlations between intrajudge errors can be estimated given a minimum of three raters and two…
Descriptors: Correlation, Data Analysis, Error of Measurement, Error Patterns
Tsujimoto, Richard N.; Berger, Dale E. – Child Abuse and Neglect: The International Journal, 1988
Two criteria are discussed for determining cutting scores on a predictor variable for identifying cases of likely child abuse--utility maximizing and error minimizing. Utility maximizing is the preferable criterion, as it optimizes the balance between the costs of incorrect decisions and the benefits of correct decisions. (Author/JDD)
Descriptors: Child Abuse, Cost Effectiveness, Cutting Scores, Error of Measurement
Peer reviewedRogosa, David R.; Willett, John B. – Journal of Educational Measurement, 1983
Demonstrating good reliability for the difference score in measurement, the results of this study indicate that the difference score is often highly reliable when the correlation between true change and true initial status is nonnegative. In general, when individual differences in true change are appreciable, the difference score shows strong…
Descriptors: Achievement Gains, Error of Measurement, Individual Differences, Measurement Techniques
Peer reviewedLord, Frederic M. – Journal of Educational Measurement, 1984
Four methods are outlined for estimating or approximating from a single test administration the standard error of measurement of number-right test score at specified ability levels or cutting scores. The methods are illustrated and compared on one set of real test data. (Author)
Descriptors: Academic Ability, Cutting Scores, Error of Measurement, Scoring Formulas
Peer reviewedLivingston, Samuel A. – Journal of Educational Measurement, 1972
This article is a reply to a previous paper (see TM 500 488) interpreting Livingston's original article (see TM 500 487). (CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Norm Referenced Tests, Test Construction
Peer reviewedNovick, Melvin R.; And Others – Psychometrika, 1971
Descriptors: Analysis of Variance, Bayesian Statistics, Error of Measurement, Mathematical Models
Peer reviewedHanna, Gerald S.; And Others – Journal of School Psychology, 1981
Discusses four ubiquitous major sources of measurement error for individual intelligence scales. Argues that where these sources cannot be directly investigated, they should be estimated rather than ignored. Estimated the typical magnitude of error arising from each of content sampling, time sampling, scoring, and administration. (Author)
Descriptors: Error of Measurement, Intelligence Tests, Measurement Techniques, Sampling
Peer reviewedOlsson, Ulf – Multivariate Behavioral Research, 1979
The paper discusses the consequences for maximum likelihood factor analysis which may follow if the observed variables are ordinal with only a few scale steps. Results indicate that classification may lead to a substantial lack of fit of the model--an erroneous indication that more factors are needed. (Author/CTM)
Descriptors: Classification, Factor Analysis, Goodness of Fit, Maximum Likelihood Statistics
Peer reviewedCahan, Sorel – Educational and Psychological Measurement, 1989
Statistical significance and "abnormality" have been used as criteria for the evaluation of intra-individual subtest score differences. Shortcomings of these criteria are identified, and improved estimates of the true score differences are suggested. The applicability of the abnormality criterion to these improved estimates is reviewed.…
Descriptors: Estimation (Mathematics), Evaluation Methods, Individual Differences, Mathematical Models
Peer reviewedLin, Miao-Hsiang; Hsiung, Chao A. – Psychometrika, 1992
Four bootstrap methods are identified for constructing confidence intervals for the binomial-error model. The extent to which similar results are obtained and the theoretical foundation of each method and its relevance and ranges of modeling the true score uncertainty are discussed. (SLD)
Descriptors: Bayesian Statistics, Computer Simulation, Equations (Mathematics), Estimation (Mathematics)
Wang, Tianyou; And Others – 1996
M. J. Kolen, B. A. Hanson, and R. L. Brennan (1992) presented a procedure for assessing the conditional standard error of measurement (CSEM) of scale scores using a strong true-score model. They also investigated the ways of using nonlinear transformation from number-correct raw score to scale score to equalize the conditional standard error along…
Descriptors: Ability, Classification, Error of Measurement, Goodness of Fit
Wilcox, Rand R. – 1983
This document presents a series of five papers describing issues in educational measurement. "A Simple Model for Diagnostic Testing When There Are Several Types of Misinformation" directly addresses the diagnostic issue. It describes a simple latent trait model for testing, examines use of erroneous algorithms, and illustrates the…
Descriptors: Diagnostic Tests, Estimation (Mathematics), Guessing (Tests), Latent Trait Theory
Wilcox, Rand R. – 1978
Two fundamental problems in mental test theory are to estimate true score and to estimate the amount of error when testing an examinee. In this report, three probability models which characterize a single test item in terms of a population of examinees are described. How these models may be modified to characterize a single examinee in terms of an…
Descriptors: Achievement Tests, Comparative Analysis, Error of Measurement, Mathematical Models
Yeh, Jennie P.; Moy, Raymond – 1980
The setting of a cut-off score on a mastery test usually involves a consideration of one or more of the following elements: (1) the distribution of observed test scores; (2) the type of mastery criterion used; (3) the level of acceptable risks of mis-classification; (4) the loss of functions of mis-classifications; and (5) the distribution of true…
Descriptors: Classification, Criteria, Cutting Scores, English (Second Language)


