NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Does not meet standards1
Showing 2,716 to 2,730 of 3,311 results Save | Export
PDF pending restoration PDF pending restoration
Harris, Chester W. – 1971
Livingston's work is a careful analysis of what occurs when one pools two populations with different means, but similar variances and reliability coefficients. However, his work fails to advance reliability theory for the special case of criterion-referenced testing. See ED 042 802 for Livingston's paper. (MS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Error of Measurement, Reliability
Lord, Frederic M.; Stocking, Martha – 1972
A general Computer program is described that will compute asymptotic standard errors and carry out significance tests for an endless variety of (standard and) nonstandard large-sample statistical problems, without requiring the statistician to derive asymptotic standard error formulas. The program assumes that the observations have a multinormal…
Descriptors: Bulletins, Computer Programs, Data Processing, Error of Measurement
Peer reviewed Peer reviewed
Werts, C. E.; And Others – Educational and Psychological Measurement, 1976
A procedure is presented for the analysis of rating data with correlated intrajudge and uncorrelated interjudge measurement errors. Correlations between true scores on different rating dimensions, reliabilities for each judge on each dimension and correlations between intrajudge errors can be estimated given a minimum of three raters and two…
Descriptors: Correlation, Data Analysis, Error of Measurement, Error Patterns
Tsujimoto, Richard N.; Berger, Dale E. – Child Abuse and Neglect: The International Journal, 1988
Two criteria are discussed for determining cutting scores on a predictor variable for identifying cases of likely child abuse--utility maximizing and error minimizing. Utility maximizing is the preferable criterion, as it optimizes the balance between the costs of incorrect decisions and the benefits of correct decisions. (Author/JDD)
Descriptors: Child Abuse, Cost Effectiveness, Cutting Scores, Error of Measurement
Peer reviewed Peer reviewed
Kennedy, Eugene – Journal of Experimental Education, 1988
Ridge estimates (REs) of population beta weights were compared to ordinary least squares (OLS) estimates through computer simulation to evaluate the use of REs in explanatory research. With fixed predictors, there was some question of the consistency of ridge regression, but with random predictors, REs were superior to OLS. (SLD)
Descriptors: Computer Simulation, Error of Measurement, Estimation (Mathematics), Least Squares Statistics
Peer reviewed Peer reviewed
Jarjoura, David; Kolen, Michael J. – Journal of Educational Statistics, 1985
An equating design in which two groups of examinees from slightly different populations are administered a different test form with a subset of common items is widely used. This paper presents standard errors and a simulation that verifies the equation for large samples for an equipercentile equating procedure for this design. (Author/BS)
Descriptors: Computer Simulation, Equated Scores, Error of Measurement, Estimation (Mathematics)
Peer reviewed Peer reviewed
Schaeffer, Gary A.; And Others – Evaluation Review, 1986
The reliability of criterion-referenced tests (CRTs) used in health program evaluation can be conceptualized in different ways. Formulas are presented for estimating appropriate standard error of measurement (SEM) for CRTs. The SEM can be used in computing confidence intervals for domain score estimates and for a cut-score. (Author/LMO)
Descriptors: Accountability, Criterion Referenced Tests, Cutting Scores, Error of Measurement
Peer reviewed Peer reviewed
Rogosa, David R.; Willett, John B. – Journal of Educational Measurement, 1983
Demonstrating good reliability for the difference score in measurement, the results of this study indicate that the difference score is often highly reliable when the correlation between true change and true initial status is nonnegative. In general, when individual differences in true change are appreciable, the difference score shows strong…
Descriptors: Achievement Gains, Error of Measurement, Individual Differences, Measurement Techniques
Peer reviewed Peer reviewed
Lord, Frederic M. – Journal of Educational Measurement, 1984
Four methods are outlined for estimating or approximating from a single test administration the standard error of measurement of number-right test score at specified ability levels or cutting scores. The methods are illustrated and compared on one set of real test data. (Author)
Descriptors: Academic Ability, Cutting Scores, Error of Measurement, Scoring Formulas
Sykes, Robert C.; Hou, Liling; Hanson, Brad; Wang, Zhen – 2002
This study investigated the effect on student scores of using anchor sets that differed in dimensionality in item response theory (IRT) scaled tests. Real data from a mathematics achievement test that had been documented to have dimensions aligned with item format were used. Item responses were available from a representative sample of…
Descriptors: Elementary School Students, Equated Scores, Error of Measurement, Intermediate Grades
Yi, Qing; Wang, Tianyou; Ban, Jae-Chun – 2000
Error indices (bias, standard error of estimation, and root mean square error) obtained on different scales of measurement under different test termination rules in a computerized adaptive test (CAT) context were examined. Four ability estimation methods were studied: (1) maximum likelihood estimation (MLE); (2) weighted likelihood estimation…
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Error of Measurement
Peer reviewed Peer reviewed
Livingston, Samuel A. – Journal of Educational Measurement, 1972
This article is a reply to a previous paper (see TM 500 488) interpreting Livingston's original article (see TM 500 487). (CK)
Descriptors: Criterion Referenced Tests, Error of Measurement, Norm Referenced Tests, Test Construction
Peer reviewed Peer reviewed
Kristof, Walter – Psychometrika, 1971
Descriptors: Cognitive Measurement, Error of Measurement, Mathematical Models, Psychological Testing
Peer reviewed Peer reviewed
Novick, Melvin R.; And Others – Psychometrika, 1971
Descriptors: Analysis of Variance, Bayesian Statistics, Error of Measurement, Mathematical Models
Peer reviewed Peer reviewed
Haladyna, Thomas M.; Roid, Gale H. – Journal of Educational Measurement, 1983
The present study showed that Rasch-based adaptive tests--when item domains were finite and specifiable--had greater precision in domain score estimation than test forms created by random sampling of items. Results were replicated across four data sources representing a variety of criterion-referenced, domain-based tests varying in length.…
Descriptors: Adaptive Testing, Criterion Referenced Tests, Error of Measurement, Estimation (Mathematics)
Pages: 1  |  ...  |  178  |  179  |  180  |  181  |  182  |  183  |  184  |  185  |  186  |  ...  |  221