Descriptor
Mathematical Models | 12 |
Test Length | 12 |
Test Reliability | 12 |
Cutting Scores | 6 |
Statistical Analysis | 6 |
Mastery Tests | 5 |
Criterion Referenced Tests | 4 |
Test Construction | 4 |
Test Validity | 4 |
Comparative Analysis | 3 |
Test Items | 3 |
More ▼ |
Author
Hambleton, Ronald K. | 2 |
Wilcox, Rand R. | 2 |
Budescu, David | 1 |
Cliff, Norman | 1 |
Eignor, Daniel R. | 1 |
Feldt, Leonard S. | 1 |
Gilmer, Jerry S. | 1 |
Gross, Susan K. | 1 |
Kristof, Walter | 1 |
Livingston, Samuel A. | 1 |
McKinley, Robert L. | 1 |
More ▼ |
Publication Type
Reports - Research | 10 |
Speeches/Meeting Papers | 3 |
Journal Articles | 1 |
Education Level
Audience
Researchers | 1 |
Location
Laws, Policies, & Programs
Assessments and Surveys
School and College Ability… | 1 |
Stanford Binet Intelligence… | 1 |
What Works Clearinghouse Rating
Wilcox, Rand R. – 1980
Wilcox (1977) examines two methods of estimating the probability of a false-positive on false-negative decision with a mastery test. Both procedures make assumptions about the form of the true score distribution which might not give good results in all situations. In this paper, upper and lower bounds on the two possible error types are described…
Descriptors: Cutting Scores, Mastery Tests, Mathematical Models, Student Placement

Budescu, David – Journal of Educational Measurement, 1985
An important determinant of equating process efficiency is the correlation between the anchor test and components of each form. Use of some monotonic function of this correlation as a measure of equating efficiency is suggested. A model relating anchor test length and test reliability to this measure of efficiency is presented. (Author/DWH)
Descriptors: Correlation, Equated Scores, Mathematical Models, Standardized Tests
Livingston, Samuel A. – 1984
Much previously published material for estimating the reliability of classification has been based on the assumption that a test consists of a known number of equally weighted items. The test score is the number of those items answered correctly. These methods cannot be used with classifications based on weighted composite scores, especially if…
Descriptors: Equated Scores, Essay Tests, Estimation (Mathematics), Mathematical Models

Kristof, Walter – Psychometrika, 1971
Descriptors: Cognitive Measurement, Error of Measurement, Mathematical Models, Psychological Testing

Gilmer, Jerry S.; Feldt, Leonard S. – 1982
The Feldt-Gilmer congeneric reliability coefficients make it possible to estimate the reliability of a test composed of parts of unequal, unknown length. The approximate standard errors of the Feldt-Gilmer coefficients are derived via a method using the multivariate Taylor's expansion. Monte Carlo simulation is employed to corroborate the…
Descriptors: Educational Testing, Error of Measurement, Mathematical Formulas, Mathematical Models
Eignor, Daniel R.; Hambleton, Ronald K. – 1979
The purpose of the investigation was to obtain some relationships among (1) test lengths, (2) shape of domain-score distributions, (3) advancement scores, and (4) several criterion-referenced test score reliability and validity indices. The study was conducted using computer simulation methods. The values of variables under study were set to be…
Descriptors: Comparative Analysis, Computer Assisted Testing, Criterion Referenced Tests, Cutting Scores

Hambleton, Ronald K.; And Others – Review of Educational Research, 1978
Reviewing psychometric and statistical developments in criterion- referenced testing, this paper presents six sections: uses of criterion- referenced test scores, reliability of criterion-referenced test scores, determination of test length, determination of cut-off scores, test development and validation, and summary and suggestions for further…
Descriptors: Criterion Referenced Tests, Cutting Scores, Mastery Tests, Mathematical Models
Schaefer, Mary M.; Gross, Susan K. – 1983
Viewing the reliability for criterion-referenced tests as that of mastery classification decisions, three models for determining reliability were examined using two test administrations so that two estimates could be compared to a standard. A major purpose of the research was to determine how several reliability coefficients (coefficient kappa, an…
Descriptors: Comparative Analysis, Correlation, Criterion Referenced Tests, Cutting Scores
Subkoviak, Michael J. – 1977
Four different procedures were used for estimating the proportion of persons who would be classified consistently as either passing both of two parallel tests or failing both. These four methods were applied at each of four different mastery level scores for each of three different length tests. Data were based on 50 replications of each procedure…
Descriptors: Criterion Referenced Tests, Cutting Scores, Data Analysis, Data Collection
McKinley, Robert L.; Reckase, Mark D. – 1981
A study was conducted to compare tailored testing procedures based on a Bayesian ability estimation technique and on a maximum likelihood ability estimation technique. The Bayesian tailored testing procedure selected items so as to minimize the posterior variance of the ability estimate distribution, while the maximum likelihood tailored testing…
Descriptors: Academic Ability, Adaptive Testing, Bayesian Statistics, Comparative Analysis
Wilcox, Rand R. – 1979
Mastery tests are analyzed in terms of the number of skills to be mastered and the number of items per skill, in order that correct decisions of mastery or nonmastery will be made to a desired degree of probability. It is assumed that a random sample of skills will be selected for measurement, that each skill will be measured by the same number of…
Descriptors: Achievement Tests, Cutting Scores, Decision Making, Equivalency Tests
Cliff, Norman; And Others – 1977
TAILOR is a computer program that uses the implied orders concept as the basis for computerized adaptive testing. The basic characteristics of TAILOR, which does not involve pretesting, are reviewed here and two studies of it are reported. One is a Monte Carlo simulation based on the four-parameter Birnbaum model and the other uses a matrix of…
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Programs, Difficulty Level