Descriptor
Criterion Referenced Tests | 32 |
Test Length | 32 |
Mastery Tests | 18 |
Test Construction | 16 |
Cutting Scores | 14 |
Mathematical Models | 11 |
Test Validity | 10 |
Test Reliability | 9 |
Latent Trait Theory | 8 |
Comparative Analysis | 6 |
Computer Assisted Testing | 6 |
More ▼ |
Source
Educational and Psychological… | 4 |
Journal of Educational… | 2 |
Review of Educational Research | 2 |
Applied Measurement in… | 1 |
Applied Psychological… | 1 |
Research Quarterly for… | 1 |
Author
Publication Type
Education Level
Audience
Researchers | 2 |
Location
Laws, Policies, & Programs
Assessments and Surveys
Texas Assessment of Basic… | 1 |
Texas Educational Assessment… | 1 |
What Works Clearinghouse Rating

Safrit, Margaret J.; And Others – Research Quarterly for Exercise and Sport, 1985
Constraints on criterion-referenced tests to make mastery/nonmastery classifications of motor skills can lead to excessively long tests. A sequential probability ratio test classified many subjects' golf shots quickly but required many trials for four subjects. The test's classification accuracy makes it a potentially useful device for physical…
Descriptors: Criterion Referenced Tests, Golf, Higher Education, Mastery Tests
Wilcox, Rand R. – 1980
Concern about passing those examinees who should pass, and retaining those who need remedial work, is one problem related to criterion-referenced testing. This paper deals with one aspect of that problem. When determining how many items to include on a criterion-referenced test, practitioners must resolve various non-statistical issues before a…
Descriptors: Bayesian Statistics, Criterion Referenced Tests, Latent Trait Theory, Mathematical Models

Millman, Jason – Review of Educational Research, 1973
Procedures for establishing standards and determining the number of items needed in criterion referenced measures were reviewed. Discussion of setting a passing score was organized around: performance of others, item content, educational consequences, psychological and financial costs, and error due to guessing and item sampling. (Author)
Descriptors: Criterion Referenced Tests, Educational Research, Literature Reviews, Measurement Techniques

Hambleton, Ronald K. – Educational and Psychological Measurement, 1987
This paper presents an algorithm for determining the number of items to measure each objective in a criterion-referenced test when testing time is fixed and when the objectives vary in their levels of importance, reliability, and validity. Results of four special applications of the algorithm are presented. (BS)
Descriptors: Algorithms, Behavioral Objectives, Criterion Referenced Tests, Test Construction

Spineti, John P.; Hambleton, Ronald K. – Educational and Psychological Measurement, 1977
The effectiveness of various tailored testing strategies for use in objective based instructional programs was investigated. The three factors of a tailored testing strategy under study with various hypothetical distributions of abilities across two learning hierarchies were test length, mastery cutting score, and starting point. (Author/JKS)
Descriptors: Adaptive Testing, Computer Assisted Testing, Criterion Referenced Tests, Cutting Scores

Wilcox, Rand R. – Educational and Psychological Measurement, 1982
When determining criterion-referenced test length, problems of guessing are shown to be more serious than expected. A new method of scoring is presented that corrects for guessing without assuming that guessing is random. Empirical investigations of the procedure are examined. Test length can be substantially reduced. (Author/CM)
Descriptors: Criterion Referenced Tests, Guessing (Tests), Multiple Choice Tests, Scoring

Wilcox, Rand R. – Educational and Psychological Measurement, 1979
A problem of considerable importance in certain educational settings is determining how many items to include on a mastery test. Applying ranking and selection procedures, a solution is given which includes as a special case all existing single-stage, non-Bayesian solutions based on a strong true-score model. (Author/JKS)
Descriptors: Criterion Referenced Tests, Mastery Tests, Nonparametric Statistics, Probability
Johnston, Shirley H.; And Others – 1983
A computer simulation was undertaken to determine the effects of using Huynh's single-administration estimates of the decision consistency indices for agreement and coefficient kappa, under conditions that violated the beta-binomial assumption. Included in the investigation were two unimodal score distributions that fit the model and two bimodal…
Descriptors: Bias, Criterion Referenced Tests, Data, Mastery Tests

Hambleton, Ronald K.; And Others – Journal of Educational Measurement, 1983
A new method was developed to assist in the selection of a test length by utilizing computer simulation procedures and item response theory. A demonstration of the method presents results which address the influences of item pool heterogeneity matched to the objectives of interest and the method of item selection. (Author/PN)
Descriptors: Computer Programs, Criterion Referenced Tests, Item Banks, Latent Trait Theory
Graham, Darol L. – 1974
The adequacy of a test developed for statewide assessment of basic mathematics skills was investigated. The test, comprised of multiple-choice items reflecting a series of behavioral objectives, was compared with a more extensive criterion measure generated from the same objectives by the application of a strict item sampling model. In many…
Descriptors: Comparative Testing, Criterion Referenced Tests, Educational Assessment, Item Sampling

Hambleton, Ronald K., Ed. – Applied Psychological Measurement, 1980
This special issue covers recent technical developments in the field of criterion-referenced testing. An introduction, six papers, and two commentaries dealing with test development, test score uses, and evaluation of scores review relevant literature, offer new models and/or results, and suggest directions for additional research. (SLD)
Descriptors: Criterion Referenced Tests, Mastery Tests, Measurement Techniques, Standard Setting (Scoring)
Hills, John R. – 1979
Six experimental approaches to the problems of setting cutoff scores and choosing proper test length are briefly mentioned. Most of these methods share the premise that a test is a random sample of items, from a domain associated with a carefully specified objective. Each item is independent and is scored zero or one, with no provision for…
Descriptors: Academic Standards, Aptitude Treatment Interaction, Criterion Referenced Tests, Cutting Scores
Haladyna, Tom; Roid, Gale – 1981
Two approaches to criterion-referenced test construction are compared. Classical test theory is based on the practice of random sampling from a well-defined domain of test items; latent trait theory suggests that the difficulty of the items should be matched to the achievement level of the student. In addition to these two methods of test…
Descriptors: Criterion Referenced Tests, Error of Measurement, Latent Trait Theory, Test Construction

Hambleton, Ronald K.; De Gruijter, Dato N. M. – Journal of Educational Measurement, 1983
Addressing the shortcomings of classical item statistics for selecting criterion-referenced test items, this paper describes an optimal item selection procedure utilizing item response theory (IRT) and offers examples in which random selection and optimal item selection methods are compared. Theoretical advantages of optimal selection based upon…
Descriptors: Criterion Referenced Tests, Cutting Scores, Item Banks, Latent Trait Theory
Steinheiser, Frederick H., Jr. – 1976
A computer simulation of Bayes' Theorem was conducted in order to determine the probability that an examinee was a master conditional upon his test score. The inputs were: number of mastery states assumed, test length, prior expectation of masters in the examinee population, and conditional probability of a master getting a randomly selected test…
Descriptors: Bayesian Statistics, Classification, Computer Programs, Criterion Referenced Tests