Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 1 |
Descriptor
Source
Educational and Psychological… | 32 |
Author
Publication Type
Journal Articles | 20 |
Reports - Research | 15 |
Reports - Evaluative | 6 |
Opinion Papers | 1 |
Reports - Descriptive | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
ACT Assessment | 1 |
Comprehensive Tests of Basic… | 1 |
Graduate Management Admission… | 1 |
Slosson Intelligence Test | 1 |
Stanford Diagnostic Reading… | 1 |
Wechsler Adult Intelligence… | 1 |
What Works Clearinghouse Rating
Wyse, Adam E. – Educational and Psychological Measurement, 2011
Standard setting is a method used to set cut scores on large-scale assessments. One of the most popular standard setting methods is the Bookmark method. In the Bookmark method, panelists are asked to envision a response probability (RP) criterion and move through a booklet of ordered items based on a RP criterion. This study investigates whether…
Descriptors: Testing Programs, Standard Setting (Scoring), Cutting Scores, Probability

Wilcox, Rand R. – Educational and Psychological Measurement, 1981
This paper describes and compares procedures for estimating the reliability of proficiency tests that are scored with latent structure models. Results suggest that the predictive estimate is the most accurate of the procedures. (Author/BW)
Descriptors: Criterion Referenced Tests, Scoring, Test Reliability

Rozeboom, William W. – Educational and Psychological Measurement, 1978
A strict equivalence presupposed by Kaiser and Michael to derive the coefficient of "domain validity" is defensible only as a biased approximation. But then, it is far from clear what psychometric significance this coefficient has in the first place. (Author)
Descriptors: Criterion Referenced Tests, Item Analysis, Item Banks, Test Validity

Lovett, Hubert T. – Educational and Psychological Measurement, 1977
The analysis of variance model for estimating reliability in norm referenced tests is extended to criterion referenced tests. The essential modification is that the criterion or cut-off score is substituted for the population mean. An example and discussion are presented. (JKS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Cutting Scores, Test Reliability

Young, James C.; And Others – Educational and Psychological Measurement, 1970
Descriptors: Achievement Tests, Algebra, Criterion Referenced Tests, Mathematics

Huynh, Huynh – Educational and Psychological Measurement, 1990
Within the multivariate normality framework, a formula is provided for computation of the criterion-related validity of composite scores based on the highest (or lowest) of several equivalent measures. This partial composite score has more validity than each single observation, but less validity than a composite based on all observations. (SLD)
Descriptors: Concurrent Validity, Criterion Referenced Tests, Equations (Mathematics), Mathematical Models
The Effect of Violating the Assumption of Equal Item Means in Estimating the Livingston Coefficient.

Lovett, Hubert T. – Educational and Psychological Measurement, 1978
The validity of five methods of estimating the reliability of criterion-referenced tests was evaluated across nine conditions of variability among item means. The results were analyzed by analysis of variance, the Newman-Keuls test, and a nonparametric procedure. There was a tendency for all of the methods to be conservative. (Author/JKS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Item Analysis, Nonparametric Statistics

Macready, George B.; Merwin, Jack C. – Educational and Psychological Measurement, 1973
In this paper consideration is given to the nature of the relationships among items within item forms and how these relationships compare with an ideal case for diagnostic tests in which if a person gets one item within an item form right then he would get all items within the item form correct. (Authors)
Descriptors: Criterion Referenced Tests, Diagnostic Tests, Homogeneous Grouping, Item Analysis

Raju, Nambury S. – Educational and Psychological Measurement, 1982
Rajaratnam, Cronbach and Gleser's generalizability formula for stratified-parallel tests and Raju's coefficient beta are generalized to estimate the reliability of a composite of criterion-referenced tests, where the parts have different cutting scores. (Author/GK)
Descriptors: Criterion Referenced Tests, Cutting Scores, Mathematical Formulas, Scoring Formulas

Roid, G. H.; Haladyna, Thomas M. – Educational and Psychological Measurement, 1978
Two techniques for writing achievement test items to accompany instructional materials are contrasted: writing items from statements of instructional objectives, and writing items from semi-automated rules for transforming instructional statements. Both systems resulted in about the same number of faulty items. (Author/JKS)
Descriptors: Achievement Tests, Comparative Analysis, Criterion Referenced Tests, Difficulty Level

Hambleton, Ronald K. – Educational and Psychological Measurement, 1987
This paper presents an algorithm for determining the number of items to measure each objective in a criterion-referenced test when testing time is fixed and when the objectives vary in their levels of importance, reliability, and validity. Results of four special applications of the algorithm are presented. (BS)
Descriptors: Algorithms, Behavioral Objectives, Criterion Referenced Tests, Test Construction

Martin, John D.; Rudolph, Linda – Educational and Psychological Measurement, 1972
The SIT Correlates highly enough with ACT scores to be considered a valid instrument for predicting acceptance and success in college. (Authors)
Descriptors: Comparative Analysis, Criterion Referenced Tests, Grade Point Average, Intelligence Tests

Wilcox, Rand R. – Educational and Psychological Measurement, 1981
The paper considers the problem of selecting the t best of k normal populations and simultaneously determining whether the selected populations have a mean larger than a known standard. Illustrations are given for selecting the t best of k examinees when the binomial error model applies. (Author)
Descriptors: Competitive Selection, Criterion Referenced Tests, Decision Making, Mathematical Models

Spineti, John P.; Hambleton, Ronald K. – Educational and Psychological Measurement, 1977
The effectiveness of various tailored testing strategies for use in objective based instructional programs was investigated. The three factors of a tailored testing strategy under study with various hypothetical distributions of abilities across two learning hierarchies were test length, mastery cutting score, and starting point. (Author/JKS)
Descriptors: Adaptive Testing, Computer Assisted Testing, Criterion Referenced Tests, Cutting Scores

Nolan, James S.; Jacobson, James – Educational and Psychological Measurement, 1972
In general Achievement Test scores appeared to be more valid predictors of grades in English and Mathematics courses than were scores on a scholastic aptitude or general intelligence test. (Authors)
Descriptors: Achievement Tests, College Freshmen, Criterion Referenced Tests, High School Freshmen