NotesFAQContact Us
Collection
Advanced
Search Tips
Showing 1 to 15 of 32 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E. – Educational and Psychological Measurement, 2011
Standard setting is a method used to set cut scores on large-scale assessments. One of the most popular standard setting methods is the Bookmark method. In the Bookmark method, panelists are asked to envision a response probability (RP) criterion and move through a booklet of ordered items based on a RP criterion. This study investigates whether…
Descriptors: Testing Programs, Standard Setting (Scoring), Cutting Scores, Probability
Peer reviewed Peer reviewed
Wilcox, Rand R. – Educational and Psychological Measurement, 1981
This paper describes and compares procedures for estimating the reliability of proficiency tests that are scored with latent structure models. Results suggest that the predictive estimate is the most accurate of the procedures. (Author/BW)
Descriptors: Criterion Referenced Tests, Scoring, Test Reliability
Peer reviewed Peer reviewed
Rozeboom, William W. – Educational and Psychological Measurement, 1978
A strict equivalence presupposed by Kaiser and Michael to derive the coefficient of "domain validity" is defensible only as a biased approximation. But then, it is far from clear what psychometric significance this coefficient has in the first place. (Author)
Descriptors: Criterion Referenced Tests, Item Analysis, Item Banks, Test Validity
Peer reviewed Peer reviewed
Lovett, Hubert T. – Educational and Psychological Measurement, 1977
The analysis of variance model for estimating reliability in norm referenced tests is extended to criterion referenced tests. The essential modification is that the criterion or cut-off score is substituted for the population mean. An example and discussion are presented. (JKS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Cutting Scores, Test Reliability
Peer reviewed Peer reviewed
Young, James C.; And Others – Educational and Psychological Measurement, 1970
Descriptors: Achievement Tests, Algebra, Criterion Referenced Tests, Mathematics
Peer reviewed Peer reviewed
Huynh, Huynh – Educational and Psychological Measurement, 1990
Within the multivariate normality framework, a formula is provided for computation of the criterion-related validity of composite scores based on the highest (or lowest) of several equivalent measures. This partial composite score has more validity than each single observation, but less validity than a composite based on all observations. (SLD)
Descriptors: Concurrent Validity, Criterion Referenced Tests, Equations (Mathematics), Mathematical Models
Peer reviewed Peer reviewed
Lovett, Hubert T. – Educational and Psychological Measurement, 1978
The validity of five methods of estimating the reliability of criterion-referenced tests was evaluated across nine conditions of variability among item means. The results were analyzed by analysis of variance, the Newman-Keuls test, and a nonparametric procedure. There was a tendency for all of the methods to be conservative. (Author/JKS)
Descriptors: Analysis of Variance, Criterion Referenced Tests, Item Analysis, Nonparametric Statistics
Peer reviewed Peer reviewed
Macready, George B.; Merwin, Jack C. – Educational and Psychological Measurement, 1973
In this paper consideration is given to the nature of the relationships among items within item forms and how these relationships compare with an ideal case for diagnostic tests in which if a person gets one item within an item form right then he would get all items within the item form correct. (Authors)
Descriptors: Criterion Referenced Tests, Diagnostic Tests, Homogeneous Grouping, Item Analysis
Peer reviewed Peer reviewed
Raju, Nambury S. – Educational and Psychological Measurement, 1982
Rajaratnam, Cronbach and Gleser's generalizability formula for stratified-parallel tests and Raju's coefficient beta are generalized to estimate the reliability of a composite of criterion-referenced tests, where the parts have different cutting scores. (Author/GK)
Descriptors: Criterion Referenced Tests, Cutting Scores, Mathematical Formulas, Scoring Formulas
Peer reviewed Peer reviewed
Roid, G. H.; Haladyna, Thomas M. – Educational and Psychological Measurement, 1978
Two techniques for writing achievement test items to accompany instructional materials are contrasted: writing items from statements of instructional objectives, and writing items from semi-automated rules for transforming instructional statements. Both systems resulted in about the same number of faulty items. (Author/JKS)
Descriptors: Achievement Tests, Comparative Analysis, Criterion Referenced Tests, Difficulty Level
Peer reviewed Peer reviewed
Hambleton, Ronald K. – Educational and Psychological Measurement, 1987
This paper presents an algorithm for determining the number of items to measure each objective in a criterion-referenced test when testing time is fixed and when the objectives vary in their levels of importance, reliability, and validity. Results of four special applications of the algorithm are presented. (BS)
Descriptors: Algorithms, Behavioral Objectives, Criterion Referenced Tests, Test Construction
Peer reviewed Peer reviewed
Martin, John D.; Rudolph, Linda – Educational and Psychological Measurement, 1972
The SIT Correlates highly enough with ACT scores to be considered a valid instrument for predicting acceptance and success in college. (Authors)
Descriptors: Comparative Analysis, Criterion Referenced Tests, Grade Point Average, Intelligence Tests
Peer reviewed Peer reviewed
Wilcox, Rand R. – Educational and Psychological Measurement, 1981
The paper considers the problem of selecting the t best of k normal populations and simultaneously determining whether the selected populations have a mean larger than a known standard. Illustrations are given for selecting the t best of k examinees when the binomial error model applies. (Author)
Descriptors: Competitive Selection, Criterion Referenced Tests, Decision Making, Mathematical Models
Peer reviewed Peer reviewed
Spineti, John P.; Hambleton, Ronald K. – Educational and Psychological Measurement, 1977
The effectiveness of various tailored testing strategies for use in objective based instructional programs was investigated. The three factors of a tailored testing strategy under study with various hypothetical distributions of abilities across two learning hierarchies were test length, mastery cutting score, and starting point. (Author/JKS)
Descriptors: Adaptive Testing, Computer Assisted Testing, Criterion Referenced Tests, Cutting Scores
Peer reviewed Peer reviewed
Nolan, James S.; Jacobson, James – Educational and Psychological Measurement, 1972
In general Achievement Test scores appeared to be more valid predictors of grades in English and Mathematics courses than were scores on a scholastic aptitude or general intelligence test. (Authors)
Descriptors: Achievement Tests, College Freshmen, Criterion Referenced Tests, High School Freshmen
Previous Page | Next Page ยป
Pages: 1  |  2  |  3