NotesFAQContact Us
Collection
Advanced
Search Tips
Education Level
Audience
Practitioners1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 26 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Sophie Litschwartz – Society for Research on Educational Effectiveness, 2021
Background/Context: Pass/fail standardized exams frequently selectively rescore failing exams and retest failing examinees. This practice distorts the test score distribution and can confuse those who do analysis on these distributions. In 2011, the Wall Street Journal showed large discontinuities in the New York City Regent test score…
Descriptors: Standardized Tests, Pass Fail Grading, Scoring Rubrics, Scoring Formulas
Swineford, Frances – 1973
Results obtained by the Kudar-Richardson formula (20) adapted for use with R-KW scoring are compared with three other reliability formulas. Based on parallel tests administered at the same sitting the KR (20) estimates are compared with alternate-form correlations and with odd-even correlations adjusted by the Spearman-Brown prophecy formula.…
Descriptors: Aptitude Tests, Scoring Formulas, Test Interpretation, Test Reliability
Berk, Ronald A. – 1980
Seventeen statistics for measuring the reliability of criterion-referenced tests were critically reviewed. The review was organized into two sections: (1) a discussion of preliminary considerations to provide a foundation for choosing the appropriate category of "reliability" (threshold loss function, squared-error loss-function, or…
Descriptors: Criterion Referenced Tests, Cutting Scores, Scoring Formulas, Statistical Analysis
Peer reviewed Peer reviewed
Sattler, Jerome M.; And Others – Psychology in the Schools, 1978
Fabricated test protocols were used to study how effectively examiners agree in scoring ambiguous WISC-R responses. The results suggest that, even with the improved WISC-R manual, scoring remains a difficult and challenging task. (Author)
Descriptors: Comparative Analysis, Intelligence Tests, Research Projects, Scoring Formulas
Peer reviewed Peer reviewed
Frary, Robert B.; And Others – Journal of Experimental Education, 1977
To date a theoretical basis has not been developed for determining changes in reliability when score points from random guessing are eliminated and those from non-randon guessing are retained. This paper presents a derivation of an expression for the reliability coefficient which displays the effect of deleting score components due to random…
Descriptors: Data Analysis, Guessing (Tests), Multiple Choice Tests, Scoring Formulas
Peer reviewed Peer reviewed
Abu-Sayf, F. K. – Educational and Psychological Measurement, 1977
A new formula for the correction for chance success due to guessing was advanced and investigated, the mathematical solution of which has the property of equating the scores of fast and slow examinees of equal ability. (Author/JKS)
Descriptors: Academic Ability, Equated Scores, Guessing (Tests), Scoring Formulas
Peer reviewed Peer reviewed
Frary, Robert B. – Journal of Educational Statistics, 1982
Six different approaches to scoring test data, including number right, correction for guessing, and answer-until-correct, were investigated using Monte Carlo techniques. Modes permitting multiple response showed higher internal consistency, but there was little difference among modes for a validity measure. (JKS)
Descriptors: Guessing (Tests), Measurement Techniques, Multiple Choice Tests, Scoring Formulas
Peer reviewed Peer reviewed
Dolliver, Robert H.; And Others – Journal of Vocational Behavior, 1975
The 1966 SVIB scoring keys were investigated and found to contain an average of 214 fewer items per occupational scale than the 1938 scoring keys. The shorter scales are less reliable than the longer scales on the 1966 SVIB. (Author)
Descriptors: Career Counseling, Interest Inventories, Item Analysis, Occupational Tests
Knapp, Thomas R. – Measurement and Evaluation in Guidance, 1980
Supports arguments against general use of change scores and recommends the Lord/McNemar estimates of true change. Provides a numerical example illustrating the reliability problem and the problem of the prediction of true change from various linear composites of initial and final measures. (Author)
Descriptors: Counseling Techniques, Literature Reviews, Pretests Posttests, Research Methodology
Peer reviewed Peer reviewed
Diamond, James J. – Journal of Educational Measurement, 1975
Investigates the reliability and validity of scores yielded from a new scoring formula. (Author/DEP)
Descriptors: Guessing (Tests), Multiple Choice Tests, Objective Tests, Scoring
Hambleton, Ronald K.; Novick, Melvin R. – 1972
In this paper, an attempt has been made to synthesize some of the current thinking in the area of criterion-referenced testing as well as to provide the beginning of an integration of theory and method for such testing. Since criterion-referenced testing is viewed from a decision-theoretic point of view, approaches to reliability and validity…
Descriptors: Criterion Referenced Tests, Measurement Instruments, Measurement Techniques, Scaling
Koehler, Roger A. – 1974
A potentially valuable measure of overconfidence on probabilistic multiple-choice tests was evaluated. The measure of overconfidence was based on probabilistic responses to nonsense items embedded in a vocabulary test. The test was administered under both confidence response and conventional choice response directions to 208 undergraduate…
Descriptors: Confidence Testing, Guessing (Tests), Measurement Techniques, Multiple Choice Tests
Foegen, Anne – Diagnostique, 2000
A study involving 105 sixth-graders examined three aspects of technical adequacy with respect to two general outcome measures in mathematics: the effects of aggregating scores and correcting for random guessing on reliability and validity and the extent to which the measures were sensitive to changes in performance. (Contains references.)…
Descriptors: Curriculum Based Assessment, Disabilities, Grade 6, Mathematics
Bayuk, Robert J. – 1973
An investigation was conducted to determine the effects of response-category weighting and item weighting on reliability and predictive validity. Response-category weighting refers to scoring in which, for each category (including omit and "not read"), a weight is assigned that is proportional to the mean criterion score of examinees selecting…
Descriptors: Aptitude Tests, Correlation, Predictive Validity, Research Reports
Ohio State Univ., Columbus. Trade and Industrial Education Instructional Materials Lab. – 1978
The Ohio Vocational Achievement Tests are specially designed instruments for use by teachers, supervisors, and administrators to evaluate and diagnose vocational achievement for improving instruction in secondary vocational programs at the 11th and 12th grade levels. This guide explains the Ohio Vocational Achievement Tests and how they are used.…
Descriptors: Academic Achievement, Achievement Tests, High Schools, Scoring Formulas
Previous Page | Next Page ยป
Pages: 1  |  2