Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 3 |
Descriptor
Student Evaluation | 8 |
Evaluation Methods | 4 |
Validity | 4 |
Interrater Reliability | 3 |
Reliability | 3 |
Scores | 3 |
Scoring | 3 |
Test Reliability | 3 |
Comparative Analysis | 2 |
Correlation | 2 |
Educational Assessment | 2 |
More ▼ |
Source
Applied Measurement in… | 8 |
Author
Ames, Allison J. | 1 |
Boughton, Keith A. | 1 |
Calfee, Robert | 1 |
Case, Susan M. | 1 |
Fisher, Steve | 1 |
Gierl, Mark J. | 1 |
Godfrey, Alan T. K. | 1 |
Gotzmann, Andrea | 1 |
Holzman, Madison A. | 1 |
Johnson, Robert L. | 1 |
Kane, Michael | 1 |
More ▼ |
Publication Type
Journal Articles | 8 |
Reports - Research | 5 |
Reports - Evaluative | 4 |
Education Level
Higher Education | 1 |
Postsecondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Myers, Aaron J.; Ames, Allison J.; Leventhal, Brian C.; Holzman, Madison A. – Applied Measurement in Education, 2020
When rating performance assessments, raters may ascribe different scores for the same performance when rubric application does not align with the intended application of the scoring criteria. Given performance assessment score interpretation assumes raters apply rubrics as rubric developers intended, misalignment between raters' scoring processes…
Descriptors: Scoring Rubrics, Validity, Item Response Theory, Interrater Reliability
Slepkov, Aaron D.; Godfrey, Alan T. K. – Applied Measurement in Education, 2019
The answer-until-correct (AUC) method of multiple-choice (MC) testing involves test respondents making selections until the keyed answer is identified. Despite attendant benefits that include improved learning, broad student adoption, and facile administration of partial credit, the use of AUC methods for classroom testing has been extremely…
Descriptors: Multiple Choice Tests, Test Items, Test Reliability, Scores
Stone, Clement A.; Ye, Feifei; Zhu, Xiaowen; Lane, Suzanne – Applied Measurement in Education, 2010
Although reliability of subscale scores may be suspect, subscale scores are the most common type of diagnostic information included in student score reports. This research compared methods for augmenting the reliability of subscale scores for an 8th-grade mathematics assessment. Yen's Objective Performance Index, Wainer et al.'s augmented scores,…
Descriptors: Item Response Theory, Case Studies, Reliability, Scores
Johnson, Robert L.; Penny, Jim; Fisher, Steve; Kuhs, Therese – Applied Measurement in Education, 2003
When raters assign different scores to a performance task, a method for resolving rating differences is required to report a single score to the examinee. Recent studies indicate that decisions about examinees, such as pass/fail decisions, differ across resolution methods. Previous studies also investigated the interrater reliability of…
Descriptors: Test Reliability, Test Validity, Scores, Interrater Reliability
Kane, Michael; Case, Susan M. – Applied Measurement in Education, 2004
The scores on 2 distinct tests (e.g., essay and objective) are often combined to create a composite score, which is used to make decisions. The validity of the observed composite can sometimes be evaluated relative to an external criterion. However, in cases where no criterion is available, the observed composite has generally been evaluated in…
Descriptors: Validity, Weighted Scores, Reliability, Student Evaluation
Gierl, Mark J.; Gotzmann, Andrea; Boughton, Keith A. – Applied Measurement in Education, 2004
Differential item functioning (DIF) analyses are used to identify items that operate differently between two groups, after controlling for ability. The Simultaneous Item Bias Test (SIBTEST) is a popular DIF detection method that matches examinees on a true score estimate of ability. However in some testing situations, like test translation and…
Descriptors: True Scores, Simulation, Test Bias, Student Evaluation

Linn, Robert L.; And Others – Applied Measurement in Education, 1992
Ten states participated in a cross-state scoring workshop in 1991, evaluating writing from elementary school, middle school, and high school students. Correlation of scores assigned by readers from one state with those from readers from another state were generally quite high. Implications for defining common standards are discussed. (SLD)
Descriptors: Comparative Analysis, Correlation, Elementary School Students, Elementary Secondary Education

Valencia, Sheila W.; Calfee, Robert – Applied Measurement in Education, 1991
Using portfolios in assessing literacy is explored, considering student portfolios and the teacher's class portfolio. Portfolio assessment is a valuable complement to externally mandated tests, but technical issues must be addressed if the portfolio movement is to survive. Portfolios must be linked to the broader task of instructional improvement.…
Descriptors: Academic Achievement, Educational Assessment, Educational Improvement, Elementary School Teachers