NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 5 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Timmerman, Briana E. Crotwell; Strickland, Denise C.; Johnson, Robert L.; Payne, John R. – Assessment & Evaluation in Higher Education, 2011
We developed a rubric for measuring students' ability to reason and write scientifically. The Rubric for Science Writing (Rubric) was tested in a variety of undergraduate biology laboratory courses (total n = 142 laboratory reports) using science graduate students (teaching assistants) as raters. Generalisability analysis indicates that the Rubric…
Descriptors: Graduate Students, Science Laboratories, Biology, Writing Skills
Peer reviewed Peer reviewed
Penny, Jim; Johnson, Robert L.; Gordon, Belita – Journal of Experimental Education, 2000
Used an analytic rubric to score 120 writing samples from Georgia's 11th grade writing assessment. Raters augmented scores by adding a "+" or "-" to the score. Results indicate that this method of augmentation tends to improve most indices of interrater reliability, although the percentage of exact and adjacent agreement…
Descriptors: High School Students, High Schools, Interrater Reliability, Scoring Rubrics
Peer reviewed Peer reviewed
Johnson, Robert L.; Penny, James; Gordon, Belita – Applied Measurement in Education, 2000
Studied four forms of score resolution used by testing agencies and investigated the effect that each has on the interrater reliability associated with the resulting operational scores. Results, based on 120 essays from the Georgia High School Writing Test, show some forms of resolution to be associated with higher reliability and some associated…
Descriptors: Essay Tests, High School Students, High Schools, Interrater Reliability
Peer reviewed Peer reviewed
Penny, Jim; Johnson, Robert L.; Gordon, Belita – Assessing Writing, 2000
Defines a two-stage process by which a holistic rubric is applied to the assessment of open-ended items, such as writing samples. Indicates that the use of rating augmentation can improve the inter-rater reliability of holistic assessments, as indicated by generalizability phi coefficients, correlation coefficients, and percent agreement indices.…
Descriptors: Grade 5, Holistic Evaluation, Intermediate Grades, Reliability
Peer reviewed Peer reviewed
Direct linkDirect link
Johnson, Robert L.; Penny, James; Gordon, Belita; Shumate, Steven R.; Fisher, Steven P. – Language Assessment Quarterly, 2005
Many studies have indicated that at least 2 raters should score writing assessments to improve interrater reliability. However, even for assessments that characteristically demonstrate high levels of rater agreement, 2 raters of the same essay can occasionally report different, or discrepant, scores. If a single score, typically referred to as an…
Descriptors: Interrater Reliability, Scores, Evaluation, Reliability