NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 9 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Stefanie A. Wind; Yangmeng Xu – Educational Assessment, 2024
We explored three approaches to resolving or re-scoring constructed-response items in mixed-format assessments: rater agreement, person fit, and targeted double scoring (TDS). We used a simulation study to consider how the three approaches impact the psychometric properties of student achievement estimates, with an emphasis on person fit. We found…
Descriptors: Interrater Reliability, Error of Measurement, Evaluation Methods, Examiners
Peer reviewed Peer reviewed
Direct linkDirect link
Wind, Stefanie A.; Guo, Wenjing – Educational Assessment, 2021
Scoring procedures for the constructed-response (CR) items in large-scale mixed-format educational assessments often involve checks for rater agreement or rater reliability. Although these analyses are important, researchers have documented rater effects that persist despite rater training and that are not always detected in rater agreement and…
Descriptors: Scoring, Responses, Test Items, Test Format
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Minji K.; Sweeney, Kevin; Melican, Gerald J. – Educational Assessment, 2017
This study investigates the relationships among factor correlations, inter-item correlations, and the reliability estimates of subscores, providing a guideline with respect to psychometric properties of useful subscores. In addition, it compares subscore estimation methods with respect to reliability and distinctness. The subscore estimation…
Descriptors: Scores, Test Construction, Test Reliability, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Bell, Courtney A.; Jones, Nathan D.; Qi, Yi; Lewis, Jennifer M. – Educational Assessment, 2018
All 50 states use observations to evaluate practicing teachers, but we know little about how administrators actually reason when they use those observation protocols. Drawing on think-aloud and stimulated recall data, this study describes the types of strategies and warrants practicing administrators used when rating with their district's…
Descriptors: Administrators, Observation, Validity, Logical Thinking
Peer reviewed Peer reviewed
Direct linkDirect link
Reckase, Mark D.; McCrory, Raven; Floden, Robert E.; Ferrini-Mundy, Joan; Senk, Sharon L. – Educational Assessment, 2015
Numerous researchers have suggested that there are multiple mathematical knowledge and skill areas needed by teachers in order for them to be effective teachers of mathematics: knowledge of the mathematics that are the goals of instruction, advanced mathematics beyond the instructional material, and mathematical knowledge that is specific to what…
Descriptors: Algebra, Knowledge Base for Teaching, Multidimensional Scaling, Psychometrics
Peer reviewed Peer reviewed
Direct linkDirect link
Reed, Deborah K. – Educational Assessment, 2011
This narrative synthesis reviews the psychometric properties of commercially and publicly available retell instruments used to assess the reading comprehension of students in grades K-12. Eleven instruments met selection criteria and were systematically coded for data related to the administration procedures, scoring procedures, and technical…
Descriptors: Reading Comprehension, Elementary Secondary Education, Construct Validity, Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Martinez, Jose Felipe; Borko, Hilda; Stecher, Brian; Luskin, Rebecca; Kloser, Matt – Educational Assessment, 2012
We report the results of a pilot validation study of the Quality Assessment in Science Notebook, a portfolio-like instrument for measuring teacher assessment practices in middle school science classrooms. A statewide sample of 42 teachers collected 2 notebooks during the school year, corresponding to science topics taught in the fall and spring.…
Descriptors: Validity, Middle School Teachers, Evaluation Methods, Educational Assessment
Peer reviewed Peer reviewed
Kane, Michael – Educational Assessment, 1998
Examines criteria for choosing between test-centered and examinee-centered methods of standard setting in empirical terms and in terms of whether the method is consistent with the model of achievement underlying test design and interpretation and the assessment methods being used. Contains 35 references. (Author/SLD)
Descriptors: Academic Achievement, Criteria, Educational Assessment, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Dawson, Theo L.; Wilson, Mark – Educational Assessment, 2004
The evaluation of developmental interventions has been hampered by a lack of practical, reliable, and objective developmental assessment systems. This article describes the construction of a domain-general computerized developmental assessment system for texts: the Lexical Abstraction Assessment System (LAAS). The LAAS provides assessments of the…
Descriptors: Scoring, Evaluation Methods, Discriminant Analysis, Computer Uses in Education