NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 92 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Matt Homer – Advances in Health Sciences Education, 2024
Quantitative measures of systematic differences in OSCE scoring across examiners (often termed examiner stringency) can threaten the validity of examination outcomes. Such effects are usually conceptualised and operationalised based solely on checklist/domain scores in a station, and global grades are not often used in this type of analysis. In…
Descriptors: Examiners, Scoring, Validity, Cutting Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Peter Baldwin; Victoria Yaneva; Kai North; Le An Ha; Yiyun Zhou; Alex J. Mechaber; Brian E. Clauser – Journal of Educational Measurement, 2025
Recent developments in the use of large-language models have led to substantial improvements in the accuracy of content-based automated scoring of free-text responses. The reported accuracy levels suggest that automated systems could have widespread applicability in assessment. However, before they are used in operational testing, other aspects of…
Descriptors: Artificial Intelligence, Scoring, Computational Linguistics, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Xin Qiao; Akihito Kamata; Cornelis Potgieter – Grantee Submission, 2024
Oral reading fluency (ORF) assessments are commonly used to screen at-risk readers and evaluate interventions' effectiveness as curriculum-based measurements. Similar to the standard practice in item response theory (IRT), calibrated passage parameter estimates are currently used as if they were population values in model-based ORF scoring.…
Descriptors: Oral Reading, Reading Fluency, Error Patterns, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Mark White; Matt Ronfeldt – Educational Assessment, 2024
Standardized observation systems seek to reliably measure a specific conceptualization of teaching quality, managing rater error through mechanisms such as certification, calibration, validation, and double-scoring. These mechanisms both support high quality scoring and generate the empirical evidence used to support the scoring inference (i.e.,…
Descriptors: Interrater Reliability, Quality Control, Teacher Effectiveness, Error Patterns
Jessica Stinson – ProQuest LLC, 2024
Intelligence tests have been used in the United States since the early 1900s for assessing soldiers during World War I (Kaufman & Harrison, 2008; White & Hall, 1980). Presently, cognitive assessments are used in school, civil service, military, clinical, and industry settings (White & Hall, 1980). Although the results of these…
Descriptors: Graduate Students, Masters Programs, Doctoral Programs, Comparative Analysis
Atehortua, Laura – ProQuest LLC, 2022
Intelligence tests are used in a variety of settings such as schools, clinics, and courts to assess the intellectual capacity of individuals of all ages. Intelligence tests are used to make high-stakes decisions such as special education placement, employment, eligibility for social security services, and determination of the death penalty.…
Descriptors: Adults, Intelligence Tests, Children, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Liang-Yi; Huang, Wen-Lung – Educational Technology & Society, 2023
With the increasing bandwidth, videos have been gradually used as submissions for online peer assessment activities. However, their transient nature imposes a high cognitive load on students, particularly lowability students. Therefore, reviewers' ability is a key factor that may affect the reviewing process and performance in an online video peer…
Descriptors: Peer Evaluation, Undergraduate Students, Video Technology, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Almusharraf, Norah; Alotaibi, Hind – Technology, Knowledge and Learning, 2023
Evaluating written texts is believed to be a time-consuming process that can lack consistency and objectivity. Automated essay scoring (AES) can provide solutions to some of the limitations of human scoring. This research aimed to evaluate the performance of one AES system, Grammarly, in comparison to human raters. Both approaches' performances…
Descriptors: Writing Evaluation, Writing Tests, Essay Tests, Essays
Peer reviewed Peer reviewed
Direct linkDirect link
Lockwood, Adam B.; Klatka, Kelsey; Freeman, Kelli; Farmer, Ryan L.; Benson, Nicholas – Journal of Psychoeducational Assessment, 2023
Sixty-three Woodcock-Johnson IV Tests of Achievement protocols, administered by 26 school psychology trainees, were examined to determine the frequency of examiner errors. Errors were noted on all protocols and ranged from 8 to 150 per administration. Critical (e.g., start, stop, and calculation) errors were noted on roughly 97% of protocols.…
Descriptors: Achievement Tests, School Psychology, Counselor Training, Trainees
Botarleanu, Robert-Mihai; Dascalu, Mihai; Allen, Laura K.; Crossley, Scott Andrew; McNamara, Danielle S. – Grantee Submission, 2022
Automated scoring of student language is a complex task that requires systems to emulate complex and multi-faceted human evaluation criteria. Summary scoring brings an additional layer of complexity to automated scoring because it involves two texts of differing lengths that must be compared. In this study, we present our approach to automate…
Descriptors: Automation, Scoring, Documentation, Likert Scales
Abdalla, Widad – ProQuest LLC, 2019
Trend scoring is often used in large-scale assessments to monitor for rater drift when the same constructed response items are administered in multiple test administrations. In trend scoring, a set of responses from Time "A" are rescored by raters at Time "B." The purpose of this study is to examine the ability of…
Descriptors: Scoring, Interrater Reliability, Test Items, Error Patterns
Petscher, Y.; Pentimonti, J.; Stanley, C. – National Center on Improving Literacy, 2019
Reliability is the consistency of a set of scores that are designed to measure the same thing. Reliability is a statistical property of scores that must be demonstrated rather than assumed.
Descriptors: Scores, Measurement, Test Reliability, Error Patterns
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Baral, Sami; Botelho, Anthony F.; Erickson, John A.; Benachamardi, Priyanka; Heffernan, Neil T. – International Educational Data Mining Society, 2021
Open-ended questions in mathematics are commonly used by teachers to monitor and assess students' deeper conceptual understanding of content. Student answers to these types of questions often exhibit a combination of language, drawn diagrams and tables, and mathematical formulas and expressions that supply teachers with insight into the processes…
Descriptors: Scoring, Automation, Mathematics Tests, Student Evaluation
Peer reviewed Peer reviewed
Direct linkDirect link
Corcoran, Stephanie – Contemporary School Psychology, 2022
With the iPad-mediated cognitive assessment gaining popularity with school districts and the need for alternative modes for training and instruction during this COVID-19 pandemic, school psychology training programs will need to adapt to effectively train their students to be competent in administering, scoring, an interpreting cognitive…
Descriptors: School Psychologists, Professional Education, Job Skills, Cognitive Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Lockwood, Adam B.; Sealander, Karen; Gross, Thomas J.; Lanterman, Christopher – Journal of Psychoeducational Assessment, 2020
Achievement tests are used to make high-stakes (e.g., special education placement) decisions, and previous research on norm-referenced assessment suggests that errors are ubiquitous. In our study of 42 teacher trainees, utilizing five of the six core subtests of the Kaufman Test of Educational Achievement, Third Edition (KTEA-3), we found that…
Descriptors: Achievement Tests, Preservice Teachers, Testing, Scoring
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7