NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Practitioners1
What Works Clearinghouse Rating
Showing 1 to 15 of 62 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Matt Homer – Advances in Health Sciences Education, 2024
Quantitative measures of systematic differences in OSCE scoring across examiners (often termed examiner stringency) can threaten the validity of examination outcomes. Such effects are usually conceptualised and operationalised based solely on checklist/domain scores in a station, and global grades are not often used in this type of analysis. In…
Descriptors: Examiners, Scoring, Validity, Cutting Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Peter Baldwin; Victoria Yaneva; Kai North; Le An Ha; Yiyun Zhou; Alex J. Mechaber; Brian E. Clauser – Journal of Educational Measurement, 2025
Recent developments in the use of large-language models have led to substantial improvements in the accuracy of content-based automated scoring of free-text responses. The reported accuracy levels suggest that automated systems could have widespread applicability in assessment. However, before they are used in operational testing, other aspects of…
Descriptors: Artificial Intelligence, Scoring, Computational Linguistics, Accuracy
Peer reviewed Peer reviewed
Direct linkDirect link
Mark White; Matt Ronfeldt – Educational Assessment, 2024
Standardized observation systems seek to reliably measure a specific conceptualization of teaching quality, managing rater error through mechanisms such as certification, calibration, validation, and double-scoring. These mechanisms both support high quality scoring and generate the empirical evidence used to support the scoring inference (i.e.,…
Descriptors: Interrater Reliability, Quality Control, Teacher Effectiveness, Error Patterns
Peer reviewed Peer reviewed
Direct linkDirect link
Li, Liang-Yi; Huang, Wen-Lung – Educational Technology & Society, 2023
With the increasing bandwidth, videos have been gradually used as submissions for online peer assessment activities. However, their transient nature imposes a high cognitive load on students, particularly lowability students. Therefore, reviewers' ability is a key factor that may affect the reviewing process and performance in an online video peer…
Descriptors: Peer Evaluation, Undergraduate Students, Video Technology, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Almusharraf, Norah; Alotaibi, Hind – Technology, Knowledge and Learning, 2023
Evaluating written texts is believed to be a time-consuming process that can lack consistency and objectivity. Automated essay scoring (AES) can provide solutions to some of the limitations of human scoring. This research aimed to evaluate the performance of one AES system, Grammarly, in comparison to human raters. Both approaches' performances…
Descriptors: Writing Evaluation, Writing Tests, Essay Tests, Essays
Peer reviewed Peer reviewed
Direct linkDirect link
Lockwood, Adam B.; Klatka, Kelsey; Freeman, Kelli; Farmer, Ryan L.; Benson, Nicholas – Journal of Psychoeducational Assessment, 2023
Sixty-three Woodcock-Johnson IV Tests of Achievement protocols, administered by 26 school psychology trainees, were examined to determine the frequency of examiner errors. Errors were noted on all protocols and ranged from 8 to 150 per administration. Critical (e.g., start, stop, and calculation) errors were noted on roughly 97% of protocols.…
Descriptors: Achievement Tests, School Psychology, Counselor Training, Trainees
Peer reviewed Peer reviewed
Direct linkDirect link
Corcoran, Stephanie – Contemporary School Psychology, 2022
With the iPad-mediated cognitive assessment gaining popularity with school districts and the need for alternative modes for training and instruction during this COVID-19 pandemic, school psychology training programs will need to adapt to effectively train their students to be competent in administering, scoring, an interpreting cognitive…
Descriptors: School Psychologists, Professional Education, Job Skills, Cognitive Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Lockwood, Adam B.; Sealander, Karen; Gross, Thomas J.; Lanterman, Christopher – Journal of Psychoeducational Assessment, 2020
Achievement tests are used to make high-stakes (e.g., special education placement) decisions, and previous research on norm-referenced assessment suggests that errors are ubiquitous. In our study of 42 teacher trainees, utilizing five of the six core subtests of the Kaufman Test of Educational Achievement, Third Edition (KTEA-3), we found that…
Descriptors: Achievement Tests, Preservice Teachers, Testing, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Henbest, Victoria S.; Apel, Kenn – Language, Speech, and Hearing Services in Schools, 2021
Purpose: As an initial step in determining whether a spelling error analysis might be useful in measuring children's linguistic knowledge, the relation between the frequency of types of scores from a spelling error analysis and children's performance on measures of phonological and orthographic pattern awareness was examined. Method: The spellings…
Descriptors: Elementary School Students, Grade 1, Spelling, Orthographic Symbols
Peer reviewed Peer reviewed
Direct linkDirect link
Oak, Erika; Viezel, Kathleen D.; Dumont, Ron; Willis, John – Journal of Psychoeducational Assessment, 2019
Individuals trained in the use of cognitive tests should be able to complete an assessment without making administrative, scoring, or recording errors. However, an examination of 295 Wechsler protocols completed by graduate students and practicing school psychologists revealed that errors are the norm, not the exception. The most common errors…
Descriptors: Intelligence Tests, Children, Adults, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Chan, Sathena; May, Lyn – Language Testing, 2023
Despite the increased use of integrated tasks in high-stakes academic writing assessment, research on rating criteria which reflect the unique construct of integrated summary writing skills is comparatively rare. Using a mixed-method approach of expert judgement, text analysis, and statistical analysis, this study examines writing features that…
Descriptors: Scoring, Writing Evaluation, Reading Tests, Listening Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Treiman, Rebecca; Kessler, Brett; Caravolas, Markéta – Journal of Research in Reading, 2019
Background: Children's spellings are often scored as correct or incorrect, but other measures may be better predictors of later spelling performance. Method: We examined seven measures of spelling in Reception Year and Year 1 (5-6 years old) as predictors of performance on a standardised spelling test in Year 2 (age 7). Results: Correctness was…
Descriptors: Spelling, Scoring, Predictor Variables, Elementary School Students
Peer reviewed Peer reviewed
Direct linkDirect link
Reinertsen, Nathanael – English in Australia, 2018
The difference in how humans read and how Automated Essay Scoring (AES) systems process written language leads to a situation where a portion of student responses will be comprehensible to human markers, but unable to be parsed by AES systems. This paper examines a number of pieces of student writing that were marked by trained human markers, but…
Descriptors: Qualitative Research, Writing Evaluation, Essay Tests, Computer Assisted Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Yachong Cui; Rachel Saulsburry; Kimberly Wolbers – American Annals of the Deaf, 2024
Limited access to spoken and signed language is a worldwide phenomenon affecting deaf children. Language delay caused by impeded language acquisition has negative cascading effects on deaf children's learning and development. In the event of stymied language development, deaf students exhibit highly errored writing and commit errors unseen in the…
Descriptors: Deafness, Written Language, Writing Evaluation, North Americans
Peer reviewed Peer reviewed
Direct linkDirect link
Allalouf, Avi; Gutentag, Tony; Baumer, Michal – Educational Measurement: Issues and Practice, 2017
Quality control (QC) in testing is paramount. QC procedures for tests can be divided into two types. The first type, one that has been well researched, is QC for tests administered to large population groups on few administration dates using a small set of test forms (e.g., large-scale assessment). The second type is QC for tests, usually…
Descriptors: Quality Control, Scoring, Computer Assisted Testing, Error Patterns
Previous Page | Next Page »
Pages: 1  |  2  |  3  |  4  |  5