Publication Date
In 2025 | 3 |
Descriptor
Evaluators | 3 |
Models | 3 |
Item Response Theory | 2 |
Automation | 1 |
Bias | 1 |
Computation | 1 |
Computer Uses in Education | 1 |
Cost Effectiveness | 1 |
Creativity | 1 |
Data | 1 |
Educational Assessment | 1 |
More ▼ |
Author
Akif Avcu | 1 |
Benjamin Goecke | 1 |
Boris Forthmann | 1 |
Carl Westine | 1 |
Michelle Boyer | 1 |
Roger E. Beaty | 1 |
Stella Y. Kim | 1 |
Tong Wu | 1 |
Publication Type
Journal Articles | 3 |
Reports - Research | 2 |
Information Analyses | 1 |
Education Level
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Akif Avcu – Malaysian Online Journal of Educational Technology, 2025
This scope-review presents the milestones of how Hierarchical Rater Models (HRMs) become operable to used in automated essay scoring (AES) to improve instructional evaluation. Although essay evaluations--a useful instrument for evaluating higher-order cognitive abilities--have always depended on human raters, concerns regarding rater bias,…
Descriptors: Automation, Scoring, Models, Educational Assessment
Boris Forthmann; Benjamin Goecke; Roger E. Beaty – Creativity Research Journal, 2025
Human ratings are ubiquitous in creativity research. Yet, the process of rating responses to creativity tasks -- typically several hundred or thousands of responses, per rater -- is often time-consuming and expensive. Planned missing data designs, where raters only rate a subset of the total number of responses, have been recently proposed as one…
Descriptors: Creativity, Research, Researchers, Research Methodology
Tong Wu; Stella Y. Kim; Carl Westine; Michelle Boyer – Journal of Educational Measurement, 2025
While significant attention has been given to test equating to ensure score comparability, limited research has explored equating methods for rater-mediated assessments, where human raters inherently introduce error. If not properly addressed, these errors can undermine score interchangeability and test validity. This study proposes an equating…
Descriptors: Item Response Theory, Evaluators, Error of Measurement, Test Validity