Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 3 |
Descriptor
Automation | 3 |
College Entrance Examinations | 3 |
Correlation | 3 |
Interrater Reliability | 3 |
Scoring | 3 |
Essay Tests | 2 |
Graduate Study | 2 |
Data | 1 |
Demography | 1 |
Essays | 1 |
Evaluation Methods | 1 |
More ▼ |
Author
Attali, Yigal | 1 |
Bridgeman, Brent | 1 |
Davey, Tim | 1 |
Lewis, Will | 1 |
Ramineni, Chaitanya | 1 |
Steier, Michael | 1 |
Trapani, Catherine S. | 1 |
Williamson, David M. | 1 |
Zhang, Mo | 1 |
Publication Type
Journal Articles | 3 |
Reports - Research | 2 |
Reports - Evaluative | 1 |
Education Level
Higher Education | 3 |
Postsecondary Education | 3 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 3 |
What Works Clearinghouse Rating
Zhang, Mo – ETS Research Report Series, 2013
Many testing programs use automated scoring to grade essays. One issue in automated essay scoring that has not been examined adequately is population invariance and its causes. The primary purpose of this study was to investigate the impact of sampling in model calibration on population invariance of automated scores. This study analyzed scores…
Descriptors: Automation, Scoring, Essay Tests, Sampling
Attali, Yigal; Lewis, Will; Steier, Michael – Language Testing, 2013
Automated essay scoring can produce reliable scores that are highly correlated with human scores, but is limited in its evaluation of content and other higher-order aspects of writing. The increased use of automated essay scoring in high-stakes testing underscores the need for human scoring that is focused on higher-order aspects of writing. This…
Descriptors: Scoring, Essay Tests, Reliability, High Stakes Tests
Ramineni, Chaitanya; Trapani, Catherine S.; Williamson, David M.; Davey, Tim; Bridgeman, Brent – ETS Research Report Series, 2012
Automated scoring models for the "e-rater"® scoring engine were built and evaluated for the "GRE"® argument and issue-writing tasks. Prompt-specific, generic, and generic with prompt-specific intercept scoring models were built and evaluation statistics such as weighted kappas, Pearson correlations, standardized difference in…
Descriptors: Scoring, Test Scoring Machines, Automation, Models