Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 5 |
Since 2006 (last 20 years) | 14 |
Descriptor
Essay Tests | 17 |
College Entrance Examinations | 14 |
Scoring | 12 |
Automation | 6 |
Graduate Study | 6 |
Computer Assisted Testing | 5 |
Writing Tests | 5 |
Correlation | 4 |
Interrater Reliability | 4 |
Standardized Tests | 4 |
Writing Skills | 4 |
More ▼ |
Source
Author
Attali, Yigal | 3 |
Arslan, Burcu | 2 |
Bridgeman, Brent | 2 |
Finn, Bridgid | 2 |
Fowles, Mary E. | 2 |
Powers, Donald E. | 2 |
Ramineni, Chaitanya | 2 |
Almond, Russell G. | 1 |
Beigman Klebanov, Beata | 1 |
Breyer, F. Jay | 1 |
Brown, Kevin | 1 |
More ▼ |
Publication Type
Journal Articles | 14 |
Reports - Research | 11 |
Reports - Evaluative | 5 |
Reports - General | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 12 |
Postsecondary Education | 10 |
Elementary Secondary Education | 1 |
Audience
Researchers | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 17 |
Test of English as a Foreign… | 6 |
College Level Examination… | 1 |
Praxis Series | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Finn, Bridgid; Arslan, Burcu; Walsh, Matthew – Applied Measurement in Education, 2020
To score an essay response, raters draw on previously trained skills and knowledge about the underlying rubric and score criterion. Cognitive processes such as remembering, forgetting, and skill decay likely influence rater performance. To investigate how forgetting influences scoring, we evaluated raters' scoring accuracy on TOEFL and GRE essays.…
Descriptors: Epistemology, Essay Tests, Evaluators, Cognitive Processes
Finn, Bridgid; Wendler, Cathy; Ricker-Pedley, Kathryn L.; Arslan, Burcu – ETS Research Report Series, 2018
This report investigates whether the time between scoring sessions has an influence on operational and nonoperational scoring accuracy. The study evaluates raters' scoring accuracy on constructed-response essay responses for the "GRE"® General Test. Binomial linear mixed-effect models are presented that evaluate how the effect of various…
Descriptors: Intervals, Scoring, Accuracy, Essay Tests
Breyer, F. Jay; Rupp, André A.; Bridgeman, Brent – ETS Research Report Series, 2017
In this research report, we present an empirical argument for the use of a contributory scoring approach for the 2-essay writing assessment of the analytical writing section of the "GRE"® test in which human and machine scores are combined for score creation at the task and section levels. The approach was designed to replace a currently…
Descriptors: College Entrance Examinations, Scoring, Essay Tests, Writing Evaluation
Almond, Russell G. – International Journal of Testing, 2014
Assessments consisting of only a few extended constructed response items (essays) are not typically equated using anchor test designs as there are typically too few essay prompts in each form to allow for meaningful equating. This article explores the idea that output from an automated scoring program designed to measure writing fluency (a common…
Descriptors: Automation, Equated Scores, Writing Tests, Essay Tests
Ramineni, Chaitanya; Williamson, David – ETS Research Report Series, 2018
Notable mean score differences for the "e-rater"® automated scoring engine and for humans for essays from certain demographic groups were observed for the "GRE"® General Test in use before the major revision of 2012, called rGRE. The use of e-rater as a check-score model with discrepancy thresholds prevented an adverse impact…
Descriptors: Scores, Computer Assisted Testing, Test Scoring Machines, Automation
Beigman Klebanov, Beata; Ramineni, Chaitanya; Kaufer, David; Yeoh, Paul; Ishizaki, Suguru – Language Testing, 2019
Essay writing is a common type of constructed-response task used frequently in standardized writing assessments. However, the impromptu timed nature of the essay writing tests has drawn increasing criticism for the lack of authenticity for real-world writing in classroom and workplace settings. The goal of this paper is to contribute evidence to a…
Descriptors: Test Validity, Writing Tests, Writing Skills, Persuasive Discourse
Zhang, Mo – ETS Research Report Series, 2013
Many testing programs use automated scoring to grade essays. One issue in automated essay scoring that has not been examined adequately is population invariance and its causes. The primary purpose of this study was to investigate the impact of sampling in model calibration on population invariance of automated scores. This study analyzed scores…
Descriptors: Automation, Scoring, Essay Tests, Sampling
Attali, Yigal; Lewis, Will; Steier, Michael – Language Testing, 2013
Automated essay scoring can produce reliable scores that are highly correlated with human scores, but is limited in its evaluation of content and other higher-order aspects of writing. The increased use of automated essay scoring in high-stakes testing underscores the need for human scoring that is focused on higher-order aspects of writing. This…
Descriptors: Scoring, Essay Tests, Reliability, High Stakes Tests
Bridgeman, Brent; Trapani, Catherine; Attali, Yigal – Applied Measurement in Education, 2012
Essay scores generated by machine and by human raters are generally comparable; that is, they can produce scores with similar means and standard deviations, and machine scores generally correlate as highly with human scores as scores from one human correlate with scores from another human. Although human and machine essay scores are highly related…
Descriptors: Scoring, Essay Tests, College Entrance Examinations, High Stakes Tests
Brown, Kevin – CEA Forum, 2015
In this article, the author describes his project to take every standardized exam English majors students take. During the summer and fall semesters of 2012, the author signed up for and took the GRE General Test, the Praxis Content Area Exam (English Language, Literature, and Composition: Content Knowledge), the Senior Major Field Tests in…
Descriptors: College Faculty, College English, Test Preparation, Standardized Tests
Attali, Yigal – Educational Testing Service, 2011
This paper proposes an alternative content measure for essay scoring, based on the "difference" in the relative frequency of a word in high-scored versus low-scored essays. The "differential word use" (DWU) measure is the average of these differences across all words in the essay. A positive value indicates the essay is using…
Descriptors: Scoring, Essay Tests, Word Frequency, Content Analysis
Quinlan, Thomas; Higgins, Derrick; Wolff, Susanne – Educational Testing Service, 2009
This report evaluates the construct coverage of the e-rater[R[ scoring engine. The matter of construct coverage depends on whether one defines writing skill, in terms of process or product. Originally, the e-rater engine consisted of a large set of components with a proven ability to predict human holistic scores. By organizing these capabilities…
Descriptors: Guides, Writing Skills, Factor Analysis, Writing Tests
Glass, Laura A.; Clause, Christopher B.; Kreiner, David S. – College Student Journal, 2007
We examined test-expectancy as it applies to fill-in-the-blank tests. We randomly assigned 60 college students to take a fill-in-the-blank vocabulary test in one of three conditions. Two groups took the test with a word bank available; we told one group but not the other that they would have a word bank. The third group took the test with no word…
Descriptors: Student Empowerment, College Students, Tests, Expectation
Hardison, Chaitra M.; Sackett, Paul R. – Applied Measurement in Education, 2008
Despite the growing use of writing assessments in standardized tests, little is known about coaching effects on writing assessments. Therefore, this study tested the effects of short-term coaching on standardized writing tests, and the transfer of those effects to other writing genres. College freshmen were randomly assigned to either training…
Descriptors: Control Groups, Group Membership, College Freshmen, Writing Tests

Powers, Donald E.; Fowles, Mary E. – Educational Assessment, 1999
Gathered judgments of 253 minority group students and 268 other college students who took the Graduate Record Examination about essay prompts being considered for use in a graduate writing test. Identified several features that underlie examinee perceptions of essay prompts, especially the extent to which prompts allow examinees to draw on their…
Descriptors: College Entrance Examinations, College Students, Essay Tests, Experience
Previous Page | Next Page »
Pages: 1 | 2