Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 4 |
Descriptor
Computer Assisted Testing | 5 |
Essay Tests | 5 |
Scoring | 5 |
College Entrance Examinations | 4 |
Automation | 2 |
Graduate Study | 2 |
Standardized Tests | 2 |
Writing Evaluation | 2 |
Writing Skills | 2 |
Abstract Reasoning | 1 |
African Americans | 1 |
More ▼ |
Author
Breyer, F. Jay | 1 |
Bridgeman, Brent | 1 |
Brown, Kevin | 1 |
Camp, Roberta | 1 |
Carlson, Sybil B. | 1 |
Higgins, Derrick | 1 |
Quinlan, Thomas | 1 |
Ramineni, Chaitanya | 1 |
Rupp, André A. | 1 |
Williamson, David | 1 |
Wolff, Susanne | 1 |
More ▼ |
Publication Type
Journal Articles | 3 |
Reports - Research | 3 |
Reports - Evaluative | 2 |
Speeches/Meeting Papers | 1 |
Education Level
Higher Education | 4 |
Postsecondary Education | 3 |
Elementary Secondary Education | 1 |
Audience
Researchers | 1 |
Location
China | 1 |
India | 1 |
Japan | 1 |
South Korea | 1 |
Taiwan | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Graduate Record Examinations | 5 |
Test of English as a Foreign… | 2 |
Praxis Series | 1 |
What Works Clearinghouse Rating
Breyer, F. Jay; Rupp, André A.; Bridgeman, Brent – ETS Research Report Series, 2017
In this research report, we present an empirical argument for the use of a contributory scoring approach for the 2-essay writing assessment of the analytical writing section of the "GRE"® test in which human and machine scores are combined for score creation at the task and section levels. The approach was designed to replace a currently…
Descriptors: College Entrance Examinations, Scoring, Essay Tests, Writing Evaluation
Ramineni, Chaitanya; Williamson, David – ETS Research Report Series, 2018
Notable mean score differences for the "e-rater"® automated scoring engine and for humans for essays from certain demographic groups were observed for the "GRE"® General Test in use before the major revision of 2012, called rGRE. The use of e-rater as a check-score model with discrepancy thresholds prevented an adverse impact…
Descriptors: Scores, Computer Assisted Testing, Test Scoring Machines, Automation
Brown, Kevin – CEA Forum, 2015
In this article, the author describes his project to take every standardized exam English majors students take. During the summer and fall semesters of 2012, the author signed up for and took the GRE General Test, the Praxis Content Area Exam (English Language, Literature, and Composition: Content Knowledge), the Senior Major Field Tests in…
Descriptors: College Faculty, College English, Test Preparation, Standardized Tests
Quinlan, Thomas; Higgins, Derrick; Wolff, Susanne – Educational Testing Service, 2009
This report evaluates the construct coverage of the e-rater[R[ scoring engine. The matter of construct coverage depends on whether one defines writing skill, in terms of process or product. Originally, the e-rater engine consisted of a large set of components with a proven ability to predict human holistic scores. By organizing these capabilities…
Descriptors: Guides, Writing Skills, Factor Analysis, Writing Tests
Carlson, Sybil B.; Camp, Roberta – 1985
This paper reports on Educational Testing Service research studies investigating the parameters critical to reliability and validity in both the direct and indirect writing ability assessment of higher education applicants. The studies involved: (1) formulating an operational definition of writing competence; (2) designing and pretesting writing…
Descriptors: College Entrance Examinations, Computer Assisted Testing, English (Second Language), Essay Tests