Publication Date
In 2025 | 1 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 15 |
Since 2016 (last 10 years) | 37 |
Since 2006 (last 20 years) | 70 |
Descriptor
Essay Tests | 256 |
Writing Evaluation | 256 |
Scoring | 86 |
Higher Education | 84 |
Writing Skills | 83 |
Writing (Composition) | 57 |
Holistic Evaluation | 49 |
Test Reliability | 46 |
Interrater Reliability | 40 |
Test Validity | 37 |
Writing Tests | 37 |
More ▼ |
Source
Author
Brossell, Gordon | 8 |
White, Edward M. | 8 |
Breland, Hunter M. | 7 |
Hoetker, James | 5 |
Ackerman, Terry A. | 3 |
Quellmalz, Edys | 3 |
Welch, Catherine | 3 |
Wilson, Joshua | 3 |
Anderson, Judith A. | 2 |
Attali, Yigal | 2 |
Bamberg, Betty | 2 |
More ▼ |
Publication Type
Education Level
Audience
Researchers | 21 |
Practitioners | 16 |
Teachers | 9 |
Location
California | 8 |
Florida | 7 |
Canada | 6 |
Iran | 3 |
North Carolina | 3 |
Australia | 2 |
New Jersey | 2 |
Nigeria | 2 |
South Korea | 2 |
Texas | 2 |
Turkey | 2 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Does not meet standards | 1 |
Hosnia M. M. Ahmed; Shaymaa E. Sorour – Education and Information Technologies, 2024
Evaluating the quality of university exam papers is crucial for universities seeking institutional and program accreditation. Currently, exam papers are assessed manually, a process that can be tedious, lengthy, and in some cases, inconsistent. This is often due to the focus on assessing only the formal specifications of exam papers. This study…
Descriptors: Higher Education, Artificial Intelligence, Writing Evaluation, Natural Language Processing
Jussi S. Jauhiainen; Agustin Bernardo Garagorry Guerra – Journal of Information Technology Education: Innovations in Practice, 2025
Aim/Purpose: This article investigates the process of identifying and correcting hallucinations in ChatGPT-4's recall of student-written responses as well as its evaluation of these responses, and provision of feedback. Effective prompting is examined to enhance the pre-evaluation, evaluation, and post-evaluation stages. Background: Advanced Large…
Descriptors: Artificial Intelligence, Student Evaluation, Writing Evaluation, Feedback (Response)
Almusharraf, Norah; Alotaibi, Hind – Technology, Knowledge and Learning, 2023
Evaluating written texts is believed to be a time-consuming process that can lack consistency and objectivity. Automated essay scoring (AES) can provide solutions to some of the limitations of human scoring. This research aimed to evaluate the performance of one AES system, Grammarly, in comparison to human raters. Both approaches' performances…
Descriptors: Writing Evaluation, Writing Tests, Essay Tests, Essays
Zhang, Haoran; Litman, Diane – Grantee Submission, 2020
While automated essay scoring (AES) can reliably grade essays at scale, automated writing evaluation (AWE) additionally provides formative feedback to guide essay revision. However, a neural AES typically does not provide useful feature representations for supporting AWE. This paper presents a method for linking AWE and neural AES, by extracting…
Descriptors: Computer Assisted Testing, Scoring, Essay Tests, Writing Evaluation
Kim, Hyunwoo – Language Testing in Asia, 2020
The halo effect is raters' undesirable tendency to assign more similar ratings across rating criteria than they should. The impacts of the halo effect on ratings have been studied in rater-mediated L2 writing assessment. Little is known, however, about the extent to which rating criteria order in analytic rating scales is associated with the…
Descriptors: Evaluation Criteria, Writing Evaluation, Second Languages, Item Response Theory
Yan, Xun; Chuang, Ping-Lin – Language Testing, 2023
This study employed a mixed-methods approach to examine how rater performance develops during a semester-long rater certification program for an English as a Second Language (ESL) writing placement test at a large US university. From 2016 to 2018, we tracked three groups of novice raters (n = 30) across four rounds in the certification program.…
Descriptors: Evaluators, Interrater Reliability, Item Response Theory, Certification
Chan, Kinnie Kin Yee; Bond, Trevor; Yan, Zi – Language Testing, 2023
We investigated the relationship between the scores assigned by an Automated Essay Scoring (AES) system, the Intelligent Essay Assessor (IEA), and grades allocated by trained, professional human raters to English essay writing by instigating two procedures novel to written-language assessment: the logistic transformation of AES raw scores into…
Descriptors: Computer Assisted Testing, Essays, Scoring, Scores
Leanne Havis – Journal of Response to Writing, 2022
Metacognition emphasizes an awareness and understanding of one's thoughts and cognitive processes, along with management of cognition through multiple strategies, including organizing, monitoring, and adapting. Before students can truly become effective writers, they must develop an appreciation for the amount of planning, organization, and…
Descriptors: Undergraduate Students, College Freshmen, Writing Assignments, Writing Evaluation
Atilgan, Hakan – Eurasian Journal of Educational Research, 2019
Purpose: This study intended to examine the generalizability and reliability of essay ratings within the scope of the generalizability (G) theory. Specifically, the effect of raters on the generalizability and reliability of students' essay ratings was examined. Furthermore, variations of the generalizability and reliability coefficients with…
Descriptors: Foreign Countries, Essay Tests, Test Reliability, Interrater Reliability
Erguvan, Inan Deniz; Aksu Dunya, Beyza – Language Testing in Asia, 2020
This study examined the rater severity of instructors using a multi-trait rubric in a freshman composition course offered in a private university in Kuwait. Use of standardized multi-trait rubrics is a recent development in this course and student feedback and anchor papers provided by instructors for each essay exam necessitated the assessment of…
Descriptors: Foreign Countries, College Freshmen, Freshman Composition, Writing Evaluation
Arefsadr, Sajjad; Babaii, Esmat; Hashemi, Mohammad Reza – International Journal of Language Testing, 2022
This study explored possible reasons why IELTS candidates usually score low in writing by investigating the effects of two different test designs and scoring criteria on Iranian IELTS candidates' obtained grades in IELTS and World Englishes (WEs) essay writing tests. To this end, first, a WEs essay writing test was preliminarily designed. Then, 17…
Descriptors: English (Second Language), Second Language Learning, Language Tests, Writing Evaluation
Reinertsen, Nathanael – English in Australia, 2018
The difference in how humans read and how Automated Essay Scoring (AES) systems process written language leads to a situation where a portion of student responses will be comprehensible to human markers, but unable to be parsed by AES systems. This paper examines a number of pieces of student writing that were marked by trained human markers, but…
Descriptors: Qualitative Research, Writing Evaluation, Essay Tests, Computer Assisted Testing
Michelle Herridge – ProQuest LLC, 2021
Evaluation of student written work during summative assessments is an important and critical task for instructors at all educational levels. Nevertheless, few research studies exist that provide insights into how different instructors approach this task. Chemistry faculty (FIs) and graduate student instructors (GSIs) regularly engage in the…
Descriptors: Science Instruction, Chemistry, College Faculty, Teaching Assistants
Weejeong Jeong – ProQuest LLC, 2022
This study is an investigation of the effects of linguistic features on quality of second language (L2) writers' essays for writing course placement at Indiana University Bloomington (IUB), and by implication at other universities and colleges. This study addresses the following research questions: (1) To what extent do selected linguistic…
Descriptors: Linguistics, Language Usage, Second Language Learning, College Students
Wilson, Joshua; Huang, Yue; Palermo, Corey; Beard, Gaysha; MacArthur, Charles A. – International Journal of Artificial Intelligence in Education, 2021
This study examined a naturalistic, districtwide implementation of an automated writing evaluation (AWE) software program called "MI Write" in elementary schools. We specifically examined the degree to which aspects of MI Write were implemented, teacher and student attitudes towards MI Write, and whether MI Write usage along with other…
Descriptors: Automation, Writing Evaluation, Feedback (Response), Computer Software