Publication Date
In 2025 | 3 |
Since 2024 | 14 |
Since 2021 (last 5 years) | 62 |
Since 2016 (last 10 years) | 171 |
Since 2006 (last 20 years) | 336 |
Descriptor
Essay Tests | 1063 |
Higher Education | 295 |
Writing Evaluation | 256 |
Scoring | 217 |
Writing Skills | 194 |
Multiple Choice Tests | 190 |
Foreign Countries | 185 |
Test Construction | 168 |
Student Evaluation | 156 |
Scores | 132 |
Test Reliability | 122 |
More ▼ |
Source
Author
Bridgeman, Brent | 15 |
Breland, Hunter M. | 12 |
White, Edward M. | 10 |
Attali, Yigal | 9 |
Brossell, Gordon | 8 |
Hoetker, James | 8 |
Wolfe, Edward W. | 8 |
Zhang, Mo | 8 |
Powers, Donald E. | 7 |
Deane, Paul | 5 |
Fowles, Mary E. | 5 |
More ▼ |
Publication Type
Education Level
Location
Indonesia | 37 |
Canada | 23 |
United Kingdom | 23 |
Florida | 16 |
California | 14 |
Australia | 8 |
Japan | 8 |
United Kingdom (England) | 8 |
Georgia | 7 |
Turkey | 7 |
China | 6 |
More ▼ |
Laws, Policies, & Programs
United States Constitution | 1 |
Assessments and Surveys
What Works Clearinghouse Rating
Does not meet standards | 1 |
Akif Avcu – Malaysian Online Journal of Educational Technology, 2025
This scope-review presents the milestones of how Hierarchical Rater Models (HRMs) become operable to used in automated essay scoring (AES) to improve instructional evaluation. Although essay evaluations--a useful instrument for evaluating higher-order cognitive abilities--have always depended on human raters, concerns regarding rater bias,…
Descriptors: Automation, Scoring, Models, Educational Assessment
Almusharraf, Norah; Alotaibi, Hind – Technology, Knowledge and Learning, 2023
Evaluating written texts is believed to be a time-consuming process that can lack consistency and objectivity. Automated essay scoring (AES) can provide solutions to some of the limitations of human scoring. This research aimed to evaluate the performance of one AES system, Grammarly, in comparison to human raters. Both approaches' performances…
Descriptors: Writing Evaluation, Writing Tests, Essay Tests, Essays
Zhan Wang; Ming Ming Chiu – Applied Linguistics, 2024
Although many studies modelled writing quality by analysing basic skills (spelling, grammar, etc.), few focused on top-down compositional strategies at the discourse level. We propose that using both narrative and argument discourse modes in an argumentative essay (a multi-discourse mode [MDM] strategy) capitalizes on their complementary…
Descriptors: Discourse Analysis, Writing (Composition), Essays, Persuasive Discourse
Chan, Kinnie Kin Yee; Bond, Trevor; Yan, Zi – Language Testing, 2023
We investigated the relationship between the scores assigned by an Automated Essay Scoring (AES) system, the Intelligent Essay Assessor (IEA), and grades allocated by trained, professional human raters to English essay writing by instigating two procedures novel to written-language assessment: the logistic transformation of AES raw scores into…
Descriptors: Computer Assisted Testing, Essays, Scoring, Scores
Larsson, Kristoffer – Scandinavian Journal of Educational Research, 2021
Educationalists have had serious problems in proposing well-grounded ways of teaching critical thinking. Reviews of the field have called for research to develop theory concerning the learning experience associated with critical thinking enhancement, as well as to explore more tailored teaching and evaluation methods. This article suggests an…
Descriptors: Teaching Methods, Essays, Critical Thinking, Evaluation Methods
Wheeler, Jordan M.; Engelhard, George; Wang, Jue – Measurement: Interdisciplinary Research and Perspectives, 2022
Objectively scoring constructed-response items on educational assessments has long been a challenge due to the use of human raters. Even well-trained raters using a rubric can inaccurately assess essays. Unfolding models measure rater's scoring accuracy by capturing the discrepancy between criterion and operational ratings by placing essays on an…
Descriptors: Accuracy, Scoring, Statistical Analysis, Models
Hosnia M. M. Ahmed; Shaymaa E. Sorour – Education and Information Technologies, 2024
Evaluating the quality of university exam papers is crucial for universities seeking institutional and program accreditation. Currently, exam papers are assessed manually, a process that can be tedious, lengthy, and in some cases, inconsistent. This is often due to the focus on assessing only the formal specifications of exam papers. This study…
Descriptors: Higher Education, Artificial Intelligence, Writing Evaluation, Natural Language Processing
Yanxuan Qu; Sandip Sinharay – ETS Research Report Series, 2024
The goal of this paper is to find better ways to estimate the internal consistency reliability of scores on tests with a specific type of design that are often encountered in practice: tests with constructed-response items clustered into sections that are not parallel or tau-equivalent, and one of the sections has only one item. To estimate the…
Descriptors: Test Reliability, Essay Tests, Construct Validity, Error of Measurement
Bejar, Isaac I.; Li, Chen; McCaffrey, Daniel – Applied Measurement in Education, 2020
We evaluate the feasibility of developing predictive models of rater behavior, that is, "rater-specific" models for predicting the scores produced by a rater under operational conditions. In the present study, the dependent variable is the score assigned to essays by a rater, and the predictors are linguistic attributes of the essays…
Descriptors: Scoring, Essays, Behavior, Predictive Measurement
Cleophas, Catherine; Hönnige, Christoph; Meisel, Frank; Meyer, Philipp – INFORMS Transactions on Education, 2023
As the COVID-19 pandemic motivated a shift to virtual teaching, exams have increasingly moved online too. Detecting cheating through collusion is not easy when tech-savvy students take online exams at home and on their own devices. Such online at-home exams may tempt students to collude and share materials and answers. However, online exams'…
Descriptors: Computer Assisted Testing, Cheating, Identification, Essay Tests
Jussi S. Jauhiainen; Agustin Bernardo Garagorry Guerra – Journal of Information Technology Education: Innovations in Practice, 2025
Aim/Purpose: This article investigates the process of identifying and correcting hallucinations in ChatGPT-4's recall of student-written responses as well as its evaluation of these responses, and provision of feedback. Effective prompting is examined to enhance the pre-evaluation, evaluation, and post-evaluation stages. Background: Advanced Large…
Descriptors: Artificial Intelligence, Student Evaluation, Writing Evaluation, Feedback (Response)
Angeline S. Lillard; Jessica Taggart – College Teaching, 2024
In large lecture courses, it can be challenging to imagine assessing student learning in ways other than multiple-choice exams and traditional point-based grading. Inspired by major pedagogical principles shared by Maria Montessori and Thomas Jefferson and supported by current understandings of effective teaching, assessment was reimagined in a…
Descriptors: College Faculty, Undergraduate Students, Teaching Assistants, Child Psychology
Naima Debbar – International Journal of Contemporary Educational Research, 2024
Intelligent systems of essay grading constitute important tools for educational technologies. They can significantly replace the manual scoring efforts and provide instructional feedback as well. These systems typically include two main parts: a feature extractor and an automatic grading model. The latter is generally based on computational and…
Descriptors: Test Scoring Machines, Computer Uses in Education, Artificial Intelligence, Essay Tests
Wendler, Cathy; Glazer, Nancy; Bridgeman, Brent – Applied Measurement in Education, 2020
Efficient constructed response (CR) scoring requires both accuracy and speed from human raters. This study was designed to determine if setting scoring rate expectations would encourage raters to score at a faster pace, and if so, if there would be differential effects on scoring accuracy for raters who score at different rates. Three rater groups…
Descriptors: Scoring, Expectation, Accuracy, Time
Aulia Dhita Nanda; Rusdi Hasan; Akhmad Sukri; Marheny Lukitasari; Alice Tonido Rivera – Journal of Biological Education Indonesia (Jurnal Pendidikan Biologi Indonesia), 2023
Higher-order thinking skills (HOTS) is one of the skills needed for 21st century challenges, especially for students. The aim of this study was to describe HOTS in students, especially the cognitive domain of analyzing and evaluating. This is a descriptive quantitative study employing a Pretest-Posttest One Group research design. The experiment…
Descriptors: Thinking Skills, Problem Based Learning, Teaching Methods, Skill Development