NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 202596
Since 2024281
Since 2021 (last 5 years)964
Since 2016 (last 10 years)2282
Since 2006 (last 20 years)3752
Showing 1 to 15 of 3,752 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Louise Badham – Oxford Review of Education, 2025
Different sources of assessment evidence are reviewed during International Baccalaureate (IB) grade awarding to convert marks into grades and ensure fair results for students. Qualitative and quantitative evidence are analysed to determine grade boundaries, with statistical evidence weighed against examiner judgement and teachers' feedback on…
Descriptors: Advanced Placement Programs, Grading, Interrater Reliability, Evaluative Thinking
Peer reviewed Peer reviewed
Direct linkDirect link
Jae-Sang Han; Hyun-Joo Kim – Journal of Science Education and Technology, 2025
This study explores the potential to enhance the performance of convolutional neural networks (CNNs) for automated scoring of kinematic graph answers through data augmentation using Deep Convolutional Generative Adversarial Networks (DCGANs). By developing and fine-tuning a DCGAN model to generate high-quality graph images, we explored its…
Descriptors: Performance, Automation, Scoring, Models
Saenz, David Arron – Online Submission, 2023
There is a vast body of literature documenting the positive impacts that rater training and calibration sessions have on inter-rater reliability as research indicates several factors including frequency and timing play crucial roles towards ensuring inter-rater reliability. Additionally, increasing amounts research indicate possible links in…
Descriptors: Interrater Reliability, Scoring, Training, Scoring Rubrics
Peer reviewed Peer reviewed
Direct linkDirect link
Keshav Panray Jungbadoor; Xi Hong; Liu Liu; Yunan Zhu; Xinni Huang; Viraiyan Teeroovengadum; Gwilym Croucher; Angel Calderon; Sara Bice; Hamish Coates – Tertiary Education and Management, 2025
This paper reports on a multiyear program of international collaborative research delivered with the aim of conceptualising, validating and prototyping rubrics for evaluating and reporting university activities and outcomes relevant to the UN SDGs. The paper sets foundations by building on earlier analysis of research on university engagement with…
Descriptors: Higher Education, Universities, Sustainable Development, Scoring Rubrics
Peer reviewed Peer reviewed
Direct linkDirect link
David DiSabito; Lisa Hansen; Thomas Mennella; Josephine Rodriguez – New Directions for Teaching and Learning, 2025
This chapter investigates the integration of generative AI (GenAI), specifically ChatGPT, into institutional and course-level assessment at Western New England University. It explores the potential of GenAI to streamline the assessment process, making it more efficient, equitable, and objective. Through the development of a proprietary GenAI tool,…
Descriptors: Artificial Intelligence, Technology Uses in Education, Man Machine Systems, Educational Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Casabianca, Jodi M.; Donoghue, John R.; Shin, Hyo Jeong; Chao, Szu-Fu; Choi, Ikkyu – Journal of Educational Measurement, 2023
Using item-response theory to model rater effects provides an alternative solution for rater monitoring and diagnosis, compared to using standard performance metrics. In order to fit such models, the ratings data must be sufficiently connected in order to estimate rater effects. Due to popular rating designs used in large-scale testing scenarios,…
Descriptors: Item Response Theory, Alternative Assessment, Evaluators, Research Problems
Peer reviewed Peer reviewed
Direct linkDirect link
Dhini, Bachriah Fatwa; Girsang, Abba Suganda; Sufandi, Unggul Utan; Kurniawati, Heny – Asian Association of Open Universities Journal, 2023
Purpose: The authors constructed an automatic essay scoring (AES) model in a discussion forum where the result was compared with scores given by human evaluators. This research proposes essay scoring, which is conducted through two parameters, semantic and keyword similarities, using a SentenceTransformers pre-trained model that can construct the…
Descriptors: Computer Assisted Testing, Scoring, Writing Evaluation, Essays
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Makiko Kato – Journal of Education and Learning, 2025
This study aims to examine whether differences exist in the factors influencing the difficulty of scoring English summaries and determining scores based on the raters' attributes, and to collect candid opinions, considerations, and tentative suggestions for future improvements to the analytic rubric of summary writing for English learners. In this…
Descriptors: Writing Evaluation, Scoring, Writing Skills, English (Second Language)
Peer reviewed Peer reviewed
Direct linkDirect link
Ariely, Moriah; Nazaretsky, Tanya; Alexandron, Giora – International Journal of Artificial Intelligence in Education, 2023
Machine learning algorithms that automatically score scientific explanations can be used to measure students' conceptual understanding, identify gaps in their reasoning, and provide them with timely and individualized feedback. This paper presents the results of a study that uses Hebrew NLP to automatically score student explanations in Biology…
Descriptors: Artificial Intelligence, Algorithms, Natural Language Processing, Hebrew
Peer reviewed Peer reviewed
Direct linkDirect link
Heather D. Hussey; Tara Lehan; Kate McConnell – Learning Assistance Review, 2024
Rubrics (e.g., Valid Assessment of Learning in Undergraduate Education (VALUE) rubrics) that measure specific skills exist, and researchers have demonstrated their benefits; however, most of them were designed for use with undergraduate students. Although some rubrics have been created to assess dissertations and oral defenses, few have been…
Descriptors: Scoring Rubrics, Doctoral Programs, Doctoral Dissertations, Online Courses
Peer reviewed Peer reviewed
Direct linkDirect link
Rodgers, Emily; D'Agostino, Jerome V.; Berenbon, Rebecca; Johnson, Tracy; Winkler, Christa – Journal of Early Childhood Literacy, 2023
Running Records are thought to be an excellent formative assessment tool because they generate results that educators can use to make their teaching more responsive. Despite the technical nature of scoring Running Records and the kinds of important decisions that are attached to their analysis, few studies have investigated assessor accuracy. We…
Descriptors: Formative Evaluation, Scoring, Accuracy, Difficulty Level
Yun-Kyung Kim; Li Cai – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2025
This paper introduces an application of cross-classified item response theory (IRT) modeling to an assessment utilizing the embedded standard setting (ESS) method (Lewis & Cook). The cross-classified IRT model is used to treat both item and person effects as random, where the item effects are regressed on the target performance levels (target…
Descriptors: Standard Setting (Scoring), Item Response Theory, Test Items, Difficulty Level
Peer reviewed Peer reviewed
Direct linkDirect link
Roduta Roberts, Mary; Gotch, Chad M.; Cook, Megan; Werther, Karin; Chao, Iris C. I. – Measurement: Interdisciplinary Research and Perspectives, 2022
Performance-based assessment is a common approach to assess the development and acquisition of practice competencies among health professions students. Judgments related to the quality of performance are typically operationalized as ratings against success criteria specified within a rubric. The extent to which the rubric is understood,…
Descriptors: Protocol Analysis, Scoring Rubrics, Interviews, Performance Based Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Bamdev, Pakhi; Grover, Manraj Singh; Singla, Yaman Kumar; Vafaee, Payman; Hama, Mika; Shah, Rajiv Ratn – International Journal of Artificial Intelligence in Education, 2023
English proficiency assessments have become a necessary metric for filtering and selecting prospective candidates for both academia and industry. With the rise in demand for such assessments, it has become increasingly necessary to have the automated human-interpretable results to prevent inconsistencies and ensure meaningful feedback to the…
Descriptors: Language Proficiency, Automation, Scoring, Speech Tests
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Doewes, Afrizal; Pechenizkiy, Mykola – International Educational Data Mining Society, 2021
Scoring essays is generally an exhausting and time-consuming task for teachers. Automated Essay Scoring (AES) facilitates the scoring process to be faster and more consistent. The most logical way to assess the performance of an automated scorer is by measuring the score agreement with the human raters. However, we provide empirical evidence that…
Descriptors: Man Machine Systems, Automation, Computer Assisted Testing, Scoring
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10  |  11  |  ...  |  251