NotesFAQContact Us
Collection
Advanced
Search Tips
What Works Clearinghouse Rating
Showing 1 to 15 of 147 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Tia M. Fechter; Heeyeon Yoon – Language Testing, 2024
This study evaluated the efficacy of two proposed methods in an operational standard-setting study conducted for a high-stakes language proficiency test of the U.S. government. The goal was to seek low-cost modifications to the existing Yes/No Angoff method to increase the validity and reliability of the recommended cut scores using a convergent…
Descriptors: Standard Setting, Language Proficiency, Language Tests, Evaluation Methods
Peer reviewed Peer reviewed
Direct linkDirect link
Fu Chen; Ying Cui; Alina Lutsyk-King; Yizhu Gao; Xiaoxiao Liu; Maria Cutumisu; Jacqueline P. Leighton – Education and Information Technologies, 2024
Post-secondary data literacy education is critical to students' academic and career success. However, the literature has not adequately addressed the conceptualization and assessment of data literacy for post-secondary students. In this study, we introduced a novel digital performance-based assessment for teaching and evaluating post-secondary…
Descriptors: Performance Based Assessment, College Students, Information Literacy, Evaluation Methods
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Anani Sarab, Mohammad Reza; Rahmani, Simindokht – International Journal of Language Testing, 2023
Language testing and assessment have grown in popularity and gained significance in the last few decades, and there is a rising need for assessment literate stakeholders in the field of language education. As teachers play a major role in assessing students, there is a need to make sure they have the right level of assessment knowledge and skills…
Descriptors: Language Tests, Literacy, Second Language Learning, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Lewis, Jennifer; Lim, Hwanggyu; Padellaro, Frank; Sireci, Stephen G.; Zenisky, April L. – Educational Measurement: Issues and Practice, 2022
Setting cut scores on (MSTs) is difficult, particularly when the test spans several grade levels, and the selection of items from MST panels must reflect the operational test specifications. In this study, we describe, illustrate, and evaluate three methods for mapping panelists' Angoff ratings into cut scores on the scale underlying an MST. The…
Descriptors: Cutting Scores, Adaptive Testing, Test Items, Item Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Guher Gorgun; Okan Bulut – Educational Measurement: Issues and Practice, 2025
Automatic item generation may supply many items instantly and efficiently to assessment and learning environments. Yet, the evaluation of item quality persists to be a bottleneck for deploying generated items in learning and assessment settings. In this study, we investigated the utility of using large-language models, specifically Llama 3-8B, for…
Descriptors: Artificial Intelligence, Quality Control, Technology Uses in Education, Automation
Gill, Tim – Research Matters, 2022
In Comparative Judgement (CJ) exercises, examiners are asked to look at a selection of candidate scripts (with marks removed) and order them in terms of which they believe display the best quality. By including scripts from different examination sessions, the results of these exercises can be used to help with maintaining standards. Results from…
Descriptors: Comparative Analysis, Decision Making, Scripts, Standards
Peer reviewed Peer reviewed
Direct linkDirect link
An, Lily Shiao; Ho, Andrew Dean; Davis, Laurie Laughlin – Educational Measurement: Issues and Practice, 2022
Technical documentation for educational tests focuses primarily on properties of individual scores at single points in time. Reliability, standard errors of measurement, item parameter estimates, fit statistics, and linking constants are standard technical features that external stakeholders use to evaluate items and individual scale scores.…
Descriptors: Documentation, Scores, Evaluation Methods, Longitudinal Studies
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Karen Leary Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Assessment is a topic of concern to all stakeholders in our educational system. Pattern Based Questions are an assessment tool which is an alternative to the standardized assessment tool, and they are based on generative learning pedagogy, which shows promise in engaging all learners and usefulness in teaching and learning but validity has not yet…
Descriptors: Undergraduate Students, College Mathematics, Mathematics Skills, Thinking Skills
Peer reviewed Peer reviewed
Direct linkDirect link
Philomina Abena Anyidoho; Rebecca Berenbon; Bridget McHugh – International Journal of Training and Development, 2024
Many workforce development training programmes use learning gains as a measure of programme effectiveness. However, research on K-12 education suggests that posttest scores may be influenced by pretesting effects. Pretesting may improve posttest performance by giving learners preknowledge of posttest content. Alternatively, pretesting may enhance…
Descriptors: Trainees, Trainers, Labor Force Development, High Stakes Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Meng, Yaru; Fu, Hua – Modern Language Journal, 2023
The distinguishing feature of dynamic assessment (DA) is the dialectical integration of assessment and instruction. However, how to design the targeted instruction or mediation has been relatively underexplored. To address this gap, this study proposes the attribute-based mediation model (AMM), an English-as-a-foreign-language listening mediation…
Descriptors: Evaluation Methods, Teaching Methods, Models, English (Second Language)
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Toker, Turker – International Journal of Curriculum and Instruction, 2023
Achievement tests are among the most widely used data collection tools to measure the knowledge and skill levels of individuals. For this reason, the existence of valid and reliable achievement tests that can perfectly reveal the competencies that a person should have in any discipline is of great importance. The purpose of this research is to…
Descriptors: Basic Skills, Evaluation Methods, Test Items, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Thomas Bickerton, Robert; Sangwin, Chris J. – International Journal of Mathematical Education in Science and Technology, 2022
We discuss a practical method for assessing mathematical proof online. We examine the use of faded worked examples and reading comprehension questions to understand proof. By breaking down a given proof, we formulate a checklist that can be used to generate comprehension questions which can be assessed automatically online. We then provide some…
Descriptors: Mathematics Instruction, Validity, Mathematical Logic, Evaluation Methods
Parry, James R. – Online Submission, 2020
This paper presents research and provides a method to ensure that parallel assessments, that are generated from a large test-item database, maintain equitable difficulty and content coverage each time the assessment is presented. To maintain fairness and validity it is important that all instances of an assessment, that is intended to test the…
Descriptors: Culture Fair Tests, Difficulty Level, Test Items, Test Validity
Thapelo Ncube Whitfield – ProQuest LLC, 2021
Student Experience surveys are used to measure student attitudes towards their campus as well as to initiate conversations for institutional change. Validity evidence to support the interpretations of these surveys' results, however, is lacking. The first purpose of this study was to compare three Differential Item Functioning (DIF) methods on…
Descriptors: College Students, Student Surveys, Student Experience, Student Attitudes
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Lynch, Sarah – Practical Assessment, Research & Evaluation, 2022
In today's digital age, tests are increasingly being delivered on computers. Many of these computer-based tests (CBTs) have been adapted from paper-based tests (PBTs). However, this change in mode of test administration has the potential to introduce construct-irrelevant variance, affecting the validity of score interpretations. Because of this,…
Descriptors: Computer Assisted Testing, Tests, Scores, Scoring
Previous Page | Next Page ยป
Pages: 1  |  2  |  3  |  4  |  5  |  6  |  7  |  8  |  9  |  10