Publication Date
In 2025 | 0 |
Since 2024 | 2 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 33 |
Descriptor
Source
Author
Thurlow, Martha | 8 |
Bielinski, John | 2 |
El Sawaf, Hamdy | 2 |
Erickson, Ron | 2 |
Erickson, Ronald | 2 |
Haigh, John | 2 |
Leitzel, Thomas C. | 2 |
Liu, Kristin | 2 |
Minnema, Jane | 2 |
Spicuzza, Richard | 2 |
Trimble, Scott | 2 |
More ▼ |
Publication Type
Education Level
Audience
Researchers | 3 |
Practitioners | 2 |
Location
Canada | 6 |
United Kingdom | 6 |
Australia | 2 |
Hong Kong | 2 |
Kentucky | 2 |
Maryland | 2 |
Taiwan | 2 |
California | 1 |
Cambodia | 1 |
China | 1 |
Colombia | 1 |
More ▼ |
Laws, Policies, & Programs
Elementary and Secondary… | 5 |
Individuals with Disabilities… | 2 |
No Child Left Behind Act 2001 | 1 |
Assessments and Surveys
ACT Assessment | 1 |
Advanced Placement… | 1 |
Pediatric Evaluation of… | 1 |
Program for International… | 1 |
Sequential Tests of… | 1 |
Stanford Achievement Tests | 1 |
What Works Clearinghouse Rating
Stefanie A. Wind; Yangmeng Xu – Educational Assessment, 2024
We explored three approaches to resolving or re-scoring constructed-response items in mixed-format assessments: rater agreement, person fit, and targeted double scoring (TDS). We used a simulation study to consider how the three approaches impact the psychometric properties of student achievement estimates, with an emphasis on person fit. We found…
Descriptors: Interrater Reliability, Error of Measurement, Evaluation Methods, Examiners
Atkinson, Cathy; Barrow, Joanna; Norris, Sarah – Educational Psychology in Practice, 2022
Assessment is one of the five functions of the educational psychologist's (EP's) role, yet there is a dearth of research exploring its distinctive contribution to school-based practice, and a lack of definition about what it is. In this study, the assessment practices of EPs were compared with those of other educational professionals who had…
Descriptors: Educational Psychology, School Psychologists, Evaluation Methods, Educational Testing
Sainan Xu; Jing Lu; Jiwei Zhang; Chun Wang; Gongjun Xu – Grantee Submission, 2024
With the growing attention on large-scale educational testing and assessment, the ability to process substantial volumes of response data becomes crucial. Current estimation methods within item response theory (IRT), despite their high precision, often pose considerable computational burdens with large-scale data, leading to reduced computational…
Descriptors: Educational Assessment, Bayesian Statistics, Statistical Inference, Item Response Theory
Socha, Alan; DeMars, Christine E.; Zilberberg, Anna; Phan, Ha – International Journal of Testing, 2015
The Mantel-Haenszel (MH) procedure is commonly used to detect items that function differentially for groups of examinees from various demographic and linguistic backgrounds--for example, in international assessments. As in some other DIF methods, the total score is used to match examinees on ability. In thin matching, each of the total score…
Descriptors: Test Items, Educational Testing, Evaluation Methods, Ability Grouping
Zwick, Rebecca – ETS Research Report Series, 2012
Differential item functioning (DIF) analysis is a key component in the evaluation of the fairness and validity of educational tests. The goal of this project was to review the status of ETS DIF analysis procedures, focusing on three aspects: (a) the nature and stringency of the statistical rules used to flag items, (b) the minimum sample size…
Descriptors: Test Bias, Sample Size, Bayesian Statistics, Evaluation Methods
Emenogu, Barnabas C.; Falenchuk, Olesya; Childs, Ruth A. – Alberta Journal of Educational Research, 2010
Most implementations of the Mantel-Haenszel differential item functioning procedure delete records with missing responses or replace missing responses with scores of 0. These treatments of missing data make strong assumptions about the causes of the missing data. Such assumptions may be particularly problematic when groups differ in their patterns…
Descriptors: Foreign Countries, Test Bias, Test Items, Educational Testing
Reckase, Mark D.; Xu, Jing-Ru – Educational and Psychological Measurement, 2015
How to compute and report subscores for a test that was originally designed for reporting scores on a unidimensional scale has been a topic of interest in recent years. In the research reported here, we describe an application of multidimensional item response theory to identify a subscore structure in a test designed for reporting results using a…
Descriptors: English, Language Skills, English Language Learners, Scores
Marshall, Jeffery H.; Chinna, Ung; Hok, Ung Ngo; Tinon, Souer; Veasna, Meung; Nissay, Put – Educational Assessment, Evaluation and Accountability, 2012
The global spread of national assessment testing activities, and the growing pressure to move beyond basic measures of participation in educational monitoring, means that student achievement measures are likely to become increasingly relevant indicators of systemic progress in the developing world. Using data from the CESSP project in Cambodia,…
Descriptors: Foreign Countries, Academic Achievement, Developing Nations, Evaluation Methods
Lee, Won-Chan; Ban, Jae-Chun – Applied Measurement in Education, 2010
Various applications of item response theory often require linking to achieve a common scale for item parameter estimates obtained from different groups. This article used a simulation to examine the relative performance of four different item response theory (IRT) linking procedures in a random groups equating design: concurrent calibration with…
Descriptors: Item Response Theory, Simulation, Comparative Analysis, Measurement Techniques
Ventouras, Errikos; Triantis, Dimos; Tsiakas, Panagiotis; Stergiopoulos, Charalampos – Computers & Education, 2010
The aim of the present research was to compare the use of multiple-choice questions (MCQs) as an examination method, to the examination based on constructed-response questions (CRQs). Despite that MCQs have an advantage concerning objectivity in the grading process and speed in production of results, they also introduce an error in the final…
Descriptors: Computer Assisted Instruction, Scoring, Grading, Comparative Analysis
Armstrong, Ronald D.; Shi, Min – Journal of Educational Measurement, 2009
This article demonstrates the use of a new class of model-free cumulative sum (CUSUM) statistics to detect person fit given the responses to a linear test. The fundamental statistic being accumulated is the likelihood ratio of two probabilities. The detection performance of this CUSUM scheme is compared to other model-free person-fit statistics…
Descriptors: Probability, Simulation, Models, Psychometrics
Fan, Ya-Ching; Wang, Tzu-Hua; Wang, Kuo-Hua – Computers & Education, 2011
This research investigates the effect of a web-based model, named "Practicing, Reflecting, and Revising with Web-based Assessment and Test Analysis system (P2R-WATA) Assessment Literacy Development Model," on enhancing assessment knowledge and perspectives of secondary in-service teachers, and adopts a single group experimental research…
Descriptors: Research Design, Test Items, Summer Programs, Prior Learning
Bryant, Darren A.; Carless, David R. – Educational Research for Policy and Practice, 2010
The literature suggests that peer assessment contributes to the development of student learning and promotes ownership of assessment processes. These claims emerge from research conducted primarily in Western contexts. This exploratory paper reports on the perspectives that a class of Hong Kong primary school students and their teachers have on…
Descriptors: Feedback (Response), Peer Evaluation, Foreign Countries, Language Proficiency
Johnson, Jeff, Ed. – Educational Testing Service, 2009
In four articles adapted from the Educational Testing Service (ETS) Research Report Series, Issue 2 of ETS Research Spotlight provides a small taste of the range of assessment-related research capabilities of the ETS Research and Development Division. Those articles cover assessment-related research aimed at developing models of student learning,…
Descriptors: Basic Writing, Educational Testing, Research Reports, Measures (Individuals)
Holling, Heinz; Bertling, Jonas P.; Zeuch, Nina – Studies in Educational Evaluation, 2009
Mathematical word problems represent a common item format for assessing student competencies. Automatic item generation (AIG) is an effective way of constructing many items with predictable difficulties, based on a set of predefined task parameters. The current study presents a framework for the automatic generation of probability word problems…
Descriptors: Word Problems (Mathematics), Probability, Automation, College Students