Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 1 |
Since 2006 (last 20 years) | 8 |
Descriptor
Source
Author
Publication Type
Reports - Evaluative | 32 |
Journal Articles | 14 |
Speeches/Meeting Papers | 14 |
Reports - Research | 1 |
Education Level
Higher Education | 2 |
Postsecondary Education | 2 |
Adult Education | 1 |
Audience
Location
Australia | 1 |
Sweden | 1 |
United Kingdom | 1 |
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 1 |
SAT (College Admission Test) | 1 |
Test of English as a Foreign… | 1 |
What Works Clearinghouse Rating
Reichle, Erik D.; Drieghe, Denis – Journal of Experimental Psychology: Learning, Memory, and Cognition, 2015
There is an ongoing debate about whether fixation durations during reading are only influenced by the processing difficulty of the words being fixated (i.e., the serial-attention hypothesis) or whether they are also influenced by the processing difficulty of the previous and/or upcoming words (i.e., the attention-gradient hypothesis). This article…
Descriptors: Reading, Eye Movements, Error of Measurement, Difficulty Level
Abrandt Dahlgren, Madeleine; Fenwick, Tara; Hopwood, Nick – Teaching in Higher Education, 2016
Despite the widespread interest in using and researching simulation in higher education, little discussion has yet to address a key pedagogical concern: difficulty. A "sociomaterial" view of learning, explained in this paper, goes beyond cognitive considerations to highlight dimensions of material, situational, representational and…
Descriptors: Simulation, Higher Education, Social Theories, Experiential Learning
Attali, Yigal; Saldivia, Luis; Jackson, Carol; Schuppan, Fred; Wanamaker, Wilbur – ETS Research Report Series, 2014
Previous investigations of the ability of content experts and test developers to estimate item difficulty have, for themost part, produced disappointing results. These investigations were based on a noncomparative method of independently rating the difficulty of items. In this article, we argue that, by eliciting comparative judgments of…
Descriptors: Test Items, Difficulty Level, Comparative Analysis, College Entrance Examinations
Wang, Wen-Chung; Jin, Kuan-Yu – Educational and Psychological Measurement, 2010
In this study, the authors extend the standard item response model with internal restrictions on item difficulty (MIRID) to fit polytomous items using cumulative logits and adjacent-category logits. Moreover, the new model incorporates discrimination parameters and is rooted in a multilevel framework. It is a nonlinear mixed model so that existing…
Descriptors: Difficulty Level, Test Items, Item Response Theory, Generalization
Thompson, Nathan A. – Practical Assessment, Research & Evaluation, 2011
Computerized classification testing (CCT) is an approach to designing tests with intelligent algorithms, similar to adaptive testing, but specifically designed for the purpose of classifying examinees into categories such as "pass" and "fail." Like adaptive testing for point estimation of ability, the key component is the…
Descriptors: Adaptive Testing, Computer Assisted Testing, Classification, Probability
Weitzman, R. A. – Educational and Psychological Measurement, 2009
Building on the Kelley and Gulliksen versions of classical test theory, this article shows that a logistic model having only a single item parameter can account for varying item discrimination, as well as difficulty, by using item-test correlations to adjust incorrect-correct (0-1) item responses prior to an initial model fit. The fit occurs…
Descriptors: Item Response Theory, Test Items, Difficulty Level, Test Bias
Dawber, Teresa; Rogers, W. Todd; Carbonaro, Michael – Alberta Journal of Educational Research, 2009
Lord (1980) proposed formulas that provide direct relationships between IRT discrimination and difficulty parameters and conventional item statistics. The purpose of the present study was to determine the robustness of the formulas beyond the initial and restrictive conditions identified by Lord. Simulation and real achievement data were employed.…
Descriptors: Test Items, Simulation, Achievement Tests, Robustness (Statistics)
Meyers, Jason L.; Miller, G. Edward; Way, Walter D. – Applied Measurement in Education, 2009
In operational testing programs using item response theory (IRT), item parameter invariance is threatened when an item appears in a different location on the live test than it did when it was field tested. This study utilizes data from a large state's assessments to model change in Rasch item difficulty (RID) as a function of item position change,…
Descriptors: Test Items, Test Content, Testing Programs, Simulation

Kim, Seock-Ho; Cohen, Allan S. – Applied Psychological Measurement, 1998
Compared three methods for developing a common metric under item response theory through simulation. For smaller numbers of common items, linking using the characteristic curve method yielded smaller root mean square differences for both item discrimination and difficulty parameters. For larger numbers of common items, the three methods were…
Descriptors: Comparative Analysis, Difficulty Level, Item Response Theory, Simulation
Mazor, Kathleen M.; And Others – 1991
The Mantel-Haenszel (MH) procedure has become one of the most popular procedures for detecting differential item functioning. Valid results with relatively small numbers of examinees represent one of the advantages typically attributed to this procedure. In this study, examinee item responses were simulated to contain differentially functioning…
Descriptors: Difficulty Level, Item Bias, Item Response Theory, Sample Size
Lau, C. Allen; Wang, Tianyou – 1999
A study was conducted to extend the sequential probability ratio testing (SPRT) procedure with the polytomous model under some practical constraints in computerized classification testing (CCT), such as methods to control item exposure rate, and to study the effects of other variables, including item information algorithms, test difficulties, item…
Descriptors: Algorithms, Computer Assisted Testing, Difficulty Level, Item Banks

Wang, Wen-chung; Wilson, Mark; Adams, Raymond J. – Journal of Outcome Measurement, 1998
Another Rasch approach to the measurement of change, the multidimensional random coefficient multinomial logit model (MRCML), is proposed. The MRCML model can be applied to polytomous items and the investigation of variations in item difficulties. Some simulation studies demonstrate good parameter recovery for the MRCML model under various testing…
Descriptors: Change, Difficulty Level, Individual Differences, Item Response Theory
Clauser, Brian E.; And Others – 1991
Item bias has been a major concern for test developers during recent years. The Mantel-Haenszel statistic has been among the preferred methods for identifying biased items. The statistic's performance in identifying uniform bias in simulated data modeled by producing various levels of difference in the (item difficulty) b-parameter for reference…
Descriptors: Comparative Testing, Difficulty Level, Item Bias, Item Response Theory
Dodeen, Hamzeh – Journal of Educational Measurement, 2004
The effect of item parameters (discrimination, difficulty, and level of guessing) on the item-fit statistic was investigated using simulated dichotomous data. Nine tests were simulated using 1,000 persons, 50 items, three levels of item discrimination, three levels of item difficulty, and three levels of guessing. The item fit was estimated using…
Descriptors: Item Response Theory, Difficulty Level, Test Items, Guessing (Tests)
Gershon, Richard; Bergstrom, Betty – 1995
When examinees are allowed to review responses on an adaptive test, can they "cheat" the adaptive algorithm in order to take an easier test and improve their performance? Theoretically, deliberately answering items incorrectly will lower the examinee ability estimate and easy test items will be administered. If review is then allowed,…
Descriptors: Adaptive Testing, Algorithms, Cheating, Computer Assisted Testing