NotesFAQContact Us
Collection
Advanced
Search Tips
Publication Date
In 20250
Since 20240
Since 2021 (last 5 years)0
Since 2016 (last 10 years)0
Since 2006 (last 20 years)26
Publication Type
Reports - Evaluative39
Journal Articles32
Speeches/Meeting Papers4
Education Level
High Schools1
Audience
Practitioners1
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing 1 to 15 of 39 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Suh, Youngsuk; Bolt, Daniel M. – Journal of Educational Measurement, 2011
In multiple-choice items, differential item functioning (DIF) in the correct response may or may not be caused by differentially functioning distractors. Identifying distractors as causes of DIF can provide valuable information for potential item revision or the design of new test items. In this paper, we examine a two-step approach based on…
Descriptors: Test Items, Test Bias, Multiple Choice Tests, Simulation
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Elosua, Paula – Psicologica: International Journal of Methodology and Experimental Psychology, 2011
Assessing measurement equivalence in the framework of the common factor linear models (CFL) is known as factorial invariance. This methodology is used to evaluate the equivalence among the parameters of a measurement model among different groups. However, when dichotomous, Likert, or ordered responses are used, one of the assumptions of the CFL is…
Descriptors: Measurement, Models, Data, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Woods, Carol M.; Grimm, Kevin J. – Applied Psychological Measurement, 2011
In extant literature, multiple indicator multiple cause (MIMIC) models have been presented for identifying items that display uniform differential item functioning (DIF) only, not nonuniform DIF. This article addresses, for apparently the first time, the use of MIMIC models for testing both uniform and nonuniform DIF with categorical indicators. A…
Descriptors: Test Bias, Testing, Interaction, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Shih, Ching-Lin; Wang, Wen-Chung – Applied Psychological Measurement, 2009
The multiple indicators, multiple causes (MIMIC) method with a pure short anchor was proposed to detect differential item functioning (DIF). A simulation study showed that the MIMIC method with an anchor of 1, 2, 4, or 10 DIF-free items yielded a well-controlled Type I error rate even when such tests contained as many as 40% DIF items. In general,…
Descriptors: Test Bias, Simulation, Methods, Factor Analysis
Peer reviewed Peer reviewed
Direct linkDirect link
Carvajal, Jorge; Skorupski, William P. – Educational and Psychological Measurement, 2010
This study is an evaluation of the behavior of the Liu-Agresti estimator of the cumulative common odds ratio when identifying differential item functioning (DIF) with polytomously scored test items using small samples. The Liu-Agresti estimator has been proposed by Penfield and Algina as a promising approach for the study of polytomous DIF but no…
Descriptors: Test Bias, Sample Size, Test Items, Computation
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Young-Sun; Cohen, Allan; Toro, Maritsa – Asia Pacific Education Review, 2009
In this study, the effectiveness of detection of differential item functioning (DIF) and testlet DIF using SIBTEST and Poly-SIBTEST were examined in tests composed of testlets. An example using data from a reading comprehension test showed that results from SIBTEST and Poly-SIBTEST were not completely consistent in the detection of DIF and testlet…
Descriptors: Test Bias, Reading Comprehension, Simulation, Reading Tests
Peer reviewed Peer reviewed
Direct linkDirect link
Guler, Nese; Penfield, Randall D. – Journal of Educational Measurement, 2009
In this study, we investigate the logistic regression (LR), Mantel-Haenszel (MH), and Breslow-Day (BD) procedures for the simultaneous detection of both uniform and nonuniform differential item functioning (DIF). A simulation study was used to assess and compare the Type I error rate and power of a combined decision rule (CDR), which assesses DIF…
Descriptors: Test Bias, Simulation, Test Items, Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Lopez Rivas, Gabriel E.; Stark, Stephen; Chernyshenko, Oleksandr S. – Applied Psychological Measurement, 2009
The purpose of this simulation study is to investigate the effects of anchor subtest composition on the accuracy of item response theory (IRT) likelihood ratio (LR) differential item functioning (DIF) detection (Thissen, Steinberg, & Wainer, 1988). Here, the IRT LR test was implemented with a free baseline approach wherein a baseline model was…
Descriptors: Simulation, Item Response Theory, Test Bias, Test Items
Baker, Eva L. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2010
This report provides an overview of what was known about alternative assessment at the time that the article was written in 1991. Topics include beliefs about assessment reform, overview of alternative assessment including research knowledge, evidence of assessment impact, and critical features of alternative assessment. The author notes that in…
Descriptors: Alternative Assessment, Evaluation Methods, Evaluation Research, Performance Based Assessment
Peer reviewed Peer reviewed
Direct linkDirect link
Weitzman, R. A. – Educational and Psychological Measurement, 2009
Building on the Kelley and Gulliksen versions of classical test theory, this article shows that a logistic model having only a single item parameter can account for varying item discrimination, as well as difficulty, by using item-test correlations to adjust incorrect-correct (0-1) item responses prior to an initial model fit. The fit occurs…
Descriptors: Item Response Theory, Test Items, Difficulty Level, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Puhan, Gautam; Moses, Timothy P.; Yu, Lei; Dorans, Neil J. – Journal of Educational Measurement, 2009
This study examined the extent to which log-linear smoothing could improve the accuracy of differential item functioning (DIF) estimates in small samples of examinees. Examinee responses from a certification test were analyzed using White examinees in the reference group and African American examinees in the focal group. Using a simulation…
Descriptors: Test Items, Reference Groups, Testing Programs, Raw Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Finch, W. Holmes; French, Brian F. – Educational and Psychological Measurement, 2008
A number of statistical methods exist for the detection of differential item functioning (DIF). The performance of DIF methods has been widely studied and generally found to be effective in the detection of both uniform and nonuniform DIF. Anecdotal reports suggest that these techniques may too often incorrectly detect the presence of one type of…
Descriptors: Test Bias, Simulation, Statistical Analysis, Probability
Peer reviewed Peer reviewed
Direct linkDirect link
Woods, Carol M. – Applied Psychological Measurement, 2009
Differential item functioning (DIF) occurs when items on a test or questionnaire have different measurement properties for one group of people versus another, irrespective of group-mean differences on the construct. Methods for testing DIF require matching members of different groups on an estimate of the construct. Preferably, the estimate is…
Descriptors: Test Results, Testing, Item Response Theory, Test Bias
Peer reviewed Peer reviewed
Direct linkDirect link
Woods, Carol M. – Multivariate Behavioral Research, 2009
Differential item functioning (DIF) occurs when an item on a test or questionnaire has different measurement properties for 1 group of people versus another, irrespective of mean differences on the construct. This study focuses on the use of multiple-indicator multiple-cause (MIMIC) structural equation models for DIF testing, parameterized as item…
Descriptors: Test Bias, Structural Equation Models, Item Response Theory, Testing
Peer reviewed Peer reviewed
Direct linkDirect link
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Previous Page | Next Page ยป
Pages: 1  |  2  |  3