Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 2 |
Since 2006 (last 20 years) | 6 |
Descriptor
Test Items | 25 |
Item Response Theory | 17 |
Item Bias | 12 |
Comparative Analysis | 6 |
Sample Size | 6 |
Simulation | 6 |
Equations (Mathematics) | 5 |
Estimation (Mathematics) | 5 |
Identification | 5 |
Mathematical Models | 5 |
Mathematics Tests | 5 |
More ▼ |
Source
Applied Psychological… | 7 |
Journal of Educational… | 3 |
Applied Measurement in… | 2 |
International Journal of… | 2 |
Educational and Psychological… | 1 |
Grantee Submission | 1 |
Learning Disabilities… | 1 |
Turkish Journal of Education | 1 |
Author
Cohen, Allan S. | 25 |
Kim, Seock-Ho | 13 |
Wollack, James A. | 2 |
Alagoz, Cigdem | 1 |
Bottge, Brian | 1 |
Bottge, Brian A. | 1 |
Cho, Sun-Joo | 1 |
Choi, Hye-Jeong | 1 |
Deng, Meng | 1 |
DiStefano, Christine A. | 1 |
Gregg, Noel | 1 |
More ▼ |
Publication Type
Journal Articles | 17 |
Reports - Evaluative | 12 |
Reports - Research | 12 |
Speeches/Meeting Papers | 12 |
Information Analyses | 1 |
Reports - Descriptive | 1 |
Education Level
Middle Schools | 2 |
Junior High Schools | 1 |
Secondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Peabody Picture Vocabulary… | 1 |
Sequential Tests of… | 1 |
What Works Clearinghouse Rating
Sen, Sedat; Terzi, Ragip; Yildirim, Ibrahim; Cohen, Allan S. – Turkish Journal of Education, 2018
The purpose of this study was to examine the effect of equated and non-equated data on value-added assessment analyses. Several models have been proposed in the literature to apply the value-added assessment approach. This study compared two different value-added models: the unadjusted hierarchical linear model and the generalized persistence…
Descriptors: Equated Scores, Value Added Models, Hierarchical Linear Modeling, Persistence
An Application of a Random Mixture Nominal Item Response Model for Investigating Instruction Effects
Choi, Hye-Jeong; Cohen, Allan S.; Bottge, Brian A. – Grantee Submission, 2016
The purpose of this study was to apply a random item mixture nominal item response model (RIM-MixNRM) for investigating instruction effects. The host study design was a pre-test-and-post-test, school-based cluster randomized trial. A RIM-MixNRM was used to identify students' error patterns in mathematics at the pre-test and the post-test.…
Descriptors: Item Response Theory, Instructional Effectiveness, Test Items, Models
Wells, Craig S.; Cohen, Allan S.; Patton, Jeffrey – International Journal of Testing, 2009
A primary concern with testing differential item functioning (DIF) using a traditional point-null hypothesis is that a statistically significant result does not imply that the magnitude of DIF is of practical interest. Similarly, for a given sample size, a non-significant result does not allow the researcher to conclude the item is free of DIF. To…
Descriptors: Test Bias, Test Items, Statistical Analysis, Hypothesis Testing
Cho, Sun-Joo; Cohen, Allan S.; Kim, Seock-Ho; Bottge, Brian – Applied Psychological Measurement, 2010
A latent transition analysis (LTA) model was described with a mixture Rasch model (MRM) as the measurement model. Unlike the LTA, which was developed with a latent class measurement model, the LTA-MRM permits within-class variability on the latent variable, making it more useful for measuring treatment effects within latent classes. A simulation…
Descriptors: Item Response Theory, Measurement, Models, Statistical Analysis
Webb, Mi-young Lee; Cohen, Allan S.; Schwanenflugel, Paula J. – Educational and Psychological Measurement, 2008
This study investigated the use of latent class analysis for the detection of differences in item functioning on the Peabody Picture Vocabulary Test-Third Edition (PPVT-III). A two-class solution for a latent class model appeared to be defined in part by ability because Class 1 was lower in ability than Class 2 on both the PPVT-III and the…
Descriptors: Item Response Theory, Test Items, Test Format, Cognitive Ability
Kim, Seock-Ho; Cohen, Allan S.; Alagoz, Cigdem; Kim, Sukwoo – Journal of Educational Measurement, 2007
Data from a large-scale performance assessment (N = 105,731) were analyzed with five differential item functioning (DIF) detection methods for polytomous items to examine the congruence among the DIF detection methods. Two different versions of the item response theory (IRT) model-based likelihood ratio test, the logistic regression likelihood…
Descriptors: Performance Based Assessment, Performance Tests, Item Response Theory, Test Bias
Kim, Seock-Ho; Cohen, Allan S.; DiStefano, Christine A.; Kim, Sooyeon – 1998
Type I error rates of the likelihood ratio test for the detection of differential item functioning (DIF) in the partial credit model were investigated using simulated data. The partial credit model with four ordered performance levels was used to generate data sets of a 30-item test for samples of 300 and 1,000 simulated examinees. Three different…
Descriptors: Item Bias, Simulation, Test Items

Wollack, James A.; Cohen, Allan S. – Applied Psychological Measurement, 1998
Investigated empirical Type I error rates and the power of omega (index of answer copying developed by J. Wollack, 1997) when item and trait (theta) parameters were unknown and estimated from datasets of 100 and 500 examinees. Type I error was unaffected by estimating item parameters, with power slightly lower for the smaller sample. (SLD)
Descriptors: Cheating, Estimation (Mathematics), Plagiarism, Sample Size

Kim, Seock-Ho; Cohen, Allan S. – Applied Psychological Measurement, 1998
Compared three methods for developing a common metric under item response theory through simulation. For smaller numbers of common items, linking using the characteristic curve method yielded smaller root mean square differences for both item discrimination and difficulty parameters. For larger numbers of common items, the three methods were…
Descriptors: Comparative Analysis, Difficulty Level, Item Response Theory, Simulation
Cohen, Allan S.; Kim, Seock-Ho; Wollack, James A. – 1998
This paper provides a review of procedures for detection of differential item functioning (DIF) for item response theory (IRT) and observed score methods for the graded response model. In addition, data from a test anxiety scale were analyzed to examine the congruence among these procedures. Data from Nasser, Takahashi, and Benson (1997) were…
Descriptors: Identification, Item Bias, Item Response Theory, Scores
Kang, Taehoon; Cohen, Allan S. – 2003
A number of methods exist for detection of differential item functioning (DIF), but these methods tell us little about the causes of DIF. DIF is typically defined based on a relationship with some manifest group characteristic, such as gender or ethnicity, which is only weakly associated with DIF. What is lacking is a method that will lead to…
Descriptors: Ethnicity, Item Bias, Item Response Theory, Racial Differences
Li, Yanmei; Cohen, Allan S.; Ibarra, Robert A. – International Journal of Testing, 2004
Most research on differential item functioning (DIF) focuses on methods for detection rather than on understanding why DIF might occur. This study was designed to investigate whether two alternative approaches to parsing items based on structural characteristics related to particular cognitive strategies could be used to help explain gender DIF.…
Descriptors: Test Items, Cognitive Structures, Gender Differences, Mathematics Tests

Kim, Seock-Ho; Cohen, Allan S. – Applied Psychological Measurement, 1998
Investigated Type I error rates of the likelihood-ratio test for the detection of differential item functioning (DIF) using Monte Carlo simulations under the graded-response model. Type I error rates were within theoretically expected values for all six combinations of sample sizes and ability-matching conditions at each of the nominal alpha…
Descriptors: Ability, Item Bias, Item Response Theory, Monte Carlo Methods

Cohen, Allan S.; And Others – Applied Psychological Measurement, 1996
Type I error rates for the likelihood ratio test for detecting differential item functioning (DIF) were investigated using Monte Carlo simulations. Type I error rates for the two-parameter model were within theoretically expected values at each alpha level, but those for the three-parameter model were not. (SLD)
Descriptors: Identification, Item Bias, Item Response Theory, Maximum Likelihood Statistics

Cohen, Allan S.; And Others – Applied Psychological Measurement, 1993
Three measures of differential item functioning for the dichotomous response model are extended to include Samejima's graded response model. Two are based on area differences between item true score functions, and one is a chi-square statistic for comparing differences in item parameters. (SLD)
Descriptors: Chi Square, Comparative Analysis, Identification, Item Bias
Previous Page | Next Page ยป
Pages: 1 | 2