Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 6 |
Descriptor
Source
Educational and Psychological… | 2 |
International Journal of… | 2 |
Applied Measurement in… | 1 |
Psychological Methods | 1 |
Studies in Educational… | 1 |
Author
Sachse, Karoline A. | 2 |
Cai, Li | 1 |
Fernández, Daniel | 1 |
Gu, Peter Yongqi | 1 |
Haag, Nicole | 1 |
Hansen, Mark | 1 |
Harvey, Samuel | 1 |
Liu, Ivy | 1 |
Mahler, Nicole | 1 |
Mapuranga, Raymond | 1 |
Pohl, Steffi | 1 |
More ▼ |
Publication Type
Journal Articles | 7 |
Reports - Research | 4 |
Reports - Evaluative | 3 |
Education Level
Secondary Education | 3 |
Elementary Secondary Education | 1 |
Audience
Location
Australia | 1 |
Ireland | 1 |
New Zealand | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Program for International… | 7 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Liu, Ivy; Suesse, Thomas; Harvey, Samuel; Gu, Peter Yongqi; Fernández, Daniel; Randal, John – Educational and Psychological Measurement, 2023
The Mantel-Haenszel estimator is one of the most popular techniques for measuring differential item functioning (DIF). A generalization of this estimator is applied to the context of DIF to compare items by taking the covariance of odds ratio estimators between dependent items into account. Unlike the Item Response Theory, the method does not rely…
Descriptors: Test Bias, Computation, Statistical Analysis, Achievement Tests
Sachse, Karoline A.; Mahler, Nicole; Pohl, Steffi – Educational and Psychological Measurement, 2019
Mechanisms causing item nonresponses in large-scale assessments are often said to be nonignorable. Parameter estimates can be biased if nonignorable missing data mechanisms are not adequately modeled. In trend analyses, it is plausible for the missing data mechanism and the percentage of missing values to change over time. In this article, we…
Descriptors: International Assessment, Response Style (Tests), Achievement Tests, Foreign Countries
Sachse, Karoline A.; Haag, Nicole – Applied Measurement in Education, 2017
Standard errors computed according to the operational practices of international large-scale assessment studies such as the Programme for International Student Assessment's (PISA) or the Trends in International Mathematics and Science Study (TIMSS) may be biased when cross-national differential item functioning (DIF) and item parameter drift are…
Descriptors: Error of Measurement, Test Bias, International Assessment, Computation
Rutkowski, Leslie; Rutkowski, David; Zhou, Yan – International Journal of Testing, 2016
Using an empirically-based simulation study, we show that typically used methods of choosing an item calibration sample have significant impacts on achievement bias and system rankings. We examine whether recent PISA accommodations, especially for lower performing participants, can mitigate some of this bias. Our findings indicate that standard…
Descriptors: Simulation, International Programs, Adolescents, Student Evaluation
Cai, Li; Yang, Ji Seung; Hansen, Mark – Psychological Methods, 2011
Full-information item bifactor analysis is an important statistical method in psychological and educational measurement. Current methods are limited to single-group analysis and inflexible in the types of item response models supported. We propose a flexible multiple-group item bifactor analysis framework that supports a variety of…
Descriptors: Item Analysis, Item Response Theory, Factor Analysis, Maximum Likelihood Statistics
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Wu, Margaret – Studies in Educational Evaluation, 2005
In large-scale assessment programs such as NAEP, TIMSS and PISA, students' achievement data sets provided for secondary analysts contain so-called "plausible values." Plausible values are multiple imputations of the unobservable latent achievement for each student. In this article it has been shown how plausible values are used to: (1)…
Descriptors: Error of Measurement, Computation, Educational Research, Educational Assessment