Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 15 |
Descriptor
Comparative Analysis | 16 |
Test Bias | 16 |
Foreign Countries | 8 |
Item Response Theory | 8 |
Scores | 5 |
Test Items | 5 |
Effect Size | 4 |
Mathematics Tests | 4 |
Computation | 3 |
Educational Assessment | 3 |
English | 3 |
More ▼ |
Source
International Journal of… | 16 |
Author
Ercikan, Kadriye | 4 |
Oliveri, Maria Elena | 3 |
Zumbo, Bruno D. | 3 |
Sandilands, Debra | 2 |
Alexeev, Natalia | 1 |
Beland, Sebastien | 1 |
Berberoglu, Giray | 1 |
Breland, Hunter | 1 |
Brown, Richard S. | 1 |
Chen, Michelle Y. | 1 |
Choi, Youn-Jeng | 1 |
More ▼ |
Publication Type
Journal Articles | 16 |
Reports - Research | 10 |
Reports - Evaluative | 5 |
Information Analyses | 1 |
Reports - Descriptive | 1 |
Education Level
Elementary Education | 3 |
Grade 4 | 3 |
Intermediate Grades | 2 |
Grade 8 | 1 |
High Schools | 1 |
Higher Education | 1 |
Secondary Education | 1 |
Audience
Location
United States | 3 |
Australia | 2 |
Canada | 2 |
Hong Kong | 2 |
Qatar | 2 |
Austria | 1 |
Colombia | 1 |
El Salvador | 1 |
Kenya | 1 |
Kuwait | 1 |
Singapore | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Program for International… | 4 |
Progress in International… | 3 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Ercikan, Kadriye; Chen, Michelle Y.; Lyons-Thomas, Juliette; Goodrich, Shawna; Sandilands, Debra; Roth, Wolff-Michael; Simon, Marielle – International Journal of Testing, 2015
The purpose of this research is to examine the comparability of mathematics and science scores for students from English language backgrounds (ELB) and non-English language backgrounds (NELB). We examine the relationship between English reading proficiency and performance on mathematics and science assessments in Australia, Canada, the United…
Descriptors: Scores, Mathematics Tests, Science Tests, Native Speakers
Oshima, T. C.; Wright, Keith; White, Nick – International Journal of Testing, 2015
Raju, van der Linden, and Fleer (1995) introduced a framework for differential functioning of items and tests (DFIT) for unidimensional dichotomous models. Since then, DFIT has been shown to be a quite versatile framework as it can handle polytomous as well as multidimensional models both at the item and test levels. However, DFIT is still limited…
Descriptors: Test Bias, Item Response Theory, Test Items, Simulation
Oliveri, María Elena; Ercikan, Kadriye; Zumbo, Bruno D.; Lawless, René – International Journal of Testing, 2014
In this study, we contrast results from two differential item functioning (DIF) approaches (manifest and latent class) by the number of items and sources of items identified as DIF using data from an international reading assessment. The latter approach yielded three latent classes, presenting evidence of heterogeneity in examinee response…
Descriptors: Test Bias, Comparative Analysis, Reading Tests, Effect Size
Choi, Youn-Jeng; Alexeev, Natalia; Cohen, Allan S. – International Journal of Testing, 2015
The purpose of this study was to explore what may be contributing to differences in performance in mathematics on the Trends in International Mathematics and Science Study 2007. This was done by using a mixture item response theory modeling approach to first detect latent classes in the data and then to examine differences in performance on items…
Descriptors: Test Bias, Mathematics Achievement, Mathematics Tests, Item Response Theory
Oliveri, Maria Elena; von Davier, Matthias – International Journal of Testing, 2014
In this article, we investigate the creation of comparable score scales across countries in international assessments. We examine potential improvements to current score scale calibration procedures used in international large-scale assessments. Our approach seeks to improve fairness in scoring international large-scale assessments, which often…
Descriptors: Test Bias, Scores, International Programs, Educational Assessment
Rios, Joseph A.; Sireci, Stephen G. – International Journal of Testing, 2014
The International Test Commission's "Guidelines for Translating and Adapting Tests" (2010) provide important guidance on developing and evaluating tests for use across languages. These guidelines are widely applauded, but the degree to which they are followed in practice is unknown. The objective of this study was to perform a…
Descriptors: Guidelines, Translation, Adaptive Testing, Second Languages
Mucherah, Winnie; Finch, W. Holmes; Keaikitse, Setlhomo – International Journal of Testing, 2012
Understanding adolescent self-concept is of great concern for educators, mental health professionals, and parents, as research consistently demonstrates that low self-concept is related to a number of problem behaviors and poor outcomes. Thus, accurate measurements of self-concept are key, and the validity of such measurements, including the…
Descriptors: Test Bias, Mental Health Workers, Validity, Self Concept Measures
Wiberg, Marie – International Journal of Testing, 2009
The aim of this study was to examine log linear modelling (LLM) compared with logistic regression (LR) and Mantel-Haenszel (MH) test for detecting Differential Item Functioning (DIF) in a mastery test. The three methods were chosen because they have similar components. The results showed fairly high matching percentages together with high…
Descriptors: Test Bias, Mastery Tests, Comparative Analysis, Regression (Statistics)
Magis, David; Raiche, Gilles; Beland, Sebastien; Gerard, Paul – International Journal of Testing, 2011
We present an extension of the logistic regression procedure to identify dichotomous differential item functioning (DIF) in the presence of more than two groups of respondents. Starting from the usual framework of a single focal group, we propose a general approach to estimate the item response functions in each group and to test for the presence…
Descriptors: Language Skills, Identification, Foreign Countries, Evaluation Methods
Oliveri, Maria Elena; Olson, Brent F.; Ercikan, Kadriye; Zumbo, Bruno D. – International Journal of Testing, 2012
In this study, the Canadian English and French versions of the Problem-Solving Measure of the Programme for International Student Assessment 2003 were examined to investigate their degree of measurement comparability at the item- and test-levels. Three methods of differential item functioning (DIF) were compared: parametric and nonparametric item…
Descriptors: Foreign Students, Test Bias, Speech Communication, Effect Size
Sandilands, Debra; Oliveri, Maria Elena; Zumbo, Bruno D.; Ercikan, Kadriye – International Journal of Testing, 2013
International large-scale assessments of achievement often have a large degree of differential item functioning (DIF) between countries, which can threaten score equivalence and reduce the validity of inferences based on comparisons of group performances. It is important to understand potential sources of DIF to improve the validity of future…
Descriptors: Validity, Measures (Individuals), International Studies, Foreign Countries
Wyse, Adam E.; Mapuranga, Raymond – International Journal of Testing, 2009
Differential item functioning (DIF) analysis is a statistical technique used for ensuring the equity and fairness of educational assessments. This study formulates a new DIF analysis method using the information similarity index (ISI). ISI compares item information functions when data fits the Rasch model. Through simulations and an international…
Descriptors: Test Bias, Evaluation Methods, Test Items, Educational Assessment
Yildirim, Huseyin Husnu; Berberoglu, Giray – International Journal of Testing, 2009
Comparisons of human characteristics across different language groups and cultures become more important in today's educational assessment practices as evidenced by the increasing interest in international comparative studies. Within this context, the fairness of the results across different language and cultural groups draws the attention of…
Descriptors: Test Bias, Cross Cultural Studies, Comparative Analysis, Factor Analysis
Lamprianou, Iasonas – International Journal of Testing, 2008
This study investigates the effect of reporting the unadjusted raw scores in a high-stakes language exam when raters differ significantly in severity and self-selected questions differ significantly in difficulty. More sophisticated models, introducing meaningful facets and parameters, are successively used to investigate the characteristics of…
Descriptors: High Stakes Tests, Raw Scores, Item Response Theory, Language Tests
Brown, Richard S.; Villarreal, Julio C. – International Journal of Testing, 2007
There has been considerable research regarding the extent to which psychometric sound assessments sometimes yield individual score estimates that are inconsistent with the response patterns of the individual. It has been suggested that individual response patterns may differ from expectations for a number of reasons, including subject motivation,…
Descriptors: Psychometrics, Test Bias, Testing, Simulation
Previous Page | Next Page »
Pages: 1 | 2