Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 16 |
Since 2006 (last 20 years) | 31 |
Descriptor
Difficulty Level | 35 |
Foreign Countries | 35 |
Test Bias | 35 |
Test Items | 30 |
Mathematics Tests | 14 |
Achievement Tests | 12 |
Item Response Theory | 12 |
International Assessment | 10 |
Comparative Analysis | 8 |
Scores | 7 |
Secondary School Students | 7 |
More ▼ |
Source
Author
Baird, Jo-Anne | 2 |
Liaw, Yuan-Ling | 2 |
Rutkowski, David | 2 |
Rutkowski, Leslie | 2 |
Abedalaziz, Nabeel | 1 |
Abedlaziz, Nabeel | 1 |
Alahmadi, Ahlam | 1 |
Albano, Anthony D. | 1 |
Allalouf, Avi | 1 |
Babiar, Tasha Calvert | 1 |
Baghaei, Purya | 1 |
More ▼ |
Publication Type
Journal Articles | 33 |
Reports - Research | 29 |
Reports - Evaluative | 4 |
Books | 1 |
Collected Works - General | 1 |
Information Analyses | 1 |
Reports - Descriptive | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Audience
Researchers | 1 |
Location
Turkey | 5 |
United States | 5 |
Germany | 4 |
Belgium | 3 |
Finland | 2 |
Hong Kong | 2 |
Algeria | 1 |
Austria | 1 |
Canada | 1 |
Colombia | 1 |
Denmark | 1 |
More ▼ |
Laws, Policies, & Programs
Assessments and Surveys
Program for International… | 7 |
Trends in International… | 4 |
Progress in International… | 3 |
Cattell Culture Fair… | 1 |
Flesch Kincaid Grade Level… | 1 |
Flesch Reading Ease Formula | 1 |
Test of English as a Foreign… | 1 |
What Works Clearinghouse Rating
Qi Huang; Daniel M. Bolt; Weicong Lyu – Large-scale Assessments in Education, 2024
Large scale international assessments depend on invariance of measurement across countries. An important consideration when observing cross-national differential item functioning (DIF) is whether the DIF actually reflects a source of bias, or might instead be a methodological artifact reflecting item response theory (IRT) model misspecification.…
Descriptors: Test Items, Item Response Theory, Test Bias, Test Validity
Kuang, Huan; Sahin, Fusun – Large-scale Assessments in Education, 2023
Background: Examinees may not make enough effort when responding to test items if the assessment has no consequence for them. These disengaged responses can be problematic in low-stakes, large-scale assessments because they can bias item parameter estimates. However, the amount of bias, and whether this bias is similar across administrations, is…
Descriptors: Test Items, Comparative Analysis, Mathematics Tests, Reaction Time
Bundsgaard, Jeppe – Large-scale Assessments in Education, 2019
International large-scale assessments like international computer and information literacy study (ICILS) (Fraillon et al. in International Association for the Evaluation of Educational Achievement (IEA), 2015) provide important empirically-based knowledge through the proficiency scales, of what characterizes tasks at different difficulty levels,…
Descriptors: Test Bias, International Assessment, Test Items, Difficulty Level
Rutkowski, Leslie; Rutkowski, David; Liaw, Yuan-Ling – Assessment in Education: Principles, Policy & Practice, 2019
Modern international studies of educational achievement have grown in terms of participating educational systems. Accompanying this development is an increase in heterogeneity, as more and different kinds of educational systems take part. This growth has been particularly pronounced among low-performing, less economically developed systems.…
Descriptors: International Assessment, Secondary School Students, Foreign Countries, Achievement Tests
Retnawati, Heri – International Journal of Assessment Tools in Education, 2018
The study was to identify the load, the type and the significance of differential item functioning (DIF) in constructed response item using the partial credit model (PCM). The data in the study were the students' instruments and the students' responses toward the PISA-like test items that had been completed by 386 ninth grade students and 460…
Descriptors: Test Bias, Test Items, Responses, Grade 9
Rutkowski, David; Rutkowski, Leslie; Liaw, Yuan-Ling – Educational Measurement: Issues and Practice, 2018
Participation in international large-scale assessments has grown over time with the largest, the Programme for International Student Assessment (PISA), including more than 70 education systems that are economically and educationally diverse. To help accommodate for large achievement differences among participants, in 2009 PISA offered…
Descriptors: Educational Assessment, Foreign Countries, Achievement Tests, Secondary School Students
Balta, Ebru; Omur Sunbul, Secil – Eurasian Journal of Educational Research, 2017
Purpose: Position effects may influence examinees' test performances in several ways and trigger other psychometric issues, such as Differential Item Functioning (DIF) .This study aims to supply test forms in which items in the test are ordered differently, depending on their difficulty level (from easy to difficult or difficult to easy), to…
Descriptors: Test Items, Sequential Approach, Difficulty Level, Mathematics Tests
Ozdemir, Burhanettin – International Journal of Progressive Education, 2017
The purpose of this study is to equate Trends in International Mathematics and Science Study (TIMSS) mathematics subtest scores obtained from TIMSS 2011 to scores obtained from TIMSS 2007 form with different nonlinear observed score equating methods under Non-Equivalent Anchor Test (NEAT) design where common items are used to link two or more test…
Descriptors: Achievement Tests, Elementary Secondary Education, Foreign Countries, International Assessment
Çokluk, Ömay; Gül, Emrah; Dogan-Gül, Çilem – Educational Sciences: Theory and Practice, 2016
The study aims to examine whether differential item function is displayed in three different test forms that have item orders of random and sequential versions (easy-to-hard and hard-to-easy), based on Classical Test Theory (CTT) and Item Response Theory (IRT) methods and bearing item difficulty levels in mind. In the correlational research, the…
Descriptors: Test Bias, Test Items, Difficulty Level, Test Theory
Baghaei, Purya; Ravand, Hamdollah – SAGE Open, 2019
In many reading comprehension tests, different test formats are employed. Two commonly used test formats to measure reading comprehension are sustained passages followed by some questions and cloze items. Individual differences in handling test format peculiarities could constitute a source of score variance. In this study, a bifactor Rasch model…
Descriptors: Cloze Procedure, Test Bias, Individual Differences, Difficulty Level
Hopfenbeck, Therese N.; Lenkeit, Jenny; El Masri, Yasmine; Cantrell, Kate; Ryan, Jeanne; Baird, Jo-Anne – Scandinavian Journal of Educational Research, 2018
International large-scale assessments are on the rise, with the Programme for International Student Assessment (PISA) seen by many as having strategic prominence in education policy debates. The present article reviews PISA-related English-language peer-reviewed articles from the programme's first cycle in 2000 to its most current in 2015. Five…
Descriptors: Foreign Countries, Achievement Tests, International Assessment, Secondary School Students
Cheng, Maurice M. W.; Oon, Pey-Tee – International Journal of Science Education, 2016
This paper reports the results of a survey of 3006 Year 10-12 students on their understandings of metallic bonding. The instrument was developed based on Chi's ontological categories of scientific concepts and students' understanding of metallic bonding as reported in the literature. The instrument has two parts. Part one probed into students'…
Descriptors: Chemistry, Item Response Theory, Science Instruction, Foreign Countries
Levi-Keren, Michal – Cogent Education, 2016
This study explains mathematical difficulties of students who immigrated from the Former Soviet Union (FSU) vis-à-vis Israeli students, by identifying the existing bias factors in achievement tests. These factors are irrelevant to the mathematical knowledge being measured, and therefore threaten the test results. The bias factors were identified…
Descriptors: Mathematics Achievement, Mathematics Tests, Immigrants, Interviews
Finch, W. Holmes; Hernández Finch, Maria E.; French, Brian F. – International Journal of Testing, 2016
Differential item functioning (DIF) assessment is key in score validation. When DIF is present scores may not accurately reflect the construct of interest for some groups of examinees, leading to incorrect conclusions from the scores. Given rising immigration, and the increased reliance of educational policymakers on cross-national assessments…
Descriptors: Test Bias, Scores, Native Language, Language Usage
Abedalaziz, Nabeel; Leng, Chin Hai; Alahmadi, Ahlam – Malaysian Online Journal of Educational Sciences, 2014
The purpose of the study was to examine gender differences in performance on multiple-choice mathematical ability test, administered within the context of high school graduation test that was designed to match eleventh grade curriculum. The transformed item difficulty (TID) was used to detect a gender related DIF. A random sample of 1400 eleventh…
Descriptors: Test Bias, Test Items, Difficulty Level, Gender Differences