Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 9 |
Since 2006 (last 20 years) | 13 |
Descriptor
Source
Author
Carnoy, Martin | 1 |
Chengyu Cui | 1 |
Chun Wang | 1 |
Dirlik, Ezgi Mor | 1 |
Gongjun Xu | 1 |
Grund, Simon | 1 |
Gustafsson, Jan-Eric | 1 |
Haag, Nicole | 1 |
Hansson, Ase | 1 |
Kang, Yoonjeong | 1 |
Lüdtke, Oliver | 1 |
More ▼ |
Publication Type
Journal Articles | 10 |
Reports - Research | 7 |
Reports - Evaluative | 3 |
Reports - Descriptive | 2 |
Dissertations/Theses -… | 1 |
Opinion Papers | 1 |
Education Level
Elementary Secondary Education | 11 |
Secondary Education | 3 |
Elementary Education | 1 |
Grade 8 | 1 |
Audience
Location
China (Shanghai) | 1 |
Kuwait | 1 |
Singapore | 1 |
Sweden | 1 |
Turkey | 1 |
United States | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Trends in International… | 13 |
Program for International… | 3 |
National Assessment of… | 2 |
Big Five Inventory | 1 |
Early Childhood Longitudinal… | 1 |
Schools and Staffing Survey… | 1 |
What Works Clearinghouse Rating
Chengyu Cui; Chun Wang; Gongjun Xu – Grantee Submission, 2024
Multidimensional item response theory (MIRT) models have generated increasing interest in the psychometrics literature. Efficient approaches for estimating MIRT models with dichotomous responses have been developed, but constructing an equally efficient and robust algorithm for polytomous models has received limited attention. To address this gap,…
Descriptors: Item Response Theory, Accuracy, Simulation, Psychometrics
Yanan Feng – ProQuest LLC, 2021
This dissertation aims to investigate the effect size measures of differential item functioning (DIF) detection in the context of cognitive diagnostic models (CDMs). A variety of DIF detection techniques have been developed in the context of CDMs. However, most of the DIF detection procedures focus on the null hypothesis significance test. Few…
Descriptors: Effect Size, Item Response Theory, Cognitive Measurement, Models
Grund, Simon; Lüdtke, Oliver; Robitzsch, Alexander – Journal of Educational and Behavioral Statistics, 2021
Large-scale assessments (LSAs) use Mislevy's "plausible value" (PV) approach to relate student proficiency to noncognitive variables administered in a background questionnaire. This method requires background variables to be completely observed, a requirement that is seldom fulfilled. In this article, we evaluate and compare the…
Descriptors: Data Analysis, Error of Measurement, Research Problems, Statistical Inference
Wang, Jianjun; Ma, Xin – Athens Journal of Education, 2019
This rejoinder keeps the original focus on statistical computing pertaining to the correlation of student achievement between mathematics and science from the Trend in Mathematics and Science Study (TIMSS). Albeit the availability of student performance data in TIMSS and the emphasis of the inter-subject connection in the Next Generation Science…
Descriptors: Scores, Correlation, Achievement Tests, Elementary Secondary Education
Dirlik, Ezgi Mor – International Journal of Progressive Education, 2019
Item response theory (IRT) has so many advantages than its precedent Classical Test Theory (CTT) such as non-changing item parameters, ability parameter estimations free from the items. However, in order to get these advantages, some assumptions should be met and they are; unidimensionality, normality and local independence. However, it is not…
Descriptors: Comparative Analysis, Nonparametric Statistics, Item Response Theory, Models
Martin, Michael O.; Mullis, Ina V. S. – Journal of Educational and Behavioral Statistics, 2019
International large-scale assessments of student achievement such as International Association for the Evaluation of Educational Achievement's Trends in International Mathematics and Science Study (TIMSS) and Progress in International Reading Literacy Study and Organization for Economic Cooperation and Development's Program for International…
Descriptors: Achievement Tests, International Assessment, Mathematics Tests, Science Achievement
Yalcin, Seher – Eurasian Journal of Educational Research, 2018
Purpose: Studies in the literature have generally demonstrated that the causes of differential item functioning (DIF) are complex and not directly related to defined groups. The purpose of this study is to determine the DIF according to the mixture item response theory (MixIRT) model, based on the latent group approach, as well as the…
Descriptors: Item Response Theory, Test Items, Test Bias, Error of Measurement
Stapleton, Laura M.; Kang, Yoonjeong – Sociological Methods & Research, 2018
This research empirically evaluates data sets from the National Center for Education Statistics (NCES) for design effects of ignoring the sampling design in weighted two-level analyses. Currently, researchers may ignore the sampling design beyond the levels that they model which might result in incorrect inferences regarding hypotheses due to…
Descriptors: Probability, Hierarchical Linear Modeling, Sampling, Inferences
Pokropek, Artur – Sociological Methods & Research, 2015
This article combines statistical and applied research perspective showing problems that might arise when measurement error in multilevel compositional effects analysis is ignored. This article focuses on data where independent variables are constructed measures. Simulation studies are conducted evaluating methods that could overcome the…
Descriptors: Error of Measurement, Hierarchical Linear Modeling, Simulation, Evaluation Methods
Sachse, Karoline A.; Roppelt, Alexander; Haag, Nicole – Journal of Educational Measurement, 2016
Trend estimation in international comparative large-scale assessments relies on measurement invariance between countries. However, cross-national differential item functioning (DIF) has been repeatedly documented. We ran a simulation study using national item parameters, which required trends to be computed separately for each country, to compare…
Descriptors: Comparative Analysis, Measurement, Test Bias, Simulation
Svetina, Dubravka; Rutkowski, Leslie – Large-scale Assessments in Education, 2014
Background: When studying student performance across different countries or cultures, an important aspect for comparisons is that of score comparability. In other words, it is imperative that the latent variable (i.e., construct of interest) is understood and measured equivalently across all participating groups or countries, if our inferences…
Descriptors: Test Items, Item Response Theory, Item Analysis, Regression (Statistics)
Hansson, Ase; Gustafsson, Jan-Eric – Scandinavian Journal of Educational Research, 2013
Socioeconomic status (SES) is often used as control variable when relations between academic outcomes and students' migrational background are investigated. When measuring SES, indicators used must have the same meaning across groups. This study aims to examine the measurement invariance of SES, using data from TIMSS, 2003. The study shows that a…
Descriptors: Socioeconomic Status, Error of Measurement, Immigrants, Outcomes of Education
Carnoy, Martin – National Education Policy Center, 2015
Stanford education professor Martin Carnoy examines four main critiques of how international test results are used in policymaking. Of particular interest are critiques of the policy analyses published by the Program for International Student Assessment (PISA). Using average PISA scores as a comparative measure of student achievement is misleading…
Descriptors: Criticism, Reputation, Test Validity, Error of Measurement