NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
What Works Clearinghouse Rating
Showing all 11 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Maritza Casas; Stephen G. Sireci – International Journal of Testing, 2025
In this study, we take a critical look at the degree to which the measurement of bullying and sense of belonging at school is invariant across groups of students defined by immigrant status. Our study focuses on the invariance of these constructs as measured on a recent PISA administration and includes a discussion of two statistical methods for…
Descriptors: Error of Measurement, Immigrants, Peer Groups, Bullying
Peer reviewed Peer reviewed
Direct linkDirect link
Wang, Ze – Large-scale Assessments in Education, 2022
In educational and psychological research, it is common to use latent factors to represent constructs and then to examine covariate effects on these latent factors. Using empirical data, this study applied three approaches to covariate effects on latent factors: the multiple-indicator multiple-cause (MIMIC) approach, multiple group confirmatory…
Descriptors: Comparative Analysis, Evaluation Methods, Grade 8, Mathematics Achievement
Peer reviewed Peer reviewed
Direct linkDirect link
Robert Meyer; Sara Hu; Michael Christian – Society for Research on Educational Effectiveness, 2022
This paper develops models to measure growth in student achievement with a focus on the possibility of differential growth in achievement for low and high-achieving students. We consider a gap-closing model that evaluates the degree to which students in a target group -- students in the bottom quartile of measured achievement -- perform better…
Descriptors: Academic Achievement, Achievement Gap, Models, Measurement Techniques
Gulsah Gurkan – ProQuest LLC, 2021
Secondary analyses of international large-scale assessments (ILSA) commonly characterize relationships between variables of interest using correlations. However, the accuracy of correlation estimates is impaired by artefacts such as measurement error and clustering. Despite advancements in methodology, conventional correlation estimates or…
Descriptors: Secondary School Students, Achievement Tests, International Assessment, Foreign Countries
Reardon, Sean F.; Ho, Andrew D.; Kalogrides, Demetra – Stanford Center for Education Policy Analysis, 2019
Linking score scales across different tests is considered speculative and fraught, even at the aggregate level (Feuer et al., 1999; Thissen, 2007). We introduce and illustrate validation methods for aggregate linkages, using the challenge of linking U.S. school district average test scores across states as a motivating example. We show that…
Descriptors: Test Validity, Evaluation Methods, School Districts, Scores
Peer reviewed Peer reviewed
Direct linkDirect link
Grund, Simon; Lüdtke, Oliver; Robitzsch, Alexander – Journal of Educational and Behavioral Statistics, 2021
Large-scale assessments (LSAs) use Mislevy's "plausible value" (PV) approach to relate student proficiency to noncognitive variables administered in a background questionnaire. This method requires background variables to be completely observed, a requirement that is seldom fulfilled. In this article, we evaluate and compare the…
Descriptors: Data Analysis, Error of Measurement, Research Problems, Statistical Inference
Peer reviewed Peer reviewed
Direct linkDirect link
Pokropek, Artur – Sociological Methods & Research, 2015
This article combines statistical and applied research perspective showing problems that might arise when measurement error in multilevel compositional effects analysis is ignored. This article focuses on data where independent variables are constructed measures. Simulation studies are conducted evaluating methods that could overcome the…
Descriptors: Error of Measurement, Hierarchical Linear Modeling, Simulation, Evaluation Methods
Reardon, Sean F.; Kalogrides, Demetra; Ho, Andrew D. – Stanford Center for Education Policy Analysis, 2017
There is no comprehensive database of U.S. district-level test scores that is comparable across states. We describe and evaluate a method for constructing such a database. First, we estimate linear, reliability-adjusted linking transformations from state test score scales to the scale of the National Assessment of Educational Progress (NAEP). We…
Descriptors: School Districts, Scores, Statistical Distributions, Database Design
Peer reviewed Peer reviewed
Direct linkDirect link
Sachse, Karoline A.; Roppelt, Alexander; Haag, Nicole – Journal of Educational Measurement, 2016
Trend estimation in international comparative large-scale assessments relies on measurement invariance between countries. However, cross-national differential item functioning (DIF) has been repeatedly documented. We ran a simulation study using national item parameters, which required trends to be computed separately for each country, to compare…
Descriptors: Comparative Analysis, Measurement, Test Bias, Simulation
Ackerman, Matthew; Egalite, Anna J. – Program on Education Policy and Governance, 2015
When lotteries are infeasible, researchers must rely on observational methods to estimate charter effectiveness at raising student test scores. Considerable attention has been paid to observational studies by the Stanford Center for Research on Education Outcomes (CREDO), which have analyzed charter performance in 27 states. However, the…
Descriptors: Charter Schools, Observation, Special Education, Lunch Programs
Peer reviewed Peer reviewed
Direct linkDirect link
Papay, John P. – American Educational Research Journal, 2011
Recently, educational researchers and practitioners have turned to value-added models to evaluate teacher performance. Although value-added estimates depend on the assessment used to measure student achievement, the importance of outcome selection has received scant attention in the literature. Using data from a large, urban school district, I…
Descriptors: Urban Schools, Teacher Effectiveness, Reading Achievement, Achievement Tests