NotesFAQContact Us
Collection
Advanced
Search Tips
Audience
Laws, Policies, & Programs
No Child Left Behind Act 20011
Assessments and Surveys
What Works Clearinghouse Rating
Does not meet standards1
Showing all 10 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Stephen M. Leach; Jason C. Immekus; Jeffrey C. Valentine; Prathiba Batley; Dena Dossett; Tamara Lewis; Thomas Reece – Assessment for Effective Intervention, 2025
Educators commonly use school climate survey scores to inform and evaluate interventions for equitably improving learning and reducing educational disparities. Unfortunately, validity evidence to support these (and other) score uses often falls short. In response, Whitehouse et al. proposed a collaborative, two-part validity testing framework for…
Descriptors: School Surveys, Measurement, Hierarchical Linear Modeling, Educational Environment
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Chine, Danielle R.; Larwin, Karen H. – International Journal of Research in Education and Science, 2022
Hierarchical linear modeling (HLM) has become an increasingly popular multilevel method of analyzing data among nested datasets, in particular, the effect of specialized academic programming within schools. The purpose of this methodological study is to demonstrate the use of HLM to determine the effectiveness of STEM programming in an Ohio middle…
Descriptors: Middle Schools, STEM Education, Instructional Effectiveness, Program Development
Peer reviewed Peer reviewed
Direct linkDirect link
Sulis, Isabella; Toland, Michael D. – Journal of Early Adolescence, 2017
Item response theory (IRT) models are the main psychometric approach for the development, evaluation, and refinement of multi-item instruments and scaling of latent traits, whereas multilevel models are the primary statistical method when considering the dependence between person responses when primary units (e.g., students) are nested within…
Descriptors: Hierarchical Linear Modeling, Item Response Theory, Psychometrics, Evaluation Methods
Buckley, Pamela; Moore, Brooke; Boardman, Alison G.; Arya, Diana J.; Maul, Andrew – American Educational Research Journal, 2017
K-12 intervention studies often include fidelity of implementation (FOI) as a mediating variable, though most do not report the validity of fidelity measures. This article discusses the critical need for validated FOI scales. To illustrate our point, we describe the development and validation of the Implementation Validity Checklist (IVC-R), an…
Descriptors: Intervention, Fidelity, Program Implementation, Test Validity
Peer reviewed Peer reviewed
Direct linkDirect link
Hecht, Martin; Weirich, Sebastian; Siegle, Thilo; Frey, Andreas – Educational and Psychological Measurement, 2015
The selection of an appropriate booklet design is an important element of large-scale assessments of student achievement. Two design properties that are typically optimized are the "balance" with respect to the positions the items are presented and with respect to the mutual occurrence of pairs of items in the same booklet. The purpose…
Descriptors: Measurement, Computation, Test Format, Test Items
Jaciw, Andrew P.; Nguyen, Thanh; Lin, Li; Zacamy, Jenna L.; Kwong, Connie; Lau, Sze-Shun – Grantee Submission, 2020
These appendices accompany the report "Final Report of the i3 Impact Study of Making Sense of SCIENCE, 2016-17 through 2017-18." Science education has experienced a significant transition over the last decade, catalyzed by a re-envisioning of what students should know and be able to do in science. That re-envisioning culminated in the…
Descriptors: Faculty Development, Science Instruction, Science Achievement, Elementary School Teachers
Peer reviewed Peer reviewed
Direct linkDirect link
Cui, Ying; Mousavi, Amin – International Journal of Testing, 2015
The current study applied the person-fit statistic, l[subscript z], to data from a Canadian provincial achievement test to explore the usefulness of conducting person-fit analysis on large-scale assessments. Item parameter estimates were compared before and after the misfitting student responses, as identified by l[subscript z], were removed. The…
Descriptors: Measurement, Achievement Tests, Comparative Analysis, Test Items
Peer reviewed Peer reviewed
Direct linkDirect link
Murphy, Daniel L.; Beretvas, S. Natasha – Applied Measurement in Education, 2015
This study examines the use of cross-classified random effects models (CCrem) and cross-classified multiple membership random effects models (CCMMrem) to model rater bias and estimate teacher effectiveness. Effect estimates are compared using CTT versus item response theory (IRT) scaling methods and three models (i.e., conventional multilevel…
Descriptors: Teacher Effectiveness, Comparative Analysis, Hierarchical Linear Modeling, Test Theory
Cho, Sun-Joo; Cohen, Allan S.; Bottge, Brian – Grantee Submission, 2013
A multilevel latent transition analysis (LTA) with a mixture IRT measurement model (MixIRTM) is described for investigating the effectiveness of an intervention. The addition of a MixIRTM to the multilevel LTA permits consideration of both potential heterogeneity in students' response to instructional intervention as well as a methodology for…
Descriptors: Intervention, Item Response Theory, Statistical Analysis, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Singh, Malkeet – Educational Research and Evaluation, 2013
Eliminating inequity in public education is a central goal of the No Child left Behind (NCLB) act. Controlling for 3rd-grade performance, the impact of English language learner (ELL) status, socioeconomic status (SES), and special education (SPED) status on a cohort's reading performance was investigated from elementary to high school through a…
Descriptors: Reading Tests, Educational Assessment, Measurement, Longitudinal Studies