Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 13 |
Descriptor
Comparative Analysis | 14 |
Hierarchical Linear Modeling | 14 |
Item Response Theory | 14 |
Computation | 7 |
Monte Carlo Methods | 5 |
Correlation | 4 |
Longitudinal Studies | 4 |
Markov Processes | 4 |
Maximum Likelihood Statistics | 4 |
Measurement | 4 |
Scores | 4 |
More ▼ |
Source
Author
Cho, Sun-Joo | 2 |
Algina, James | 1 |
Aydin, Burak | 1 |
Beretvas, S. Natasha | 1 |
Bottge, Brian | 1 |
Bottge, Brian A. | 1 |
Carstensen, Claus H. | 1 |
Chun Wang | 1 |
Cohen, Allan S. | 1 |
Cui, Ying | 1 |
He, Wei | 1 |
More ▼ |
Publication Type
Reports - Research | 12 |
Journal Articles | 11 |
Dissertations/Theses -… | 2 |
Education Level
Middle Schools | 6 |
Secondary Education | 5 |
Junior High Schools | 4 |
Elementary Education | 3 |
Grade 8 | 3 |
Grade 6 | 2 |
Intermediate Grades | 2 |
Elementary Secondary Education | 1 |
Grade 4 | 1 |
Grade 5 | 1 |
Grade 7 | 1 |
More ▼ |
Audience
Laws, Policies, & Programs
Assessments and Surveys
National Assessment of… | 1 |
National Education… | 1 |
What Works Clearinghouse Rating
Xue Zhang; Chun Wang – Grantee Submission, 2021
Among current state-of-art estimation methods for multilevel IRT models, the two-stage divide-and-conquer strategy has practical advantages, such as clearer definition of factors, convenience for secondary data analysis, convenience for model calibration and fit evaluation, and avoidance of improper solutions. However, various studies have shown…
Descriptors: Error of Measurement, Error Correction, Item Response Theory, Comparative Analysis
Kabasakal, Kübra Atalay; Kelecioglu, Hülya – Educational Sciences: Theory and Practice, 2015
This study examines the effect of differential item functioning (DIF) items on test equating through multilevel item response models (MIRMs) and traditional IRMs. The performances of three different equating models were investigated under 24 different simulation conditions, and the variables whose effects were examined included sample size, test…
Descriptors: Test Bias, Equated Scores, Item Response Theory, Simulation
Zhu, Xiaoshu – ProQuest LLC, 2013
The current study introduced a general modeling framework, multilevel mixture IRT (MMIRT) which detects and describes characteristics of population heterogeneity, while accommodating the hierarchical data structure. In addition to introducing both continuous and discrete approaches to MMIRT, the main focus of the current study was to distinguish…
Descriptors: Item Response Theory, Models, Comparative Analysis, Goodness of Fit
Cho, Sun-Joo; Bottge, Brian A. – Grantee Submission, 2015
In a pretest-posttest cluster-randomized trial, one of the methods commonly used to detect an intervention effect involves controlling pre-test scores and other related covariates while estimating an intervention effect at post-test. In many applications in education, the total post-test and pre-test scores that ignores measurement error in the…
Descriptors: Item Response Theory, Hierarchical Linear Modeling, Pretests Posttests, Scores
Itzek-Greulich, Heike; Vollmer, Christian – Journal of Research in Science Teaching, 2017
Students' interest in science declines in secondary school. Therefore, motivating students to become competent and engaged in science topics that are relevant for their everyday lives is an important goal, so they can be better citizens and decision makers with socioscientific issues (e.g., climate change and waste disposal). The present study…
Descriptors: Science Laboratories, Item Response Theory, Hierarchical Linear Modeling, Student Interests
Aydin, Burak; Leite, Walter L.; Algina, James – Educational and Psychological Measurement, 2016
We investigated methods of including covariates in two-level models for cluster randomized trials to increase power to detect the treatment effect. We compared multilevel models that included either an observed cluster mean or a latent cluster mean as a covariate, as well as the effect of including Level 1 deviation scores in the model. A Monte…
Descriptors: Error of Measurement, Predictor Variables, Randomized Controlled Trials, Experimental Groups
Cui, Ying; Mousavi, Amin – International Journal of Testing, 2015
The current study applied the person-fit statistic, l[subscript z], to data from a Canadian provincial achievement test to explore the usefulness of conducting person-fit analysis on large-scale assessments. Item parameter estimates were compared before and after the misfitting student responses, as identified by l[subscript z], were removed. The…
Descriptors: Measurement, Achievement Tests, Comparative Analysis, Test Items
Murphy, Daniel L.; Beretvas, S. Natasha – Applied Measurement in Education, 2015
This study examines the use of cross-classified random effects models (CCrem) and cross-classified multiple membership random effects models (CCMMrem) to model rater bias and estimate teacher effectiveness. Effect estimates are compared using CTT versus item response theory (IRT) scaling methods and three models (i.e., conventional multilevel…
Descriptors: Teacher Effectiveness, Comparative Analysis, Hierarchical Linear Modeling, Test Theory
Jiao, Hong; Wang, Shudong; He, Wei – Journal of Educational Measurement, 2013
This study demonstrated the equivalence between the Rasch testlet model and the three-level one-parameter testlet model and explored the Markov Chain Monte Carlo (MCMC) method for model parameter estimation in WINBUGS. The estimation accuracy from the MCMC method was compared with those from the marginalized maximum likelihood estimation (MMLE)…
Descriptors: Computation, Item Response Theory, Models, Monte Carlo Methods
Jeon, Minjeong – ProQuest LLC, 2012
Maximum likelihood (ML) estimation of generalized linear mixed models (GLMMs) is technically challenging because of the intractable likelihoods that involve high dimensional integrations over random effects. The problem is magnified when the random effects have a crossed design and thus the data cannot be reduced to small independent clusters. A…
Descriptors: Hierarchical Linear Modeling, Computation, Measurement, Maximum Likelihood Statistics
Cho, Sun-Joo; Cohen, Allan S.; Bottge, Brian – Grantee Submission, 2013
A multilevel latent transition analysis (LTA) with a mixture IRT measurement model (MixIRTM) is described for investigating the effectiveness of an intervention. The addition of a MixIRTM to the multilevel LTA permits consideration of both potential heterogeneity in students' response to instructional intervention as well as a methodology for…
Descriptors: Intervention, Item Response Theory, Statistical Analysis, Models
Yen, Wendy M.; Lall, Venessa F.; Monfils, Lora – ETS Research Report Series, 2012
Alternatives to vertical scales are compared for measuring longitudinal academic growth and for producing school-level growth measures. The alternatives examined were empirical cross-grade regression, ordinary least squares and logistic regression, and multilevel models. The student data used for the comparisons were Arabic Grades 4 to 10 in…
Descriptors: Foreign Countries, Scaling, Item Response Theory, Test Interpretation
von Davier, Alina A.; Carstensen, Claus H.; von Davier, Matthias – ETS Research Report Series, 2006
Measuring and linking competencies require special instruments, special data collection designs, and special statistical models. The measurement instruments are tests or tests forms, which can be used in the following situations: The same test can be given repeatedly; two or more parallel tests forms (i.e., forms intended to be similar in…
Descriptors: Scores, Measurement Techniques, Competence, Comparative Analysis
Johnson, Matthew S.; Jenkins, Frank – ETS Research Report Series, 2005
Large-scale educational assessments such as the National Assessment of Educational Progress (NAEP) sample examinees to whom an exam will be administered. In most situations the sampling design is not a simple random sample and must be accounted for in the estimating model. After reviewing the current operational estimation procedure for NAEP, this…
Descriptors: Bayesian Statistics, Hierarchical Linear Modeling, National Competency Tests, Sampling