Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 4 |
Since 2016 (last 10 years) | 7 |
Since 2006 (last 20 years) | 14 |
Descriptor
Error of Measurement | 17 |
Simulation | 17 |
Computation | 7 |
Item Response Theory | 7 |
Models | 6 |
Statistical Analysis | 5 |
Bayesian Statistics | 4 |
Statistical Inference | 4 |
Data Analysis | 3 |
Equated Scores | 3 |
Evaluation Methods | 3 |
More ▼ |
Source
Journal of Educational and… | 17 |
Author
Grund, Simon | 2 |
Lüdtke, Oliver | 2 |
Robitzsch, Alexander | 2 |
Bartoš, František | 1 |
Brabec, Marek | 1 |
Browne, Michael W. | 1 |
Cao, Jing | 1 |
Fan, Weihua | 1 |
Hancock, Gregory R. | 1 |
Hoskens, Machteld | 1 |
Jiang, Yanlin | 1 |
More ▼ |
Publication Type
Journal Articles | 17 |
Reports - Research | 10 |
Reports - Evaluative | 4 |
Reports - Descriptive | 2 |
Information Analyses | 1 |
Speeches/Meeting Papers | 1 |
Education Level
Elementary Secondary Education | 2 |
Grade 8 | 1 |
Secondary Education | 1 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Behavioral Risk Factor… | 1 |
National Assessment of… | 1 |
Program for International… | 1 |
Trends in International… | 1 |
What Works Clearinghouse Rating
Grund, Simon; Lüdtke, Oliver; Robitzsch, Alexander – Journal of Educational and Behavioral Statistics, 2023
Multiple imputation (MI) is a popular method for handling missing data. In education research, it can be challenging to use MI because the data often have a clustered structure that need to be accommodated during MI. Although much research has considered applications of MI in hierarchical data, little is known about its use in cross-classified…
Descriptors: Educational Research, Data Analysis, Error of Measurement, Computation
Martinková, Patrícia; Bartoš, František; Brabec, Marek – Journal of Educational and Behavioral Statistics, 2023
Inter-rater reliability (IRR), which is a prerequisite of high-quality ratings and assessments, may be affected by contextual variables, such as the rater's or ratee's gender, major, or experience. Identification of such heterogeneity sources in IRR is important for the implementation of policies with the potential to decrease measurement error…
Descriptors: Interrater Reliability, Bayesian Statistics, Statistical Inference, Hierarchical Linear Modeling
Pashley, Nicole E.; Miratrix, Luke W. – Journal of Educational and Behavioral Statistics, 2021
Evaluating blocked randomized experiments from a potential outcomes perspective has two primary branches of work. The first focuses on larger blocks, with multiple treatment and control units in each block. The second focuses on matched pairs, with a single treatment and control unit in each block. These literatures not only provide different…
Descriptors: Causal Models, Statistical Inference, Research Methodology, Computation
Monroe, Scott – Journal of Educational and Behavioral Statistics, 2019
In item response theory (IRT) modeling, the Fisher information matrix is used for numerous inferential procedures such as estimating parameter standard errors, constructing test statistics, and facilitating test scoring. In principal, these procedures may be carried out using either the expected information or the observed information. However, in…
Descriptors: Item Response Theory, Error of Measurement, Scoring, Inferences
Parsons, Eric; Koedel, Cory; Tan, Li – Journal of Educational and Behavioral Statistics, 2019
We study the relative performance of two policy-relevant value-added models--a one-step fixed effect model and a two-step aggregated residuals model--using a simulated data set well grounded in the value-added literature. A key feature of our data generating process is that student achievement depends on a continuous measure of economic…
Descriptors: Value Added Models, Economically Disadvantaged, Academic Achievement, Low Income Students
Grund, Simon; Lüdtke, Oliver; Robitzsch, Alexander – Journal of Educational and Behavioral Statistics, 2021
Large-scale assessments (LSAs) use Mislevy's "plausible value" (PV) approach to relate student proficiency to noncognitive variables administered in a background questionnaire. This method requires background variables to be completely observed, a requirement that is seldom fulfilled. In this article, we evaluate and compare the…
Descriptors: Data Analysis, Error of Measurement, Research Problems, Statistical Inference
Magnus, Brooke E.; Thissen, David – Journal of Educational and Behavioral Statistics, 2017
Questionnaires that include items eliciting count responses are becoming increasingly common in psychology. This study proposes methodological techniques to overcome some of the challenges associated with analyzing multivariate item response data that exhibit zero inflation, maximum inflation, and heaping at preferred digits. The modeling…
Descriptors: Item Response Theory, Models, Multivariate Analysis, Questionnaires
Liang, Longjuan; Browne, Michael W. – Journal of Educational and Behavioral Statistics, 2015
If standard two-parameter item response functions are employed in the analysis of a test with some newly constructed items, it can be expected that, for some items, the item response function (IRF) will not fit the data well. This lack of fit can also occur when standard IRFs are fitted to personality or psychopathology items. When investigating…
Descriptors: Item Response Theory, Statistical Analysis, Goodness of Fit, Bayesian Statistics
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2015
Person-fit assessment may help the researcher to obtain additional information regarding the answering behavior of persons. Although several researchers examined person fit, there is a lack of research on person-fit assessment for mixed-format tests. In this article, the lz statistic and the ?2 statistic, both of which have been used for tests…
Descriptors: Test Format, Goodness of Fit, Item Response Theory, Bayesian Statistics
Fan, Weihua; Hancock, Gregory R. – Journal of Educational and Behavioral Statistics, 2012
This study proposes robust means modeling (RMM) approaches for hypothesis testing of mean differences for between-subjects designs in order to control the biasing effects of nonnormality and variance inequality. Drawing from structural equation modeling (SEM), the RMM approaches make no assumption of variance homogeneity and employ robust…
Descriptors: Robustness (Statistics), Hypothesis Testing, Monte Carlo Methods, Simulation
Cao, Jing; Stokes, S. Lynne; Zhang, Song – Journal of Educational and Behavioral Statistics, 2010
We develop a Bayesian hierarchical model for the analysis of ordinal data from multirater ranking studies. The model for a rater's score includes four latent factors: one is a latent item trait determining the true order of items and the other three are the rater's performance characteristics, including bias, discrimination, and measurement error…
Descriptors: Bayesian Statistics, Data Analysis, Bias, Measurement
Wang, Tianyou – Journal of Educational and Behavioral Statistics, 2009
Holland and colleagues derived a formula for analytical standard error of equating using the delta-method for the kernel equating method. Extending their derivation, this article derives an analytical standard error of equating procedure for the conventional percentile rank-based equipercentile equating with log-linear smoothing. This procedure is…
Descriptors: Error of Measurement, Equated Scores, Statistical Analysis, Statistical Inference
Li, Deping; Oranje, Andreas; Jiang, Yanlin – Journal of Educational and Behavioral Statistics, 2009
To find population proficiency distributions, a two-level hierarchical linear model may be applied to large-scale survey assessments such as the National Assessment of Educational Progress (NAEP). The model and parameter estimation are developed and a simulation was carried out to evaluate parameter recovery. Subsequently, both a hierarchical and…
Descriptors: Computation, National Competency Tests, Measurement, Regression (Statistics)
Moses, Tim – Journal of Educational and Behavioral Statistics, 2008
Equating functions are supposed to be population invariant, meaning that the choice of subpopulation used to compute the equating function should not matter. The extent to which equating functions are population invariant is typically assessed in terms of practical difference criteria that do not account for equating functions' sampling…
Descriptors: Equated Scores, Error of Measurement, Sampling, Evaluation Methods

Wilson, Mark; Hoskens, Machteld – Journal of Educational and Behavioral Statistics, 2001
Introduces the Rater Bundle Model, an item response model for repeated ratings of student work. Applies the model to real and simulated data to illustrate the approach, which was motivated by the observation that when repeated ratings occur, the assumption of conditional independence is violated, and current item response models can then…
Descriptors: Error of Measurement, Evaluators, Item Response Theory, Models
Previous Page | Next Page »
Pages: 1 | 2