NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 12 results Save | Export
Peer reviewed Peer reviewed
Direct linkDirect link
Pavlov, Goran; Maydeu-Olivares, Alberto; Shi, Dexin – Educational and Psychological Measurement, 2021
We examine the accuracy of p values obtained using the asymptotic mean and variance (MV) correction to the distribution of the sample standardized root mean squared residual (SRMR) proposed by Maydeu-Olivares to assess the exact fit of SEM models. In a simulation study, we found that under normality, the MV-corrected SRMR statistic provides…
Descriptors: Structural Equation Models, Goodness of Fit, Simulation, Error of Measurement
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational Measurement, 2018
Response-time models are of increasing interest in educational and psychological testing. This article focuses on the lognormal model for response times, which is one of the most popular response-time models, and suggests a simple person-fit statistic for the model. The distribution of the statistic under the null hypothesis of no misfit is proved…
Descriptors: Reaction Time, Educational Testing, Psychological Testing, Models
Sinharay, Sandip – Grantee Submission, 2018
Response-time models are of increasing interest in educational and psychological testing. This paper focuses on the lognormal model for response times (van der Linden, 2006), which is one of the most popular response-time models, and suggests a simple person-fit statistic for the model. The distribution of the statistic under the null hypothesis…
Descriptors: Reaction Time, Educational Testing, Psychological Testing, Models
Peer reviewed Peer reviewed
Direct linkDirect link
Ames, Allison J. – Measurement: Interdisciplinary Research and Perspectives, 2018
Bayesian item response theory (IRT) modeling stages include (a) specifying the IRT likelihood model, (b) specifying the parameter prior distributions, (c) obtaining the posterior distribution, and (d) making appropriate inferences. The latter stage, and the focus of this research, includes model criticism. Choice of priors with the posterior…
Descriptors: Bayesian Statistics, Item Response Theory, Statistical Inference, Prediction
Peer reviewed Peer reviewed
PDF on ERIC Download full text
Li, Zhen; Cai, Li – Grantee Submission, 2017
In standard item response theory (IRT) applications, the latent variable is typically assumed to be normally distributed. If the normality assumption is violated, the item parameter estimates can become biased. Summed score likelihood based statistics may be useful for testing latent variable distribution fit. We develop Satorra-Bentler type…
Descriptors: Scores, Goodness of Fit, Statistical Distributions, Item Response Theory
Peer reviewed Peer reviewed
Direct linkDirect link
Sinharay, Sandip – Journal of Educational and Behavioral Statistics, 2015
Person-fit assessment may help the researcher to obtain additional information regarding the answering behavior of persons. Although several researchers examined person fit, there is a lack of research on person-fit assessment for mixed-format tests. In this article, the lz statistic and the ?2 statistic, both of which have been used for tests…
Descriptors: Test Format, Goodness of Fit, Item Response Theory, Bayesian Statistics
Peer reviewed Peer reviewed
Direct linkDirect link
Beretvas, S. Natasha; Murphy, Daniel L. – Journal of Experimental Education, 2013
The authors assessed correct model identification rates of Akaike's information criterion (AIC), corrected criterion (AICC), consistent AIC (CAIC), Hannon and Quinn's information criterion (HQIC), and Bayesian information criterion (BIC) for selecting among cross-classified random effects models. Performance of default values for the 5…
Descriptors: Models, Goodness of Fit, Evaluation Criteria, Educational Research
Peer reviewed Peer reviewed
Direct linkDirect link
Verkuilen, Jay; Smithson, Michael – Journal of Educational and Behavioral Statistics, 2012
Doubly bounded continuous data are common in the social and behavioral sciences. Examples include judged probabilities, confidence ratings, derived proportions such as percent time on task, and bounded scale scores. Dependent variables of this kind are often difficult to analyze using normal theory models because their distributions may be quite…
Descriptors: Responses, Regression (Statistics), Statistical Analysis, Models
Peer reviewed Peer reviewed
Tanguma, Jesus – Educational and Psychological Measurement, 2001
Studied the effects of sample size on the cumulative distribution of selected fit indices using Monte Carlo simulation. Generally, the comparative fit index exhibited very stable patterns and was less influenced by sample size or data types than were other fit indices. (SLD)
Descriptors: Goodness of Fit, Monte Carlo Methods, Sample Size, Simulation
Peer reviewed Peer reviewed
van Krimpen-Stoop, Edith M. L. A.; Meijer, Rob – Applied Psychological Measurement, 1999
Theoretical null distributions of several fit statistic have been derived for paper-and-pencil tests. Examined whether these distributions also hold for computerized adaptive tests through simulation. Rates for two statistics studied were found to be similar in most cases. (SLD)
Descriptors: Adaptive Testing, Computer Assisted Testing, Goodness of Fit, Item Response Theory
Peer reviewed Peer reviewed
Mount, Robert E.; Schumacker, Randall E. – Journal of Outcome Measurement, 1998
A Monte Carlo study was conducted using simulated dichotomous data to determine the effects of guessing on Rasch item fit statistics and the Logit Residual Index. Results indicate that no significant differences were found between the mean Rasch item fit statistics for each distribution type as the probability of guessing the correct answer…
Descriptors: Goodness of Fit, Guessing (Tests), Item Response Theory, Monte Carlo Methods
Peer reviewed Peer reviewed
Smith, Richard M. – Educational and Psychological Measurement, 1994
Simulated data are used to assess the appropriateness of using separate calibration and between-fit approaches to detecting item bias in the Rasch rating scale model. Results indicate that Type I error rates for the null distribution hold even when there are different ability levels for reference and focal groups. (SLD)
Descriptors: Ability, Goodness of Fit, Identification, Item Bias