Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 5 |
Since 2006 (last 20 years) | 14 |
Descriptor
Evaluation Methods | 17 |
Item Response Theory | 17 |
Nonparametric Statistics | 17 |
Models | 8 |
Simulation | 7 |
Statistical Analysis | 6 |
Sample Size | 5 |
Test Items | 5 |
Computation | 4 |
Test Bias | 4 |
Error of Measurement | 3 |
More ▼ |
Source
Author
Sijtsma, Klaas | 2 |
Anthony Petrosino | 1 |
Benjamin Lugu | 1 |
Bolt, Daniel M. | 1 |
Cheng, Ying | 1 |
Corey Brady | 1 |
DeMars, Christine E. | 1 |
Dorans, Neil J. | 1 |
Douglas, Jeff | 1 |
Edwards, Julianne M. | 1 |
Emons, Wilco H. M. | 1 |
More ▼ |
Publication Type
Journal Articles | 16 |
Reports - Research | 8 |
Reports - Evaluative | 5 |
Reports - Descriptive | 4 |
Speeches/Meeting Papers | 1 |
Education Level
Elementary Education | 1 |
Grade 6 | 1 |
Intermediate Grades | 1 |
Middle Schools | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
State of Texas Assessments of… | 1 |
What Works Clearinghouse Rating
Stefanie A. Wind; Benjamin Lugu – Applied Measurement in Education, 2024
Researchers who use measurement models for evaluation purposes often select models with stringent requirements, such as Rasch models, which are parametric. Mokken Scale Analysis (MSA) offers a theory-driven nonparametric modeling approach that may be more appropriate for some measurement applications. Researchers have discussed using MSA as a…
Descriptors: Item Response Theory, Data Analysis, Simulation, Nonparametric Statistics
Walter M. Stroup; Anthony Petrosino; Corey Brady; Karen Duseau – North American Chapter of the International Group for the Psychology of Mathematics Education, 2023
Tests of statistical significance often play a decisive role in establishing the empirical warrant of evidence-based research in education. The results from pattern-based assessment items, as introduced in this paper, are categorical and multimodal and do not immediately support the use of measures of central tendency as typically related to…
Descriptors: Statistical Significance, Comparative Analysis, Research Methodology, Evaluation Methods
Wind, Stefanie A. – Educational Measurement: Issues and Practice, 2017
Mokken scale analysis (MSA) is a probabilistic-nonparametric approach to item response theory (IRT) that can be used to evaluate fundamental measurement properties with less strict assumptions than parametric IRT models. This instructional module provides an introduction to MSA as a probabilistic-nonparametric framework in which to explore…
Descriptors: Probability, Nonparametric Statistics, Item Response Theory, Scaling
Lathrop, Quinn N.; Cheng, Ying – Journal of Educational Measurement, 2014
When cut scores for classifications occur on the total score scale, popular methods for estimating classification accuracy (CA) and classification consistency (CC) require assumptions about a parametric form of the test scores or about a parametric response model, such as item response theory (IRT). This article develops an approach to estimate CA…
Descriptors: Cutting Scores, Classification, Computation, Nonparametric Statistics
Finch, Holmes; Edwards, Julianne M. – Educational and Psychological Measurement, 2016
Standard approaches for estimating item response theory (IRT) model parameters generally work under the assumption that the latent trait being measured by a set of items follows the normal distribution. Estimation of IRT parameters in the presence of nonnormal latent traits has been shown to generate biased person and item parameter estimates. A…
Descriptors: Item Response Theory, Computation, Nonparametric Statistics, Bayesian Statistics
Socha, Alan; DeMars, Christine E.; Zilberberg, Anna; Phan, Ha – International Journal of Testing, 2015
The Mantel-Haenszel (MH) procedure is commonly used to detect items that function differentially for groups of examinees from various demographic and linguistic backgrounds--for example, in international assessments. As in some other DIF methods, the total score is used to match examinees on ability. In thin matching, each of the total score…
Descriptors: Test Items, Educational Testing, Evaluation Methods, Ability Grouping
Kalkan, Ömür Kaya; Kelecioglu, Hülya – Educational Sciences: Theory and Practice, 2016
Linear factor analysis models used to examine constructs underlying the responses are not very suitable for dichotomous or polytomous response formats. The associated problems cannot be eliminated by polychoric or tetrachoric correlations in place of the Pearson correlation. Therefore, we considered parameters obtained from the NOHARM and FACTOR…
Descriptors: Sample Size, Nonparametric Statistics, Factor Analysis, Correlation
Karabatsos, George; Walker, Stephen G. – Psychometrika, 2009
A Bayesian nonparametric model is introduced for score equating. It is applicable to all major equating designs, and has advantages over previous equating models. Unlike the previous models, the Bayesian model accounts for positive dependence between distributions of scores from two tests. The Bayesian model and the previous equating models are…
Descriptors: Nonparametric Statistics, Item Response Theory, Models, Comparative Analysis
Wells, Craig S.; Bolt, Daniel M. – Applied Measurement in Education, 2008
Tests of model misfit are often performed to validate the use of a particular model in item response theory. Douglas and Cohen (2001) introduced a general nonparametric approach for detecting misfit under the two-parameter logistic model. However, the statistical properties of their approach, and empirical comparisons to other methods, have not…
Descriptors: Test Length, Test Items, Monte Carlo Methods, Nonparametric Statistics
Lee, Young-Sun – Applied Psychological Measurement, 2007
This study compares the performance of three nonparametric item characteristic curve (ICC) estimation procedures: isotonic regression, smoothed isotonic regression, and kernel smoothing. Smoothed isotonic regression, employed along with an appropriate kernel function, provides better estimates and also satisfies the assumption of strict…
Descriptors: Nonparametric Statistics, Computation, Item Response Theory, Evaluation Methods
Mapuranga, Raymond; Dorans, Neil J.; Middleton, Kyndra – ETS Research Report Series, 2008
In many practical settings, essentially the same differential item functioning (DIF) procedures have been in use since the late 1980s. Since then, examinee populations have become more heterogeneous, and tests have included more polytomously scored items. This paper summarizes and classifies new DIF methods and procedures that have appeared since…
Descriptors: Test Bias, Educational Development, Evaluation Methods, Statistical Analysis

Junker, Brian; Sijtsma, Klaas – Applied Psychological Measurement, 2001
Discusses usability and interpretation issues for single-strategy cognitive assessment models that posit a stochastic, conjunctive relationship between a set of cognitive attributes to be assessed and performance on particular items/tasks of the assessment. Also discusses stochastic ordering and monotonicity properties that enhance the…
Descriptors: Cognitive Processes, Evaluation Methods, Item Response Theory, Models
Johnson, Matthew S. – Psychometrika, 2006
Unlike their monotone counterparts, nonparametric unfolding response models, which assume the item response function is unimodal, have seen little attention in the psychometric literature. This paper studies the nonparametric behavior of unfolding models by building on the work of Post (1992). The paper provides rigorous justification for a class…
Descriptors: Psychometrics, Nonparametric Statistics, Item Response Theory, Models
Xu, Xueli; Douglas, Jeff – Psychometrika, 2006
Nonparametric item response models have been developed as alternatives to the relatively inflexible parametric item response models. An open question is whether it is possible and practical to administer computerized adaptive testing with nonparametric models. This paper explores the possibility of computerized adaptive testing when using…
Descriptors: Simulation, Nonparametric Statistics, Item Analysis, Item Response Theory
Emons, Wilco H. M.; Sijtsma, Klaas; Meijer, Rob R. – Psychological Methods, 2005
Person-fit statistics test whether the likelihood of a respondent's complete vector of item scores on a test is low given the hypothesized item response theory model. This binary information may be insufficient for diagnosing the cause of a misfitting item-score vector. The authors propose a comprehensive methodology for person-fit analysis in the…
Descriptors: Evaluation Methods, Item Response Theory, Evaluation Research, Goodness of Fit
Previous Page | Next Page »
Pages: 1 | 2