Publication Date
In 2025 | 0 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 10 |
Since 2006 (last 20 years) | 23 |
Descriptor
Goodness of Fit | 51 |
Nonparametric Statistics | 51 |
Item Response Theory | 24 |
Statistical Analysis | 17 |
Comparative Analysis | 14 |
Models | 11 |
Hypothesis Testing | 10 |
Simulation | 10 |
Sample Size | 8 |
Statistical Significance | 8 |
Test Items | 8 |
More ▼ |
Source
Author
Wells, Craig S. | 4 |
Liang, Tie | 3 |
Baker, Frank B. | 2 |
Douglas, Jeffrey | 2 |
Emons, Wilco H. M. | 2 |
McSweeney, Maryellen | 2 |
Meijer, Rob R. | 2 |
Porter, Andrew C. | 2 |
Sinharay, Sandip | 2 |
Tendeiro, Jorge N. | 2 |
Abad, Francisco J. | 1 |
More ▼ |
Publication Type
Journal Articles | 34 |
Reports - Research | 31 |
Reports - Evaluative | 6 |
Reports - Descriptive | 3 |
Speeches/Meeting Papers | 2 |
Book/Product Reviews | 1 |
Dissertations/Theses -… | 1 |
Non-Print Media | 1 |
Education Level
Elementary Education | 2 |
Elementary Secondary Education | 2 |
Secondary Education | 2 |
Grade 8 | 1 |
Junior High Schools | 1 |
Middle Schools | 1 |
Audience
Laws, Policies, & Programs
Assessments and Surveys
Trends in International… | 1 |
What Works Clearinghouse Rating
Roya Shoahosseini; Purya Baghaei; Hossein Khodabakhshzadeh; Hamid Ashraf – Language Testing in Asia, 2024
C-Test is a gap-filling test designed to measure first and second language proficiency. Over the past four decades, researchers have shown the fit of C-Test data to parametric item response theory (IRT) models, but no study so far has shown the fit of C-Tests to nonparametric IRT models. The purpose of this study is to contribute to the ongoing…
Descriptors: Item Response Theory, Nonparametric Statistics, Language Proficiency, Second Language Learning
Reimers, Jennifer; Turner, Ronna C.; Tendeiro, Jorge N.; Lo, Wen-Juo; Keiffer, Elizabeth – Measurement: Interdisciplinary Research and Perspectives, 2023
Person-fit analyses are commonly used to detect aberrant responding in self-report data. Nonparametric person fit statistics do not require fitting a parametric test theory model and have performed well compared to other person-fit statistics. However, detection of aberrant responding has primarily focused on dominance response data, thus the…
Descriptors: Goodness of Fit, Nonparametric Statistics, Error of Measurement, Comparative Analysis
Wind, Stefanie A. – Measurement: Interdisciplinary Research and Perspectives, 2020
Rater fit analyses provide insight into the degree to which rater judgments correspond to expected properties, as defined within a measurement framework. Parametric models such as the Rasch model provide a useful framework for evaluating rating quality; however, these models are not appropriate for all assessment contexts. The purpose of this…
Descriptors: Evaluators, Goodness of Fit, Simulation, Psychometrics
Dimitrov, Dimiter M.; Atanasov, Dimitar V.; Luo, Yong – Measurement: Interdisciplinary Research and Perspectives, 2020
This study examines and compares four person-fit statistics (PFSs) in the framework of the "D"- scoring method (DSM): (a) van der Flier's "U3" statistic; (b) "Ud" statistic, as a modification of "U3" under the DSM; (c) "Zd" statistic, as a modification of the "Z3 (l[subscript z])"…
Descriptors: Goodness of Fit, Item Analysis, Item Response Theory, Scoring
Delafontaine, Jolien; Chen, Changsheng; Park, Jung Yeon; Van den Noortgate, Wim – Large-scale Assessments in Education, 2022
In cognitive diagnosis assessment (CDA), the impact of misspecified item-attribute relations (or "Q-matrix") designed by subject-matter experts has been a great challenge to real-world applications. This study examined parameter estimation of the CDA with the expert-designed Q-matrix and two refined Q-matrices for international…
Descriptors: Q Methodology, Matrices, Cognitive Measurement, Diagnostic Tests
Sinharay, Sandip – Journal of Educational Measurement, 2017
Person-fit assessment (PFA) is concerned with uncovering atypical test performance as reflected in the pattern of scores on individual items on a test. Existing person-fit statistics (PFSs) include both parametric and nonparametric statistics. Comparison of PFSs has been a popular research topic in PFA, but almost all comparisons have employed…
Descriptors: Goodness of Fit, Testing, Test Items, Scores
Berger, Moritz; Tutz, Gerhard – Journal of Educational and Behavioral Statistics, 2016
Detection of differential item functioning (DIF) by use of the logistic modeling approach has a long tradition. One big advantage of the approach is that it can be used to investigate nonuniform (NUDIF) as well as uniform DIF (UDIF). The classical approach allows one to detect DIF by distinguishing between multiple groups. We propose an…
Descriptors: Test Bias, Regression (Statistics), Nonparametric Statistics, Statistical Analysis
Sinharay, Sandip – Applied Measurement in Education, 2017
Karabatsos compared the power of 36 person-fit statistics using receiver operating characteristics curves and found the "H[superscript T]" statistic to be the most powerful in identifying aberrant examinees. He found three statistics, "C", "MCI", and "U3", to be the next most powerful. These four statistics,…
Descriptors: Nonparametric Statistics, Goodness of Fit, Simulation, Comparative Analysis
Liang, Tie; Wells, Craig S. – Applied Measurement in Education, 2015
Investigating the fit of a parametric model plays a vital role in validating an item response theory (IRT) model. An area that has received little attention is the assessment of multiple IRT models used in a mixed-format test. The present study extends the nonparametric approach, proposed by Douglas and Cohen (2001), to assess model fit of three…
Descriptors: Nonparametric Statistics, Goodness of Fit, Item Response Theory, Test Format
Liang, Longjuan; Browne, Michael W. – Journal of Educational and Behavioral Statistics, 2015
If standard two-parameter item response functions are employed in the analysis of a test with some newly constructed items, it can be expected that, for some items, the item response function (IRF) will not fit the data well. This lack of fit can also occur when standard IRFs are fitted to personality or psychopathology items. When investigating…
Descriptors: Item Response Theory, Statistical Analysis, Goodness of Fit, Bayesian Statistics
Sengul Avsar, Asiye; Tavsancil, Ezel – Educational Sciences: Theory and Practice, 2017
This study analysed polytomous items' psychometric properties according to nonparametric item response theory (NIRT) models. Thus, simulated datasets--three different test lengths (10, 20 and 30 items), three sample distributions (normal, right and left skewed) and three samples sizes (100, 250 and 500)--were generated by conducting 20…
Descriptors: Test Items, Psychometrics, Nonparametric Statistics, Item Response Theory
Golino, Hudson F.; Gomes, Cristiano M. A. – International Journal of Research & Method in Education, 2016
This paper presents a non-parametric imputation technique, named random forest, from the machine learning field. The random forest procedure has two main tuning parameters: the number of trees grown in the prediction and the number of predictors used. Fifty experimental conditions were created in the imputation procedure, with different…
Descriptors: Item Response Theory, Regression (Statistics), Difficulty Level, Goodness of Fit
Liang, Tie; Wells, Craig S.; Hambleton, Ronald K. – Journal of Educational Measurement, 2014
As item response theory has been more widely applied, investigating the fit of a parametric model becomes an important part of the measurement process. There is a lack of promising solutions to the detection of model misfit in IRT. Douglas and Cohen introduced a general nonparametric approach, RISE (Root Integrated Squared Error), for detecting…
Descriptors: Item Response Theory, Measurement Techniques, Nonparametric Statistics, Models
Tendeiro, Jorge N.; Meijer, Rob R. – Applied Psychological Measurement, 2013
To classify an item score pattern as not fitting a nonparametric item response theory (NIRT) model, the probability of exceedance (PE) of an observed response vector x can be determined as the sum of the probabilities of all response vectors that are, at most, as likely as x, conditional on the test's total score. Vector x is to be considered…
Descriptors: Probability, Nonparametric Statistics, Goodness of Fit, Test Length
Collins, Anne G. E.; Frank, Michael J. – Psychological Review, 2013
Learning and executive functions such as task-switching share common neural substrates, notably prefrontal cortex and basal ganglia. Understanding how they interact requires studying how cognitive control facilitates learning but also how learning provides the (potentially hidden) structure, such as abstract rules or task-sets, needed for…
Descriptors: Learning, Executive Function, Models, Bayesian Statistics