Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 8 |
Descriptor
Nonparametric Statistics | 15 |
Scores | 15 |
Test Items | 15 |
Item Response Theory | 10 |
Simulation | 4 |
Comparative Analysis | 3 |
Error of Measurement | 3 |
Identification | 3 |
Psychometrics | 3 |
Statistical Analysis | 3 |
Behavior Disorders | 2 |
More ▼ |
Source
Author
Meijer, Rob R. | 4 |
Sijtsma, Klaas | 3 |
Cui, Zhongmin | 1 |
Davis, John L. | 1 |
DeMars, Christine E. | 1 |
Emons, Wilco H. M. | 1 |
Evans, Steven W. | 1 |
Gierl, Mark J. | 1 |
Guo, Hongwen | 1 |
Haberman, Shelby | 1 |
Harrison, Judith R. | 1 |
More ▼ |
Publication Type
Journal Articles | 13 |
Reports - Research | 7 |
Reports - Evaluative | 6 |
Information Analyses | 1 |
Numerical/Quantitative Data | 1 |
Reports - Descriptive | 1 |
Education Level
High Schools | 2 |
Higher Education | 2 |
Postsecondary Education | 2 |
Secondary Education | 2 |
Audience
Location
Laws, Policies, & Programs
Assessments and Surveys
Raven Progressive Matrices | 1 |
SAT (College Admission Test) | 1 |
What Works Clearinghouse Rating
Myszkowski, Nils – Journal of Intelligence, 2020
Raven's Standard Progressive Matrices (Raven 1941) is a widely used 60-item long measure of general mental ability. It was recently suggested that, for situations where taking this test is too time consuming, a shorter version, comprised of only the last series of the Standard Progressive Matrices (Myszkowski and Storme 2018) could be used, while…
Descriptors: Intelligence Tests, Psychometrics, Nonparametric Statistics, Item Response Theory
Sinharay, Sandip – Journal of Educational Measurement, 2017
Person-fit assessment (PFA) is concerned with uncovering atypical test performance as reflected in the pattern of scores on individual items on a test. Existing person-fit statistics (PFSs) include both parametric and nonparametric statistics. Comparison of PFSs has been a popular research topic in PFA, but almost all comparisons have employed…
Descriptors: Goodness of Fit, Testing, Test Items, Scores
Guo, Hongwen; Rios, Joseph A.; Haberman, Shelby; Liu, Ou Lydia; Wang, Jing; Paek, Insu – Applied Measurement in Education, 2016
Unmotivated test takers using rapid guessing in item responses can affect validity studies and teacher and institution performance evaluation negatively, making it critical to identify these test takers. The authors propose a new nonparametric method for finding response-time thresholds for flagging item responses that result from rapid-guessing…
Descriptors: Guessing (Tests), Reaction Time, Nonparametric Statistics, Models
Socha, Alan; DeMars, Christine E.; Zilberberg, Anna; Phan, Ha – International Journal of Testing, 2015
The Mantel-Haenszel (MH) procedure is commonly used to detect items that function differentially for groups of examinees from various demographic and linguistic backgrounds--for example, in international assessments. As in some other DIF methods, the total score is used to match examinees on ability. In thin matching, each of the total score…
Descriptors: Test Items, Educational Testing, Evaluation Methods, Ability Grouping
Harrison, Judith R.; State, Talida M.; Evans, Steven W.; Schamberg, Terah – Journal of Positive Behavior Interventions, 2016
The purpose of this study was to evaluate the construct and predictive validity of scores on a measure of social acceptability of class-wide and individual student intervention, the School Intervention Rating Form (SIRF), with high school teachers. Utilizing scores from 158 teachers, exploratory factor analysis revealed a three-factor (i.e.,…
Descriptors: Construct Validity, Predictive Validity, Likert Scales, Intervention
Vannest, Kimberly J.; Parker, Richard I.; Davis, John L.; Soares, Denise A.; Smith, Stacey L. – Behavioral Disorders, 2012
More and more, schools are considering the use of progress monitoring data for high-stakes decisions such as special education eligibility, program changes to more restrictive environments, and major changes in educational goals. Those high-stakes types of data-based decisions will need methodological defensibility. Current practice for…
Descriptors: Decision Making, Educational Change, Regression (Statistics), Field Tests
Cui, Zhongmin; Kolen, Michael J. – Applied Psychological Measurement, 2008
This article considers two methods of estimating standard errors of equipercentile equating: the parametric bootstrap method and the nonparametric bootstrap method. Using a simulation study, these two methods are compared under three sample sizes (300, 1,000, and 3,000), for two test content areas (the Iowa Tests of Basic Skills Maps and Diagrams…
Descriptors: Test Length, Test Content, Simulation, Computation
Emons, Wilco H. M. – Applied Psychological Measurement, 2008
Person-fit methods are used to uncover atypical test performance as reflected in the pattern of scores on individual items in a test. Unlike parametric person-fit statistics, nonparametric person-fit statistics do not require fitting a parametric test theory model. This study investigates the effectiveness of generalizations of nonparametric…
Descriptors: Simulation, Nonparametric Statistics, Item Response Theory, Goodness of Fit
Meijer, Rob R.; Sijtsma, Klaas – 1994
Methods for detecting item score patterns that are unlikely (aberrant) given that a parametric item response theory (IRT) model gives an adequate description of the data or given the responses of the other persons in the group are discussed. The emphasis here is on the latter group of statistics. These statistics can be applied when a…
Descriptors: Foreign Countries, Identification, Item Response Theory, Nonparametric Statistics

Meijer, Rob R.; Sijtsma, Klaas – Applied Measurement in Education, 1995
Methods for detecting item score patterns that are unlikely, given that a parametric item response theory model gives an adequate description of the data or given the responses of other persons in the group, are discussed. The use of person-fit statistics in empirical data analysis is briefly discussed. (SLD)
Descriptors: Identification, Item Response Theory, Nonparametric Statistics, Patterns in Mathematics

Sijtsma, Klaas – Applied Psychological Measurement, 1998
Reviews developments in nonparametric item-response theory (NIRT), from its historic origins in item-response theory (IRT) and scale analysis to new theoretical results for practical test construction. Discusses theoretical results from NIRT often relevant to IRT. Contains 134 references. (SLD)
Descriptors: Item Response Theory, Nonparametric Statistics, Research Methodology, Scores
Hessen, David J. – Psychometrika, 2005
In the present paper, a new family of item response theory (IRT) models for dichotomous item scores is proposed. Two basic assumptions define the most general model of this family. The first assumption is local independence of the item scores given a unidimensional latent trait. The second assumption is that the odds-ratios for all item-pairs are…
Descriptors: Item Response Theory, Scores, Test Items, Models

Meijer, Rob R.; And Others – Applied Psychological Measurement, 1994
The power of the nonparametric person-fit statistic, U3, is investigated through simulations as a function of item characteristics, test characteristics, person characteristics, and the group to which examinees belong. Results suggest conditions under which relatively short tests can be used for person-fit analysis. (SLD)
Descriptors: Difficulty Level, Group Membership, Item Response Theory, Nonparametric Statistics

Meijer, Rob R.; And Others – Applied Measurement in Education, 1996
Several existing group-based statistics to detect improbable item score patterns are discussed, along with the cut scores proposed in the literature to classify an item score pattern as aberrant. A simulation study and an empirical study are used to compare the statistics and their use and to investigate the practical use of cut scores. (SLD)
Descriptors: Achievement Tests, Classification, Cutting Scores, Identification
Gierl, Mark J.; Tan, Xuan; Wang, Changjiang – College Board, 2005
The results of this study conclude that there is a multidimensional basis for test score inferences on the mathematics and critical reading sections of the SAT. Results from the exploratory analyses indicate that the data are multidimensional, as mathematics displayed two dimensions and critical reading displayed three dimensions. The correlations…
Descriptors: College Entrance Examinations, Standardized Tests, Scores, Inferences