Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 2 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 8 |
Descriptor
Sample Size | 14 |
Statistical Distributions | 14 |
Test Items | 14 |
Item Response Theory | 6 |
Simulation | 6 |
Ability | 5 |
Difficulty Level | 4 |
Error of Measurement | 4 |
Models | 4 |
Test Length | 4 |
Comparative Analysis | 3 |
More ▼ |
Source
ProQuest LLC | 3 |
Educational and Psychological… | 2 |
ACT, Inc. | 1 |
Educational Sciences: Theory… | 1 |
Journal of Educational… | 1 |
Author
Petersen, Nancy S. | 2 |
Chalmers, Robert Philip | 1 |
Cook, Linda L. | 1 |
Divers, Jasmin | 1 |
Ersen, Rabia Karatoprak | 1 |
Ito, Kyoko | 1 |
Kang, Taehoon | 1 |
Lee, Won-Chan | 1 |
Lin, Zhongtian | 1 |
MacDonald, George T. | 1 |
Meijer, Rob R. | 1 |
More ▼ |
Publication Type
Reports - Evaluative | 5 |
Reports - Research | 5 |
Speeches/Meeting Papers | 5 |
Journal Articles | 4 |
Dissertations/Theses -… | 3 |
Reports - Descriptive | 1 |
Education Level
Grade 8 | 1 |
Audience
Researchers | 1 |
Location
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Ersen, Rabia Karatoprak; Lee, Won-Chan – Journal of Educational Measurement, 2023
The purpose of this study was to compare calibration and linking methods for placing pretest item parameter estimates on the item pool scale in a 1-3 computerized multistage adaptive testing design in terms of item parameter recovery. Two models were used: embedded-section, in which pretest items were administered within a separate module, and…
Descriptors: Pretesting, Test Items, Computer Assisted Testing, Adaptive Testing
Paek, Insu; Lin, Zhongtian; Chalmers, Robert Philip – Educational and Psychological Measurement, 2023
To reduce the chance of Heywood cases or nonconvergence in estimating the 2PL or the 3PL model in the marginal maximum likelihood with the expectation-maximization (MML-EM) estimation method, priors for the item slope parameter in the 2PL model or for the pseudo-guessing parameter in the 3PL model can be used and the marginal maximum a posteriori…
Descriptors: Models, Item Response Theory, Test Items, Intervals
Sengul Avsar, Asiye; Tavsancil, Ezel – Educational Sciences: Theory and Practice, 2017
This study analysed polytomous items' psychometric properties according to nonparametric item response theory (NIRT) models. Thus, simulated datasets--three different test lengths (10, 20 and 30 items), three sample distributions (normal, right and left skewed) and three samples sizes (100, 250 and 500)--were generated by conducting 20…
Descriptors: Test Items, Psychometrics, Nonparametric Statistics, Item Response Theory
Padilla, Miguel A.; Divers, Jasmin – Educational and Psychological Measurement, 2013
The performance of the normal theory bootstrap (NTB), the percentile bootstrap (PB), and the bias-corrected and accelerated (BCa) bootstrap confidence intervals (CIs) for coefficient omega was assessed through a Monte Carlo simulation under conditions not previously investigated. Of particular interests were nonnormal Likert-type and binary items.…
Descriptors: Sampling, Statistical Inference, Computation, Statistical Analysis
Quesen, Sarah – ProQuest LLC, 2016
When studying differential item functioning (DIF) with students with disabilities (SWD) focal groups typically suffer from small sample size, whereas the reference group population is usually large. This makes it possible for a researcher to select a sample from the reference population to be similar to the focal group on the ability scale. Doing…
Descriptors: Test Items, Academic Accommodations (Disabilities), Testing Accommodations, Disabilities
MacDonald, George T. – ProQuest LLC, 2014
A simulation study was conducted to explore the performance of the linear logistic test model (LLTM) when the relationships between items and cognitive components were misspecified. Factors manipulated included percent of misspecification (0%, 1%, 5%, 10%, and 15%), form of misspecification (under-specification, balanced misspecification, and…
Descriptors: Simulation, Item Response Theory, Models, Test Items
Park, Sangwook – ProQuest LLC, 2011
Many studies have been conducted to evaluate the performance of DIF detection methods, when two groups have different ability distributions. Such studies typically have demonstrated factors that are associated with inflation of Type I error rates in DIF detection, such as mean ability differences. However, no study has examined how the direction…
Descriptors: Test Bias, Regression (Statistics), Sample Size, Simulation
Kang, Taehoon; Petersen, Nancy S. – ACT, Inc., 2009
This paper compares three methods of item calibration--concurrent calibration, separate calibration with linking, and fixed item parameter calibration--that are frequently used for linking item parameters to a base scale. Concurrent and separate calibrations were implemented using BILOG-MG. The Stocking and Lord (1983) characteristic curve method…
Descriptors: Standards, Testing Programs, Test Items, Statistical Distributions
Sotaridona, Leonardo S.; Meijer, Rob R. – 2001
Two new indices to detect answer copying on a multiple-choice test, S(1) and S(2) (subscripts), are proposed. The S(1) index is similar to the K-index (P. Holland, 1996) and the K-overscore(2), (K2) index (L. Sotaridona and R. Meijer, in press), but the distribution of the number of matching incorrect answers of the source (examinee s) and the…
Descriptors: Cheating, Multiple Choice Tests, Responses, Sample Size
Pommerich, Mary; And Others – 1995
The Mantel-Haenszel (MH) statistic for identifying differential item functioning (DIF) commonly conditions on the observed test score as a surrogate for conditioning on latent ability. When the comparison group distributions are not completely overlapping (i.e., are incongruent), the observed score represents different levels of latent ability…
Descriptors: Ability, Comparative Analysis, Difficulty Level, Item Bias
Narayanan, Pankaja; Swaminathan, H. – 1993
The purpose of this study was to compare two non-parametric procedures, the Mantel-Haenszel (MH) procedure and the simultaneous item bias (SIB) procedure, with respect to their Type I error rates and power, and to investigate the conditions under which asymptotic distributional properties of the SIB and MH were obtained. Data were simulated to…
Descriptors: Ability, Comparative Analysis, Computer Simulation, Control Groups
Nandakumar, Ratna; Yu, Feng – 1994
DIMTEST is a statistical test procedure for assessing essential unidimensionality of binary test item responses. The test statistic T used for testing the null hypothesis of essential unidimensionality is a nonparametric statistic. That is, there is no particular parametric distribution assumed for the underlying ability distribution or for the…
Descriptors: Ability, Content Validity, Correlation, Nonparametric Statistics
Ito, Kyoko; Sykes, Robert C. – 1994
Responses to previously calibrated items administered in a computerized adaptive testing (CAT) mode may be used to recalibrate the items. This live-data simulation study investigated the possibility, and limitations, of on-line adaptive recalibration of precalibrated items. Responses to items of a Rasch-based paper-and-pencil licensure examination…
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Difficulty Level
Cook, Linda L.; Petersen, Nancy S. – 1986
This paper examines how various equating methods are affected by: (1) sampling error; (2) sample characteristics; and (3) characteristics of anchor test items. It reviews empirical studies that investigated the invariance of equating transformations, and it discusses empirical and simulation studies that focus on how the properties of anchor tests…
Descriptors: Educational Research, Equated Scores, Error of Measurement, Evaluation Methods