Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 1 |
Since 2016 (last 10 years) | 3 |
Since 2006 (last 20 years) | 7 |
Descriptor
Computation | 7 |
Scores | 7 |
Test Length | 7 |
Test Items | 5 |
Error of Measurement | 3 |
Item Response Theory | 3 |
Sample Size | 3 |
Comparative Analysis | 2 |
Nonparametric Statistics | 2 |
Statistical Analysis | 2 |
Test Bias | 2 |
More ▼ |
Source
Applied Psychological… | 2 |
Applied Measurement in… | 1 |
International Journal of… | 1 |
Journal of Educational… | 1 |
Journal of Speech, Language,… | 1 |
ProQuest LLC | 1 |
Author
Christie, Robert E. | 1 |
Cui, Zhongmin | 1 |
Gelfand, Jessica T. | 1 |
Gelfand, Stanley A. | 1 |
Kim, Hyung Jin | 1 |
Kolen, Michael J. | 1 |
Lee, HyeSun | 1 |
Lee, Won-Chan | 1 |
Lee, Yi-Hsuan | 1 |
Md Desa, Zairul Nor Deana | 1 |
Nandakumar, Ratna | 1 |
More ▼ |
Publication Type
Journal Articles | 6 |
Reports - Research | 5 |
Dissertations/Theses -… | 1 |
Reports - Evaluative | 1 |
Education Level
Elementary Secondary Education | 1 |
Audience
Location
Iran | 1 |
Laws, Policies, & Programs
Assessments and Surveys
Trends in International… | 1 |
What Works Clearinghouse Rating
Kim, Hyung Jin; Lee, Won-Chan – Journal of Educational Measurement, 2022
Orlando and Thissen (2000) introduced the "S - X[superscript 2]" item-fit index for testing goodness-of-fit with dichotomous item response theory (IRT) models. This study considers and evaluates an alternative approach for computing "S - X[superscript 2]" values and other factors associated with collapsing tables of observed…
Descriptors: Goodness of Fit, Test Items, Item Response Theory, Computation
Lee, HyeSun – Applied Measurement in Education, 2018
The current simulation study examined the effects of Item Parameter Drift (IPD) occurring in a short scale on parameter estimates in multilevel models where scores from a scale were employed as a time-varying predictor to account for outcome scores. Five factors, including three decisions about IPD, were considered for simulation conditions. It…
Descriptors: Test Items, Hierarchical Linear Modeling, Predictor Variables, Scores
Lee, Yi-Hsuan; Zhang, Jinming – International Journal of Testing, 2017
Simulations were conducted to examine the effect of differential item functioning (DIF) on measurement consequences such as total scores, item response theory (IRT) ability estimates, and test reliability in terms of the ratio of true-score variance to observed-score variance and the standard error of estimation for the IRT ability parameter. The…
Descriptors: Test Bias, Test Reliability, Performance, Scores
Gelfand, Jessica T.; Christie, Robert E.; Gelfand, Stanley A. – Journal of Speech, Language, and Hearing Research, 2014
Purpose: Speech recognition may be analyzed in terms of recognition probabilities for perceptual wholes (e.g., words) and parts (e.g., phonemes), where j or the j-factor reveals the number of independent perceptual units required for recognition of the whole (Boothroyd, 1968b; Boothroyd & Nittrouer, 1988; Nittrouer & Boothroyd, 1990). For…
Descriptors: Phonemes, Word Recognition, Vowels, Syllables
Md Desa, Zairul Nor Deana – ProQuest LLC, 2012
In recent years, there has been increasing interest in estimating and improving subscore reliability. In this study, the multidimensional item response theory (MIRT) and the bi-factor model were combined to estimate subscores, to obtain subscores reliability, and subscores classification. Both the compensatory and partially compensatory MIRT…
Descriptors: Item Response Theory, Computation, Reliability, Classification
Nandakumar, Ratna; Yu, Feng; Zhang, Yanwei – Applied Psychological Measurement, 2011
DETECT is a nonparametric methodology to identify the dimensional structure underlying test data. The associated DETECT index, "D[subscript max]," denotes the degree of multidimensionality in data. Conditional covariances (CCOV) are the building blocks of this index. In specifying population CCOVs, the latent test composite [theta][subscript TT]…
Descriptors: Nonparametric Statistics, Statistical Analysis, Tests, Data
Cui, Zhongmin; Kolen, Michael J. – Applied Psychological Measurement, 2008
This article considers two methods of estimating standard errors of equipercentile equating: the parametric bootstrap method and the nonparametric bootstrap method. Using a simulation study, these two methods are compared under three sample sizes (300, 1,000, and 3,000), for two test content areas (the Iowa Tests of Basic Skills Maps and Diagrams…
Descriptors: Test Length, Test Content, Simulation, Computation