Publication Date
In 2025 | 1 |
Since 2024 | 1 |
Since 2021 (last 5 years) | 3 |
Since 2016 (last 10 years) | 4 |
Since 2006 (last 20 years) | 7 |
Descriptor
Error of Measurement | 10 |
Simulation | 10 |
Validity | 10 |
Monte Carlo Methods | 4 |
Reliability | 4 |
Item Response Theory | 3 |
Scores | 3 |
Comparative Analysis | 2 |
Estimation (Mathematics) | 2 |
Evaluation Methods | 2 |
Factor Analysis | 2 |
More ▼ |
Source
Educational and Psychological… | 2 |
International Journal of… | 2 |
Applied Measurement in… | 1 |
Psychometrika | 1 |
Research in the Schools | 1 |
Structural Equation Modeling:… | 1 |
Author
Publication Type
Journal Articles | 8 |
Reports - Research | 6 |
Reports - Evaluative | 4 |
Speeches/Meeting Papers | 2 |
Education Level
Audience
Location
Arkansas | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Finch, Holmes – Applied Measurement in Education, 2022
Much research has been devoted to identification of differential item functioning (DIF), which occurs when the item responses for individuals from two groups differ after they are conditioned on the latent trait being measured by the scale. There has been less work examining differential step functioning (DSF), which is present for polytomous…
Descriptors: Comparative Analysis, Item Response Theory, Item Analysis, Simulation
Manuel T. Rein; Jeroen K. Vermunt; Kim De Roover; Leonie V. D. E. Vogelsmeier – Structural Equation Modeling: A Multidisciplinary Journal, 2025
Researchers often study dynamic processes of latent variables in everyday life, such as the interplay of positive and negative affect over time. An intuitive approach is to first estimate the measurement model of the latent variables, then compute factor scores, and finally use these factor scores as observed scores in vector autoregressive…
Descriptors: Measurement Techniques, Factor Analysis, Scores, Validity
Koziol, Natalie A.; Goodrich, J. Marc; Yoon, HyeonJin – Educational and Psychological Measurement, 2022
Differential item functioning (DIF) is often used to examine validity evidence of alternate form test accommodations. Unfortunately, traditional approaches for evaluating DIF are prone to selection bias. This article proposes a novel DIF framework that capitalizes on regression discontinuity design analysis to control for selection bias. A…
Descriptors: Regression (Statistics), Item Analysis, Validity, Testing Accommodations
Raykov, Tenko; Marcoulides, George A.; Li, Tenglong – Educational and Psychological Measurement, 2017
The measurement error in principal components extracted from a set of fallible measures is discussed and evaluated. It is shown that as long as one or more measures in a given set of observed variables contains error of measurement, so also does any principal component obtained from the set. The error variance in any principal component is shown…
Descriptors: Error of Measurement, Factor Analysis, Research Methodology, Psychometrics
Duong, Minh Q.; von Davier, Alina A. – International Journal of Testing, 2012
Test equating is a statistical procedure for adjusting for test form differences in difficulty in a standardized assessment. Equating results are supposed to hold for a specified target population (Kolen & Brennan, 2004; von Davier, Holland, & Thayer, 2004) and to be (relatively) independent of the subpopulations from the target population (see…
Descriptors: Ability Grouping, Difficulty Level, Psychometrics, Statistical Analysis
Schmitt, T. A.; Sass, D. A.; Sullivan, J. R.; Walker, C. M. – International Journal of Testing, 2010
Imposed time limits on computer adaptive tests (CATs) can result in examinees having difficulty completing all items, thus compromising the validity and reliability of ability estimates. In this study, the effects of speededness were explored in a simulated CAT environment by varying examinee response patterns to end-of-test items. Expectedly,…
Descriptors: Monte Carlo Methods, Simulation, Computer Assisted Testing, Adaptive Testing
Hughes, Gail D. – Research in the Schools, 2009
The impacts of incorrect responses to reverse-coded survey items were examined in this simulation study by reversing responses to traditional Likert-format items from 700 administrators in randomly selected schools in a 7-county region in central Arkansas that were obtained from an archival dataset. Specifically, the number of reverse-coded items…
Descriptors: Surveys, Coding, Context Effect, Measures (Individuals)
Romano, Jeanine; Kromrey, Jeffrey D. – 2002
The purpose of this study was to examine the potential impact of selected methodological factors on the validity of conclusions from reliability generalization (RG) studies. The study focused on four factors; (1) missing data in the primary studies; (2) transformation of sample reliability estimates; (3) use of sample weights for estimating mean…
Descriptors: Error of Measurement, Monte Carlo Methods, Reliability, Research Methodology

Hutchinson, J. Wesley; Mungale, Amitabh – Psychometrika, 1997
A nonmetric algorithm, pairwise partitioning, is developed to identify feature-based similarity structures. Presents theorems about the validity of the features identified by the algorithm, and reports results of Monte Carlo simulations that estimate the probabilities of identifying valid features for different feature structures and amounts of…
Descriptors: Algorithms, Error of Measurement, Estimation (Mathematics), Identification
Lambert, Richard G.; Curlette, William L. – 1995
Validity generalization meta-analysis (VG) examines the extent to which the validity of an instrument can be transported across settings. VG offers correction and summarization procedures designed in part to remove the effects of statistical artifacts on estimates of association between criterion and predictor. By employing a random effects model,…
Descriptors: Correlation, Error of Measurement, Estimation (Mathematics), Meta Analysis