Publication Date
In 2025 | 0 |
Since 2024 | 0 |
Since 2021 (last 5 years) | 0 |
Since 2016 (last 10 years) | 0 |
Since 2006 (last 20 years) | 7 |
Descriptor
Error Patterns | 13 |
Item Response Theory | 9 |
Simulation | 7 |
Evaluation Methods | 6 |
Test Items | 6 |
Error of Measurement | 4 |
Computation | 3 |
Higher Education | 3 |
Models | 3 |
Computer Assisted Testing | 2 |
Nonparametric Statistics | 2 |
More ▼ |
Source
Applied Psychological… | 13 |
Author
Attali, Yigal | 1 |
Baker, Frank B. | 1 |
Camilli, Gregory | 1 |
Chan, Tsze | 1 |
Chiu, Ting-Wei | 1 |
Cohen, Jon | 1 |
Doebler, Anna | 1 |
Finkelman, Matthew D. | 1 |
Hoogstraten, Joh. | 1 |
Hung, Lai-Fa | 1 |
Jiang, Tao | 1 |
More ▼ |
Publication Type
Journal Articles | 13 |
Reports - Research | 7 |
Reports - Evaluative | 4 |
Information Analyses | 1 |
Reports - Descriptive | 1 |
Reports - General | 1 |
Education Level
Higher Education | 1 |
Audience
Researchers | 1 |
Location
Taiwan | 1 |
Laws, Policies, & Programs
Assessments and Surveys
What Works Clearinghouse Rating
Chiu, Ting-Wei; Camilli, Gregory – Applied Psychological Measurement, 2013
Guessing behavior is an issue discussed widely with regard to multiple choice tests. Its primary effect is on number-correct scores for examinees at lower levels of proficiency. This is a systematic error or bias, which increases observed test scores. Guessing also can inflate random error variance. Correction or adjustment for guessing formulas…
Descriptors: Item Response Theory, Guessing (Tests), Multiple Choice Tests, Error of Measurement
Doebler, Anna – Applied Psychological Measurement, 2012
It is shown that deviations of estimated from true values of item difficulty parameters, caused for example by item calibration errors, the neglect of randomness of item difficulty parameters, testlet effects, or rule-based item generation, can lead to systematic bias in point estimation of person parameters in the context of adaptive testing.…
Descriptors: Adaptive Testing, Computer Assisted Testing, Computation, Item Response Theory
Hung, Lai-Fa – Applied Psychological Measurement, 2012
Rasch used a Poisson model to analyze errors and speed in reading tests. An important property of the Poisson distribution is that the mean and variance are equal. However, in social science research, it is very common for the variance to be greater than the mean (i.e., the data are overdispersed). This study embeds the Rasch model within an…
Descriptors: Social Science Research, Markov Processes, Reading Tests, Social Sciences
Attali, Yigal – Applied Psychological Measurement, 2011
Recently, Attali and Powers investigated the usefulness of providing immediate feedback on the correctness of answers to constructed response questions and the opportunity to revise incorrect answers. This article introduces an item response theory (IRT) model for scoring revised responses to questions when several attempts are allowed. The model…
Descriptors: Feedback (Response), Item Response Theory, Models, Error Correction
Finkelman, Matthew D.; Weiss, David J.; Kim-Kang, Gyenam – Applied Psychological Measurement, 2010
Assessing individual change is an important topic in both psychological and educational measurement. An adaptive measurement of change (AMC) method had previously been shown to exhibit greater efficiency in detecting change than conventional nonadaptive methods. However, little work had been done to compare different procedures within the AMC…
Descriptors: Computer Assisted Testing, Hypothesis Testing, Measurement, Item Analysis
Roberts, James S. – Applied Psychological Measurement, 2008
Orlando and Thissen (2000) developed an item fit statistic for binary item response theory (IRT) models known as S-X[superscript 2]. This article generalizes their statistic to polytomous unfolding models. Four alternative formulations of S-X[superscript 2] are developed for the generalized graded unfolding model (GGUM). The GGUM is a…
Descriptors: Item Response Theory, Goodness of Fit, Test Items, Models
Cohen, Jon; Chan, Tsze; Jiang, Tao; Seburn, Mary – Applied Psychological Measurement, 2008
U.S. state educational testing programs administer tests to track student progress and hold schools accountable for educational outcomes. Methods from item response theory, especially Rasch models, are usually used to equate different forms of a test. The most popular method for estimating Rasch models yields inconsistent estimates and relies on…
Descriptors: Testing Programs, Educational Testing, Item Response Theory, Computation

Meijer, Rob R. – Applied Psychological Measurement, 1994
Through simulation, the power of the U3 statistic was compared with the power of one of the simplest person-fit statistics, the sum of the number of Guttman errors. In most cases, a weighted version of the latter statistic performed as well as the U3 statistic. (SLD)
Descriptors: Error Patterns, Item Response Theory, Nonparametric Statistics, Power (Statistics)

Whitely, Susan E. – Applied Psychological Measurement, 1979
A model which gives maximum likelihood estimates of measurement error within the context of a simplex model for practice effects is presented. The appropriateness of the model is tested for five traits, and error estimates are compared to the classical formula estimates. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Higher Education, Mathematical Models

Baker, Frank B. – Applied Psychological Measurement, 1993
Using simulation, the effect that misspecification of elements in the weight matrix has on estimates of basic parameters of the linear logistic test model was studied. Results indicate that, because specifying elements of the weight matrix is a subjective process, it must be done with great care. (SLD)
Descriptors: Error Patterns, Estimation (Mathematics), Item Response Theory, Matrices

Kleinke, David J. – Applied Psychological Measurement, 1979
Lord's, Millman's and Saupe's methods of approximating the standard error of measurement are reviewed. Through an empirical demonstration involving 200 university classroom tests, all three approximations are shown to be biased. (Author/JKS)
Descriptors: Error of Measurement, Error Patterns, Higher Education, Mathematical Formulas

Hoogstraten, Joh. – Applied Psychological Measurement, 1979
The biasing effects of a pretest on subsequent post-test results were investigated in two experimental studies. In general, the results argue for using designs without pretests. (Author/JKS)
Descriptors: Control Groups, Error Patterns, Evaluation Methods, Higher Education
Wang, Wen-Chung; Su, Ya-Hui – Applied Psychological Measurement, 2004
Eight independent variables (differential item functioning [DIF] detection method, purification procedure, item response model, mean latent trait difference between groups, test length, DIF pattern, magnitude of DIF, and percentage of DIF items) were manipulated, and two dependent variables (Type I error and power) were assessed through…
Descriptors: Test Length, Test Bias, Simulation, Item Response Theory