NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 7 results Save | Export
Peer reviewed Peer reviewed
Wilcox, Rand R. – Journal of Educational Measurement, 1987
Four procedures are discussed for obtaining a confidence interval when answer-until-correct scoring is used in multiple choice tests. Simulated data show that the choice of procedure depends upon sample size. (GDC)
Descriptors: Computer Simulation, Multiple Choice Tests, Sample Size, Scoring
Peer reviewed Peer reviewed
Direct linkDirect link
Lee, Won-Chan; Brennan, Robert L.; Kolen, Michael J. – Journal of Educational and Behavioral Statistics, 2006
Assuming errors of measurement are distributed binomially, this article reviews various procedures for constructing an interval for an individual's true number-correct score; presents two general interval estimation procedures for an individual's true scale score (i.e., normal approximation and endpoints conversion methods); compares various…
Descriptors: Probability, Intervals, Guidelines, Computer Simulation
Peer reviewed Peer reviewed
Zwick, Rebecca; And Others – Journal of Educational Measurement, 1995
In a simulation study of ability and estimation of differential item functioning (DIF) in computerized adaptive tests, Rasch-based DIF statistics were highly correlated with generating DIF, but DIF statistics tended to be slightly smaller than in the three-parameter logistic model analyses. (SLD)
Descriptors: Ability, Adaptive Testing, Computer Assisted Testing, Computer Simulation
Peer reviewed Peer reviewed
Lin, Miao-Hsiang; Hsiung, Chao A. – Psychometrika, 1992
Four bootstrap methods are identified for constructing confidence intervals for the binomial-error model. The extent to which similar results are obtained and the theoretical foundation of each method and its relevance and ranges of modeling the true score uncertainty are discussed. (SLD)
Descriptors: Bayesian Statistics, Computer Simulation, Equations (Mathematics), Estimation (Mathematics)
Peer reviewed Peer reviewed
Hirsch, Thomas M. – Journal of Educational Measurement, 1989
Equatings were performed on both simulated and real data sets using common-examinee design and two abilities for each examinee. Results indicate that effective equating, as measured by comparability of true scores, is possible with the techniques used in this study. However, the stability of the ability estimates proved unsatisfactory. (TJH)
Descriptors: Academic Ability, College Students, Comparative Analysis, Computer Assisted Testing
Peer reviewed Peer reviewed
Houston, Walter M.; And Others – Applied Psychological Measurement, 1991
The effectiveness of alternative procedures to correct for rater leniency/stringency effects was studied when true scores were known. Ordinary least squares, weighted least squares, and imputation of the missing data consistently outperformed averaging the observed ratings; and the imputation technique was superior to the least squares methods.…
Descriptors: Comparative Analysis, Computer Simulation, Educational Assessment, Equations (Mathematics)
Peer reviewed Peer reviewed
Donoghue, John R.; Cliff, Norman – Applied Psychological Measurement, 1991
The validity of the assumptions under which the ordinal true score test theory was derived was examined using (1) simulation based on classical test theory; (2) a long empirical test with data from 321 sixth graders; and (3) an extensive simulation with 480 datasets based on the 3-parameter model. (SLD)
Descriptors: Computer Simulation, Elementary Education, Elementary School Students, Equations (Mathematics)