NotesFAQContact Us
Collection
Advanced
Search Tips
Showing all 12 results Save | Export
Mislevy, Robert J.; Behrens, John T.; Bennett, Randy E.; Demark, Sarah F.; Frezzo, Dennis C.; Levy, Roy; Robinson, Daniel H.; Rutstein, Daisy Wise; Shute, Valerie J.; Stanley, Ken; Winters, Fielding I. – National Center for Research on Evaluation, Standards, and Student Testing (CRESST), 2007
People use external knowledge representations (EKRs) to identify, depict, transform, store, share, and archive information. Learning how to work with EKRs is central to becoming proficient in virtually every discipline. As such, EKRs play central roles in curriculum, instruction, and assessment. Five key roles of EKRs in educational assessment are…
Descriptors: Educational Assessment, Computer Networks, Test Construction, Computer Assisted Testing
Peer reviewed Peer reviewed
Huynh, Huynh – Journal of Educational Statistics, 1986
Under the assumptions of classical measurement theory and the condition of normality, a formula is derived for the reliability of composite scores. The formula represents an extension of the Spearman-Brown formula to the case of truncated data. (Author/JAZ)
Descriptors: Computer Simulation, Error of Measurement, Expectancy Tables, Scoring Formulas
Peer reviewed Peer reviewed
Harris, Deborah J.; Subkoviak, Michael J. – Educational and Psychological Measurement, 1986
This study examined three statistical methods for selecting items for mastery tests: (1) pretest-posttest; (2) latent trait; and (3) agreement statistics. The correlation between the latent trait method and agreement statistics, proposed here as an alternative, was substantial. Results for the pretest-posttest method confirmed its reputed…
Descriptors: Computer Simulation, Correlation, Item Analysis, Latent Trait Theory
Morrison, Carol A.; Fitzpatrick, Steven J. – 1992
An attempt was made to determine which item response theory (IRT) equating method results in the least amount of equating error or "scale drift" when equating scores across one or more test forms. An internal anchor test design was employed with five different test forms, each consisting of 30 items, 10 in common with the base test and 5…
Descriptors: Comparative Analysis, Computer Simulation, Equated Scores, Error of Measurement
Peer reviewed Peer reviewed
Jarjoura, David; Kolen, Michael J. – Journal of Educational Statistics, 1985
An equating design in which two groups of examinees from slightly different populations are administered a different test form with a subset of common items is widely used. This paper presents standard errors and a simulation that verifies the equation for large samples for an equipercentile equating procedure for this design. (Author/BS)
Descriptors: Computer Simulation, Equated Scores, Error of Measurement, Estimation (Mathematics)
van der Linden, Wim J.; Adema, Jos J. – 1988
Two optimalization models for the construction of tests with a maximal value of coefficient alpha are given. Both models have a linear form and can be solved by using a branch-and-bound algorithm. The first model assumes an item bank calibrated under the Rasch model and can be used, for instance, when classical test theory has to serve as an…
Descriptors: Algorithms, Computer Simulation, Estimation (Mathematics), Foreign Countries
Peer reviewed Peer reviewed
Stout, William – Psychometrika, 1987
A procedure--based on item response theory--for testing the hypothesis of unidimensionality of the latent space is proposed. Use of the procedure is supported by an asymptotic theory and a Monte Carlo simulation study. The procedure tests for unidimensionality in test construction and/or compares two tests. (SLD)
Descriptors: College Entrance Examinations, Computer Simulation, Equations (Mathematics), Hypothesis Testing
Weitzman, R. A. – 1982
The goal of this research was to predict from a recruit's responses to the Armed Services Vocational Aptitude Battery (ASVAB) items whether the recruit would pass the Armed Forces Qualification Test (AFQT). The data consisted of the responses (correct/incorrect) of 1,020 Navy recruits to 200 items of the ASVAB together with the scores of these…
Descriptors: Adults, Armed Forces, Computer Oriented Programs, Computer Simulation
Peer reviewed Peer reviewed
Harrison, David A. – Journal of Educational Statistics, 1986
Multidimensional item response data were created. The strength of a general factor, the number of common factors, the distribution of items loadingon common factors, and the number of items in simulated tests were manipulated. LOGIST effectively recovered both item and trait parameters in nearly all of the experimental conditions. (Author/JAZ)
Descriptors: Adaptive Testing, Computer Assisted Testing, Computer Simulation, Correlation
Ackerman, Terry A. – 1987
The purpose of this study was to investigate the effect of using multidimensional items in a computer adaptive test (CAT) setting which assumes a unidimensional item response theory (IRT) framework. Previous research has suggested that the composite of multidimensional abilities being estimated by a unidimensional IRT model is not constant…
Descriptors: Adaptive Testing, College Entrance Examinations, Computer Assisted Testing, Computer Simulation
Vale, C. David; Gialluca, Kathleen A. – 1985
ASCAL is a microcomputer-based program for calibrating items according to the three-parameter logistic model of item response theory. It uses a modified multivariate Newton-Raphson procedure for estimating item parameters. This study evaluated this procedure using Monte Carlo Simulation Techniques. The current version of ASCAL was then compared to…
Descriptors: Adaptive Testing, Bayesian Statistics, Computer Assisted Testing, Computer Simulation
Levine, Michael V.; Drasgow, Fritz – 1984
Some examinees' test-taking behavior may be so idiosyncratic that their scores are not comparable to the scores of more typical examinees. Appropriateness indices, which provide quantitative measures of response-pattern atypicality, can be viewed as statistics for testing a null hypothesis of normal test-taking behavior against an alternative…
Descriptors: Cheating, College Entrance Examinations, Computer Simulation, Estimation (Mathematics)